The article by Kurzweil seemed to be insightful.

Jim Bromer

On Thu, Jan 14, 2016 at 3:24 PM, Raymond D Roberts Jr. via AGI <
[email protected]> wrote:

>
> http://www.kurzweilai.net/why-doesnt-my-phone-understand-me-yet?utm_source=KurzweilAI+Daily+Newsletter&utm_campaign=0481d44bf4-UA-946742-1&utm_medium=email&utm_term=0_6de721fb33-0481d44bf4-282058098
>
> Raymond D. Roberts Jr.
>
>
> -----Original Message-----
> From: Jim Bromer <[email protected]>
> To: AGI <[email protected]>
> Sent: Thu, Jan 14, 2016 3:16 pm
> Subject: Re: [agi] Re: If Deep Learning is It then Why Are Search Engines
> Incapable of Thinking (Outside the Box or Otherwise)?
>
> I don't think that deep learning only applies to pure deep learning. It
> could be used as part of a system which is attuned to discovering
> relations. And it could return relationships (in language for example)
> which could then be evaluated. Supervised learning is part of machine and
> deep learning so a system which returns candidate samples that can be
> evaluated still could be classified as machine learning (or could be said
> to have something in common with machine learning).
>
> Older AI paradigms have a much more fixed definition than newer ones.
> Comparing Watson, which was apparently able to learn new things about
> language, to an old Expert System does not sound right. My argument is that
> almost all contemporary AI paradigms involve some kind of network of
> relations so to presuppose that an advanced nlp program cannot 'learn'
> about nlp does not make sense. When I get some time I will ask someone at
> IBM what the phrase "Deep NLP" denotes. Does it mean deep search nlp or
> does it mean something that is closer to deep learning nlp because there is
> no reason to rule the possibility that new relationships in nlp could be
> detected in an applied network (of some kind) and then be used as an
> abstraction to search for other cases that might have a similar *kind* of
> relationship.
>
> I am interested in what you said Ben but I get the sense that Watson was
> used to detect relationships in nlp which were then evaluated (probably in
> different ways.)
>
> Jim Bromer
>
> On Thu, Jan 14, 2016 at 10:30 AM, Ben Goertzel <[email protected]> wrote:
>
>> ***
>> My question was why haven't there been clear advances in search engine
>> technology in the 2 years since Deep Learning and Watson have made very
>> obvious advances in AI?
>> ***
>>
>> The Web is very big.   Internally within Google, the API calls you can
>> make against the whole Web are many fewer than the ones you can make
>> against, say, Wikipedia
>>
>> But business-wise, there is more $$ to be made in making crude searches
>> against the whole Web slightly less crude, than in making more refined and
>> intelligent searches against smaller text-bases...
>>
>> Also,
>>
>> -- Watson is basically an expert system, albeit a very clever one....
>> Expert system methods don't scale, not even fancy ones...
>>
>> -- Deep learning in its current form works best for high-dimensional
>> floating point data, not discrete data like text ....  Also, current deep
>> learning algorithms rely essentially on bottom-up pattern recognition, with
>> limited top-down feedback.  But real language understanding can't get
>> approximated very well without sophisticated top-down feedback....  I.e.,
>> image and speech understanding can get further without cognitive feedback,
>> than language understanding...
>>
>>
>> ... ben
>>
>>
>>
>> On Thu, Jan 14, 2016 at 11:22 PM, Jim Bromer <[email protected]> wrote:
>>
>>>
>>> http://www.ibm.com/blogs/think/2016/01/14/the-next-grand-challenge-computers-that-converse-like-people/
>>>
>>> Jim Bromer
>>>
>>> On Thu, Jan 14, 2016 at 10:20 AM, Jim Bromer <[email protected]>
>>> wrote:
>>>
>>>> I watched the Presenti presentation on youtube a few days ago.
>>>>
>>>> Neural Networks can learn but they cannot use that learning efficiently
>>>> in many important ways. Discrete AI can acquire more specific
>>>> (discrete) 'objects' as they learn. So back in the 90's people started
>>>> using hybrids that combined neural networks with discrete methods. Machine
>>>> learning includes advances on hybrid methods.
>>>>
>>>> Most discrete methods are built around networks of relations between
>>>> the data objects which represent 'concepts' or 'ideas' or 'knowledge' or
>>>> 'know how' or whatever it is that you want to call the data objects that
>>>> would be used to hold knowledge in a (more) discrete AI program. So a
>>>> contemporary discrete AI program is also going to be an implementation of a
>>>> network. The network may include numerical values but even if it doesn't
>>>> it probably will represent categories of association. That definition is
>>>> not meant to be complete because I am only trying to get an idea across:
>>>> Modern discrete AI methods involve network methods that can potentially be
>>>> seen as representatives of 'thought' that are more sophisticated than
>>>> neural networks. That makes sense.
>>>>
>>>> Presenti was talking about an IBM researcher from the 70s who found
>>>> that he could use statistical methods to *learn* about speech without a
>>>> linguist. That would be a form of machine learning. Therefore it is fairly
>>>> safe for me to conclude that Watson used machine learning in what Watson
>>>> researcher's called, "Deep NLP".
>>>>
>>>> My question was why haven't there been clear advances in search engine
>>>> technology in the 2 years since Deep Learning and Watson have made very
>>>> obvious advances in AI?  I did an image search for "cats" on google and it
>>>> was very good.  I only found one dog (a small dog which had been photo
>>>> shopped with multiple legs somewhat like a caterpillar.) I tried some other
>>>> searches on images and the results were also very good. The results were
>>>> really amazing. So there have been some advances on image searches in the
>>>> past 2 years. The search for "castles on the moon" did not distinguish
>>>> between castles pictured as being on the moon from castles with the moon in
>>>> the scene. So even though I am nit-picking to some extent the point is that
>>>> it looks like you have to train a deep learning neural network with a
>>>> narrow training sample in order to teach it to recognize something that
>>>> would require a little thinking outside the box. That was also a problem
>>>> with Watson. Its Deep NLP could be trained with all the questions from past
>>>> Jeopardy shows (and Jeopardy-style questions that researchers could create)
>>>> but can it be trained to handle juxtapositions of linguistic 'concepts'
>>>> that might require some thinking outside of the box? (Incidentally I tried
>>>> "cat in a box" and Google did very well. But when I tried "full stadium" it
>>>> did include pictures of stadiums that were not empty. I could spot them as
>>>> I was paging quickly through the images.) But I guess I there have been
>>>> some significant advances in the past 2 years. They just do not include
>>>> using language to refine your searches.
>>>>
>>>> My idea of Concept Integration is that different concepts cannot always
>>>> be merged, as in a neural network to take an example, because as more
>>>> concepts are integrated the requirements of a part of the conceptual
>>>> integration may change. To restate that in another way,the integration of a
>>>> number of concepts will typically change if additional concepts are
>>>> integrated with them. This is what would happen if you tried to refine your
>>>> search using conversation.
>>>>
>>>> Jim Bromer
>>>>
>>>>
>>>> On Tue, Jan 12, 2016 at 10:21 PM, LAU <> wrote:
>>>>
>>>>> OK, as you wish ... It's just a word. We do not agree on the
>>>>> signification. But, it's OK. If you call it "deep learning", or 
>>>>> "conceptual
>>>>> learning" ... or "Hakuna Matata's learning", it's not important. Stop
>>>>> playing with words.
>>>>>
>>>>> Try to back to the topic of this thread. If I understand what you want
>>>>> to promote :
>>>>> 1) You note that deep learning implemented in the industry is not so
>>>>> intelligent than espected taking in account the computation power 
>>>>> available.
>>>>> 2) Watson seems to less "narrow" than other implementations.
>>>>> 3) What it miss there is "conceptual integration".
>>>>> Correct me if I'm wrong.
>>>>>
>>>>>
>>>>> In my humble opinion, there's no intelligent machines just because
>>>>> people don't try to, or most likely don't figure out how to, make it more
>>>>> intelligent.
>>>>> Implementing"conceptual integration" is certainly a way that some
>>>>> researchers tried, but lead to no significant results until now. If I look
>>>>> at wikipedia, the theory is dated from 1990s. Twenty years later, still
>>>>> nothing.
>>>>>
>>>>> There's no magic behind deep learning, I mean neural networks, used by
>>>>> Google or Facebook. Very roughly, it's just a kind of "universal
>>>>> approximator". And it's not the computation power that with make it
>>>>> spontaneously more intelligent.
>>>>> Deep learning becomes very popular these last years because it's
>>>>> easier to make a neural network to accomplish picture or voice recognition
>>>>> task (*I've made a small one myself from scratch in few days*) than
>>>>> handcrafted codes, and for a better result.
>>>>> But basically, a neural network is just another kind of programming.
>>>>> Instead of coding a multitude of operation to achieve a complex task, a
>>>>> neural network can do it itself by learning from examples.
>>>>> And the question will be how to teach a neural network what is
>>>>> "conceptual integration" ?
>>>>>
>>>>> In the paris tech conference video (*on youtube, but it's in french
>>>>> ...*), Jerome Pesenti said something else interesting. He cite a
>>>>> IBM's 70s researcher, Fred Jelinek, who said "*Every time I fire a
>>>>> linguist the performance of the speech recognizer goes up*". The
>>>>> Jelinek speech recognizer team was composed by part of linguist and
>>>>> engineers. By replacing a linguist who treats language as a human do by an
>>>>> engineer who does mathematics and statistics on words, the result is
>>>>> better. And it seems to be the philosophy at IBM to work differently that 
>>>>> a
>>>>> human do, and it seems to give better result. Instead of playing jeopardy
>>>>> in human way, watson applies statistics on the database (*which was
>>>>> wikipedia*).
>>>>>
>>>>> What I want to say is that may be the "conceptual integration" is a
>>>>> track to explore for building AGI. Or, may be the solution will come from
>>>>> elsewhere.
>>>>>
>>>>>
>>>>> LAU
>>>>>
>>>>>
>>>>>
>>>>> Le 12/01/2016 10:46, Jim Bromer a écrit :
>>>>>
>>>>> Deep Learning is Deep Machine Learning and Machine Learning is in no
>>>>> way limited to Neural Networks. So there is no way that Deep Learning
>>>>> is going to be forever defined to refer Machine Learning that uses
>>>>> Neural Networks (in certain ways). From that point of view I can say
>>>>> that Watson-Jeopardy probably did use a kind of deep learning.
>>>>>
>>>>>
>>>>> -------------------------------------------
>>>>> AGI
>>>>> Archives: https://www.listbox.com/member/archive/303/=now
>>>>> RSS Feed: https://www.listbox.com/member/archive/rss/303/27172223-36de8e6c
>>>>> Modify Your Subscription: https://www.listbox.com/member/?&;
>>>>> Powered by Listbox: http://www.listbox.com
>>>>>
>>>>>
>>>>> *AGI* | Archives | Modify Your Subscription
>>>>>
>>>>
>>>>
>>> *AGI* | Archives | Modify Your Subscription
>>>
>>
>>
>>
>> --
>> Ben Goertzel, PhD
>> http://goertzel.org
>>
>> "The reasonable man adapts himself to the world: the unreasonable one
>> persists in trying to adapt the world to himself. Therefore all progress
>> depends on the unreasonable man." -- George Bernard Shaw
>> *AGI* | Archives | Modify Your Subscription
>>
>
> *AGI* | Archives | Modify Your Subscription
> *AGI* | Archives <https://www.listbox.com/member/archive/303/=now>
> <https://www.listbox.com/member/archive/rss/303/24379807-653794b5> |
> Modify
> <https://www.listbox.com/member/?&;>
> Your Subscription <http://www.listbox.com>
>



-------------------------------------------
AGI
Archives: https://www.listbox.com/member/archive/303/=now
RSS Feed: https://www.listbox.com/member/archive/rss/303/21088071-f452e424
Modify Your Subscription: 
https://www.listbox.com/member/?member_id=21088071&id_secret=21088071-58d57657
Powered by Listbox: http://www.listbox.com

Reply via email to