@Matthew Lohbihler  Have you seen this site? https://www.metacademy.org

That might (or might not) be useful for compiling a list of examples for
the "overview" site you speak of?



On Tue, Jun 30, 2015 at 1:07 PM, David Ragazzi <[email protected]>
wrote:

> Really nice, Julian! This is a noble attitude!
>
> On 30 June 2015 at 14:59, Matthew Taylor <[email protected]> wrote:
>
>> Wow, this is great guys. It just goes to show that we are all in this
>> game for the same payoff, and working together will get us further
>> along than throwing spears. Thanks for this.
>>
>> ---------
>> Matt Taylor
>> OS Community Flag-Bearer
>> Numenta
>>
>>
>> On Tue, Jun 30, 2015 at 10:51 AM, Julian Samaroo <[email protected]>
>> wrote:
>> > I apologize for being so hostile. As David and Matthew stated, yes, I
>> do get
>> > rather defensive when others assault the hard work that has been done
>> > throughout the years in the Machine Learning field, as I personally
>> look up
>> > to those researchers. I also should say that HTM is what really got me
>> > interested in ML starting off, so I have no hatred towards it. While I
>> don't
>> > like many of the biases exhibited here and on the public Gitter chat
>> room, I
>> > do realize that I am a part of it. Thus, moving forwards I think it
>> would
>> > help for all of us to shed these biases, and approach the pursuit of AI
>> from
>> > a more laid-back perspective, considering all approaches equally.
>> >
>> > Therefore, as Matthew suggested, I think it would be in all of our best
>> > interests if we might each attempt to add pieces to the puzzle, so to
>> speak.
>> > I clearly have more experience with other ML techniques distinct from
>> HTM
>> > and those residing in NuPIC, and I am indeed currently putting together
>> a
>> > project to showcase a combination of these algorithms, which is the
>> approach
>> > that I believe to be most likely to produce the AGI that we seek. And of
>> > course, Numenta and friends have more experience on the HTM and
>> neocortical
>> > aspects of cognition to press forwards and develop THE cortical
>> algorithm on
>> > which our cortices operate. I think it might be good then for each of
>> us to
>> > work on our respective pieces more-or-less separately, and divide up the
>> > work that will go into creating AI.
>> >
>> > Let me then layout the work that I have cutout for myself:
>> >
>> > Reinforcement learning, specifically using a singular reward/punishment
>> > signal to produce internal and external actions
>> > Error-driven learning (such as found in the cerebellum), to allow an AI
>> to
>> > model it's environment in a way which merges sensory and motor systems
>> on a
>> > fast timescale
>> > Episodic-like memory formation, such as found in the hippocampus, for
>> the
>> > storage and retrieval of "memories" of past or current events
>> > Working memory, as found in the PFC, for temporary storage and
>> retrieval of
>> > relevant information required at some later point (useful for matching
>> > tasks)
>> >
>> > Following from that, it seems that Numenta is already in it's preferred
>> > spot:
>> >
>> > Feature learning and encoding of diverse stimuli
>> > Pooling throughout a hierarchy, spatially and temporally
>> > Anomaly detection and prediction of future events or external states
>> > Sensorimotor prediction, utilizing motor feedback signals for tracking
>> the
>> > state of the AI
>> > Invariance and generalization to similar stimuli, while retaining the
>> > ability to differentiate distinct stimuli
>> >
>> > I'm sure I missed a few things on Numenta's side, but to me this seems
>> like
>> > a good division of labor, especially given that my work is heavily
>> based off
>> > the work of previous (and current) ML research, and should therefore
>> move
>> > along a bit quicker. Hopefully within the next few years both platforms
>> will
>> > be developed enough that they can be easily combined and still function
>> > effectively. However, given that HTM has had some issues with certain
>> > cortical-based problems, such as visual recognition, I'd like to suggest
>> > some well-designed algorithms to look at for ideas for future additions
>> or
>> > modifications. Many of them are also based on the cortex, so it should
>> be
>> > easy to understand the connections to HTM:
>> >
>> > Convolutional Neural Networks, for rapidly learning generic "filters"
>> (such
>> > as in the visual and auditory cortices)
>> > Echo-State Networks, also used in auditory areas for storage and recall
>> of
>> > short or long sequences of auditory representations
>> > Recurrent Sparse Autoencoders, for reproducing the input provided and
>> > extracting higher-level, more abstract features (can also be made
>> > temporally-aware quite easily)
>> >
>> > These are just a few examples, and I'm happy to give more as needed. I
>> hope
>> > that one day our algorithms can be combined to make something that we
>> can
>> > possibly call "Human", and thus enter the era of Artificial
>> Intelligence.
>> >
>> > Julian Samaroo
>> > Manager of Information Technology
>> > BluePrint Pathways, LLC
>> > (516) 993-1150
>> >
>> > On Tue, Jun 30, 2015 at 11:32 AM, Matthew Lohbihler
>> > <[email protected]> wrote:
>> >>
>> >> A fair summary. Thanks Matt.
>> >>
>> >>
>> >> On 6/30/2015 12:06 PM, Matthew Taylor wrote:
>> >>
>> >> Encoders matter to Numenta, and those are extra-cortical structures.
>> >> And you can't do sensorimotor work without extra-cortical structures
>> >> either, so I would not say that they don't matter to us.
>> >>
>> >> I would say that we do not care so much about creating biologically
>> >> accurate versions of extra-cortical structures.
>> >> ---------
>> >> Matt Taylor
>> >> OS Community Flag-Bearer
>> >> Numenta
>> >>
>> >>
>> >> On Tue, Jun 30, 2015 at 7:37 AM, Matthew Lohbihler
>> >> <[email protected]> wrote:
>> >>
>> >> Actually, he doesn't. Jeff talks about cortex all the time. I have
>> never
>> >> seen any talk of, research into, or plans to develop any other
>> structure.
>> >> Don't get me wrong: cortex is a key thing. But let's not pretend that,
>> >> publicly anyway, anything else matters much to Numenta at the moment.
>> >>
>> >>
>> >> On 6/30/2015 10:20 AM, Dillon Bender wrote:
>> >>
>> >> Right, Jeff talks about this all the time. An isolated cortex knows
>> >> virtually nothing and can cause nothing. It requires the sub-cortical
>> >> structures like the basal ganglia for learning sensorimotor perception
>> and
>> >> control. That aspect will no doubt need to be included in HTM in some
>> >> form.
>> >> But like he also says all the time, there’s no reason it has to
>> resemble
>> >> natural, humanoid functions. All the cortical principles will be
>> applied
>> >> generally to any sensory domain, limited by our imagination. No
>> >> circumvention of the biological algorithm is planned.
>> >>
>> >>
>> >>
>> >> - Dillon
>> >>
>> >>
>> >>
>> >> From: nupic [mailto:[email protected]] On Behalf Of
>> Matthew
>> >> Lohbihler
>> >> Sent: Tuesday, June 30, 2015 9:03 AM
>> >> To: Dillon Bender
>> >> Subject: Re: Response to Jeff Hawkins interview.
>> >>
>> >>
>> >>
>> >> I tend to agree with John. I suspect that intelligence developed upon a
>> >> neurological substrate without which that cortex can't function
>> >> completely.
>> >> Maybe, maybe, MI can still be developed by circumventing the substrate,
>> >> but
>> >> we'll learn so much more by developing it too.
>> >>
>> >> On 6/30/2015 9:49 AM, Dillon Bender wrote:
>> >>
>> >> <John> "And I think we'll have to work our way through the whole animal
>> >> kingdom to get a humanoid robot working."
>> >>
>> >>
>> >>
>> >> If what you mean is that researchers should start with building simple
>> >> organisms and then bolt on the more recently evolved systems, then I
>> think
>> >> this is false. The human brain contains the entirety of non-mammal to
>> >> mammal
>> >> evolution, so there is no reason to model non-mammals.
>> >>
>> >>
>> >>
>> >> I think you have missed out on Numenta's current research goals to work
>> >> sensorimotor into CLA theory, because they realized before you that
>> >> intelligence "needs to be embodied with sensory-motor loop at the core
>> of
>> >> its functionality." They have stated many times that the previous
>> version
>> >> of
>> >> the theory modeled L2/3 of the cortex, and now adding L4 (and soon L5)
>> >> will
>> >> help close the sensorimotor loop.
>> >>
>> >>
>> >>
>> >> - Dillon
>> >>
>> >>
>> >>
>> >>
>> >>
>> >> -----Original Message-----
>> >>
>> >> From: nupic [mailto:[email protected]] On Behalf Of John
>> >> Blackburn
>> >>
>> >> Sent: Tuesday, June 30, 2015 4:55 AM
>> >>
>> >> To: Dillon Bender
>> >>
>> >> Subject: Re: Response to Jeff Hawkins interview.
>> >>
>> >>
>> >>
>> >> Sorry to reopen this thread, I missed it! David, I wanted to comment on
>> >> what
>> >> you said on Facebook:
>> >>
>> >>
>> >>
>> >> 2.) For the first time in human history, we have an algorithm which
>> models
>> >> activity in the neocortex and performs with true intelligence exactly
>> >> **how** the brain does it (its the HOW that is truly important here).
>> >> ...and
>> >> by the way, this was also contributed by Jeff Hawkins and Numenta.
>> >>
>> >>
>> >>
>> >> "performs with true intelligence" is a pretty bold claim. If this is
>> the
>> >> case, how come there are no very convincing examples of HTM working
>> with
>> >> human like intelligence? The Hotgym example is nice but it is really no
>> >> better than what could be achieved with many existing neural networks.
>> >> Echo
>> >> state networks have been around for years and can make temporal
>> >> predictions
>> >> quite well. I recently presented some time sequence data relating to a
>> >> bridge to this forum but HTM did not succeed in modelling this (ESNs
>> >> worked
>> >> much better). So outside of Hotgym, what really compelling demos do you
>> >> have? I've been away for a while so maybe I missed something...
>> >>
>> >>
>> >>
>> >> I am also rather concerned HTM needs swarming before it can model
>> >> anything.
>> >> Isn't that "cheating" in a way? It seems the HTM is rather fragile and
>> >> needs
>> >> a lot of help. The human brain does not have this luxury it just has to
>> >> cope
>> >> with whatever data it gets.
>> >>
>> >>
>> >>
>> >> I'm also not convinced the neocortex is everything as Jeff Hawkins
>> thinks.
>> >> I
>> >> seriously doubt the bulk of the brain is just scaffolding.
>> >>
>> >> I've been told birds have no neocortex but are capable of very
>> intelligent
>> >> behaviour including constructing tools. Meanwhile I don't see any AI
>> robot
>> >> capable of even ant-like intelligence. (ants are
>> >>
>> >> amazing!) Has anyone even constructed a robot based on HTM?
>> >>
>> >>
>> >>
>> >> Personally I don't think a a disembodied computer can ever be
>> intelligent
>> >> (not even ant-like intelligence). IMO a robot (and it must BE a robot)
>> >> needs
>> >> to be embodied with sensory-motor loop at the core of its
>> functionality to
>> >> start behaving like an animal. (animals are the only things we know
>> that
>> >> show intelligence: clouds don't, volcanos don't, computers don't). And
>> I
>> >> think we'll have to work our way through the whole animal kingdom to
>> get a
>> >> humanoid robot working.
>> >>
>> >>
>> >>
>> >> John.
>> >>
>> >>
>> >>
>> >> On Mon, May 25, 2015 at 10:17 PM, cogmission (David Ray)
>> >> <[email protected]> wrote:
>> >>
>> >> You're probably right :-)
>> >>
>> >>
>> >>
>> >> On Mon, May 25, 2015 at 4:16 PM, Matthew Lohbihler
>> >>
>> >> <[email protected]> wrote:
>> >>
>> >>
>> >>
>> >> Yes, I agree. Except for the part about checking up on us. As i
>> >>
>> >> mentioned before, indifference to us seems to me to be more the
>> >>
>> >> default than caring about us.
>> >>
>> >>
>> >>
>> >>
>> >>
>> >> On 5/25/2015 5:03 PM, cogmission (David Ray) wrote:
>> >>
>> >>
>> >>
>> >> Let me try and think this through. Only in the context of scarcity
>> >>
>> >> does the question of AGI **or** us come about. Where there is no
>> >>
>> >> scarcity, I think an AGI will just go about its business - peeking in
>> >>
>> >> from time to time to make sure we're doing ok. Why in a universe
>> >>
>> >> where it can go anywhere it wants and produce infinite energy and not
>> >>
>> >> be bound by our planet, would a super-super intelligent being even be
>> >>
>> >> obsessed over us, when it could merely go someplace else? I honestly
>> >>
>> >> thing that is the way it will be. (and maybe is already!)
>> >>
>> >>
>> >>
>> >> On Mon, May 25, 2015 at 3:56 PM, Matthew Lohbihler
>> >>
>> >> <[email protected]> wrote:
>> >>
>> >>
>> >>
>> >> Forgive me David, but these are very loose definitions, and i've
>> >>
>> >> lost track of how they relate back to what an AGI will think about
>> >>
>> >> humanity. But to use your terms - hopefully accurately - what if the
>> >>
>> >> AGI satisfies its sentient need for "others" by creating other AGIs,
>> >>
>> >> ones that it can love and appreciate? I doubt humans would ever be
>> >>
>> >> up such a task, unless 1) as pets, or 2) with cybernetic improvements.
>> >>
>> >>
>> >>
>> >> On 5/25/2015 4:37 PM, David Ray wrote:
>> >>
>> >>
>> >>
>> >> Observation is the phenomenon of distinction, in the domain of
>> language.
>> >>
>> >> The universe consists of two things, content and context. Content
>> >>
>> >> depends on its boundaries in order to exist. It depends on what it
>> >>
>> >> is not for it's being. Context is the space for things to be, though
>> >>
>> >> it is not quite space because space is yet another thing. It has no
>> >>
>> >> boundaries and it cannot be arrived at by assembling all of its
>> content.
>> >>
>> >>
>> >>
>> >> Ideas; love, hate, our sense of who we are, our histories what we
>> >>
>> >> know to be true all of those are content. Context is what allows for
>> >>
>> >> that stuff to be. And all of it lives in language without which there
>> >> would
>> >> be nothing.
>> >>
>> >> There maybe would be a "drift" but we wouldn't know about it and we
>> >>
>> >> wouldn't be able to observe it.
>> >>
>> >>
>> >>
>> >> Sent from my iPhone
>> >>
>> >>
>> >>
>> >> On May 25, 2015, at 3:26 PM, Matthew Lohbihler
>> >>
>> >> <[email protected]>
>> >>
>> >> wrote:
>> >>
>> >>
>> >>
>> >> You lost me. You seem to be working with definitions of
>> >>
>> >> "observation" and "space for thinking" that i'm unaware of.
>> >>
>> >>
>> >>
>> >> On 5/25/2015 4:14 PM, David Ray wrote:
>> >>
>> >>
>> >>
>> >> Matthew L.,
>> >>
>> >>
>> >>
>> >> It isn't a thought. It is there before observation or thoughts or
>> >>
>> >> thinking. It actually is the space for thinking to occur - it is the
>> >>
>> >> context that allows for thought. We bring it to the table - it is
>> >>
>> >> there before we are (ontologically speaking). (It being this sense
>> >>
>> >> of integrity/wholeness)
>> >>
>> >>
>> >>
>> >> Sent from my iPhone
>> >>
>> >>
>> >>
>> >> On May 25, 2015, at 2:59 PM, Matthew Lohbihler
>> >>
>> >> <[email protected]>
>> >>
>> >> wrote:
>> >>
>> >>
>> >>
>> >> Goodness. I thought we agreed that an AGI would not think like humans.
>> >>
>> >> And besides, "love" doesn't feel like something i want to depend on
>> >>
>> >> as obvious in a machine.
>> >>
>> >>
>> >>
>> >>
>> >>
>> >> On 5/25/2015 3:50 PM, David Ray wrote:
>> >>
>> >>
>> >>
>> >> If I can take this conversation into yet a different direction.
>> >>
>> >>
>> >>
>> >> I think we've all been dancing around The question of what belies
>> >>
>> >> the generation of morality or how will an AI derive its sense of
>> >>
>> >> ethics? Of course initially there will be those parameters that are
>> >>
>> >> programmed in - but eventually those will be gotten around.
>> >>
>> >>
>> >>
>> >> There has been a lot of research into this actually - though it's
>> >>
>> >> not common knowledge it is however knowledge developed over the
>> >>
>> >> observation of millions of people.
>> >>
>> >>
>> >>
>> >> The universe and all beings along the gradient of sentience observe
>> >>
>> >> (albeit perhaps unconsciously), a sense of what I will call
>> >>
>> >> integrity or "wholeness". We'd like to think that mankind steered
>> >>
>> >> itself through the ages toward notions of gentility and societal
>> >>
>> >> sophistication; but it didn't really. The idea that a group or
>> >>
>> >> different groups devised a grand plan to have it turn out this way is
>> >> totally preposterous.
>> >>
>> >>
>> >>
>> >> What is more likely is that there is a natural order to things and
>> >>
>> >> that is motion toward what works for the whole. I can't prove any of
>> >>
>> >> this but internally we all know when it's missing or when we are not
>> >>
>> >> in alignment with it. This ineffable sense is what love is - it's
>> concern
>> >> for the whole.
>> >>
>> >>
>> >>
>> >> So I say that any truly intelligent being, by virtue of existing in
>> >>
>> >> a substrate of integrity will have this built in and a super
>> >>
>> >> intelligent being will understand this - and that is ultimately the
>> >>
>> >> best chance for any single instance to survive is for the whole to
>> >> survive.
>> >>
>> >>
>> >>
>> >> Yes I know immediately people want to cite all the aberrations and
>> >>
>> >> of course yes there are aberrations just as there are mutations but
>> >>
>> >> those aberrations our reactions to how a person is shown love during
>> >>
>> >> their development.
>> >>
>> >>
>> >>
>> >> Like I said I can't prove any of this but eventually it will bear
>> >>
>> >> itself out and we will find it to be so in the future.
>> >>
>> >>
>> >>
>> >> You can be skeptical if you want to but ask yourself some questions.
>> >>
>> >> Why is it that we all know when it's missing
>> >>
>> >> (fairness/justice/integrity)? Why is it that we develop open source
>> >>
>> >> software and free software? Why is it that despite our greed and
>> >>
>> >> insecurity society moves toward freedom and equality for everyone?
>> >>
>> >>
>> >>
>> >> One more question. Why is it that the most advanced philosophical
>> >>
>> >> beliefs cite that where we are located as a phenomenological event,
>> >>
>> >> is not in separate bodies?
>> >>
>> >>
>> >>
>> >> I know this kind of talk doesn't go over well in this crowd of
>> >>
>> >> concrete thinkers but I know that there is some science somewhere that
>> >> backs
>> >> this up.
>> >>
>> >>
>> >>
>> >> Sent from my iPhone
>> >>
>> >>
>> >>
>> >> On May 25, 2015, at 2:12 PM, vlab <[email protected]> wrote:
>> >>
>> >>
>> >>
>> >> Small point: Even if they did decide that our diverse intelligence
>> >>
>> >> is worth keeping around (having not already mapped it into silicon)
>> >>
>> >> why would they need all of us.  Surely 10% of the population would
>> >>
>> >> give them enough 'sample size' to get their diversity ration, heck
>> maybe
>> >> 1/10 of 1% would be
>> >>
>> >> enough.   They may find that we are wasting away the planet (oh, not
>> >> maybe,
>> >>
>> >> we are) and the planet would be more efficient and they could have
>> >>
>> >> more energy without most of us.  (Unless we become 'copper tops' as
>> >>
>> >> in the Matrix movie).
>> >>
>> >>
>> >>
>> >> On 5/25/2015 2:40 PM, Fergal Byrne wrote:
>> >>
>> >>
>> >>
>> >> Matthew,
>> >>
>> >>
>> >>
>> >> You touch upon the right point. Intelligence which can self-improve
>> >>
>> >> could only come about by having an appreciation for intelligence, so
>> >>
>> >> it's not going to be interested in destroying diverse sources of
>> >>
>> >> intelligence. We represent a crap kind of intelligence to such an AI
>> >>
>> >> in a certain sense, but one which it itself would rather communicate
>> >>
>> >> with than condemn its offspring to have to live like. If these
>> >>
>> >> things appear (which looks inevitable) and then they kill us, many
>> >>
>> >> of them will look back at us as a kind of "lost civilisation" which
>> >> they'll
>> >> struggle to reconstruct.
>> >>
>> >>
>> >>
>> >> The nice thing is that they'll always be able to rebuild us from the
>> >>
>> >> human genome. It's just a file of numbers after all.
>> >>
>> >>
>> >>
>> >> So, we have these huge threats to humanity. The AGI future is the
>> >>
>> >> only reversible one.
>> >>
>> >>
>> >>
>> >> Regards
>> >>
>> >> Fergal Byrne
>> >>
>> >>
>> >>
>> >> --
>> >>
>> >>
>> >>
>> >> Fergal Byrne, Brenter IT
>> >>
>> >>
>> >>
>> >> Author, Real Machine Intelligence with Clortex and NuPIC
>> >>
>> >> https://leanpub.com/realsmartmachines
>> >>
>> >>
>> >>
>> >> Speaking on Clortex and HTM/CLA at euroClojure Krakow, June 2014:
>> >>
>> >> http://euroclojure.com/2014/
>> >>
>> >> and at LambdaJam Chicago, July 2014: http://www.lambdajam.com
>> >>
>> >>
>> >>
>> >> http://inbits.com - Better Living through Thoughtful Technology
>> >>
>> >> http://ie.linkedin.com/in/fergbyrne/ -
>> >>
>> >> https://github.com/fergalbyrne
>> >>
>> >>
>> >>
>> >> e:[email protected] t:+353 83 4214179 Join the quest for
>> >>
>> >> Machine Intelligence at http://numenta.org Formerly of Adnet
>> >>
>> >> [email protected] http://www.adnet.ie
>> >>
>> >>
>> >>
>> >>
>> >>
>> >> On Mon, May 25, 2015 at 7:27 PM, Matthew Lohbihler
>> >>
>> >> <[email protected]> wrote:
>> >>
>> >>
>> >>
>> >> I think Jeff underplays a couple of points, the main one being the
>> >>
>> >> speed at which an AGI can learn. Yes, there is a natural limit to
>> >>
>> >> how much experimentation in the real world can be done in a given
>> >>
>> >> amount of time. But we humans are already going beyond this with,
>> >>
>> >> for example, protein folding simulations, which speeds up the
>> >>
>> >> discovery of new drugs and such by many orders of magnitude. Any
>> >>
>> >> sufficiently detailed simulation could massively narrow down the
>> >>
>> >> amount of real world verification necessary, such that new
>> >>
>> >> discoveries happen more and more quickly, possibly at some point
>> >>
>> >> faster than we know the AGI is doing them. An intelligence
>> >>
>> >> explosion is not a remote possibility. The major risk here is what
>> Eliezer
>> >> Yudkowsky pointed out: not that the AGI is evil or something, but that
>> it
>> >> is
>> >> indifferent to humanity.
>> >>
>> >> No one yet goes out of their way to make any form of AI care about
>> >>
>> >> us (because we don't yet know how). What if an AI created
>> >>
>> >> self-replicating nanobots just to prove a hypothesis?
>> >>
>> >>
>> >>
>> >> I think Nick Bostrom's book is what got Stephen, Elon, and Bill all
>> >>
>> >> upset. I have to say it starts out merely interesting, but gets to
>> >>
>> >> a dark place pretty quickly. But he goes too far in the other
>> >>
>> >> direction, at the same time easily accepting that superinteligences
>> >>
>> >> have all manner of cognitive skill, but at the same time can't
>> >>
>> >> fathom the how humans might not like the idea of having our brain's
>> >>
>> >> pleasure centers constantly poked, turning us all into smiling idiots
>> (as
>> >> i
>> >> mentioned here:
>> >>
>> >> http://blog.serotoninsoftware.com/so-smart-its-stupid).
>> >>
>> >>
>> >>
>> >>
>> >>
>> >>
>> >>
>> >> On 5/25/2015 2:01 PM, Fergal Byrne wrote:
>> >>
>> >>
>> >>
>> >> Just one last idea in this. One thing that crops up every now and
>> >>
>> >> again in the Culture novels is the response of the Culture to
>> >>
>> >> Swarms, which are self-replicating viral machines or organisms.
>> >>
>> >> Once these things start consuming everything else, the AIs (mainly
>> >>
>> >> Ships and Hubs) respond by treating the swarms as a threat to the
>> >>
>> >> diversity of their Culture. They first try to negotiate, then
>> >>
>> >> they'll eradicate. If they can contain them, they'll do that.
>> >>
>> >>
>> >>
>> >> They do this even though they can themselves withdraw from real
>> >>
>> >> spacetime. They don't have to worry about their own survival. They
>> >>
>> >> do this simply because life is more interesting when it includes all
>> the
>> >> rest of us.
>> >>
>> >>
>> >>
>> >> Regards
>> >>
>> >>
>> >>
>> >> Fergal Byrne
>> >>
>> >>
>> >>
>> >> --
>> >>
>> >>
>> >>
>> >> Fergal Byrne, Brenter IT
>> >>
>> >>
>> >>
>> >> Author, Real Machine Intelligence with Clortex and NuPIC
>> >>
>> >> https://leanpub.com/realsmartmachines
>> >>
>> >>
>> >>
>> >> Speaking on Clortex and HTM/CLA at euroClojure Krakow, June 2014:
>> >>
>> >> http://euroclojure.com/2014/
>> >>
>> >> and at LambdaJam Chicago, July 2014: http://www.lambdajam.com
>> >>
>> >>
>> >>
>> >> http://inbits.com - Better Living through Thoughtful Technology
>> >>
>> >> http://ie.linkedin.com/in/fergbyrne/ -
>> >>
>> >> https://github.com/fergalbyrne
>> >>
>> >>
>> >>
>> >> e:[email protected] t:+353 83 4214179 Join the quest for
>> >>
>> >> Machine Intelligence at http://numenta.org Formerly of Adnet
>> >>
>> >> [email protected] http://www.adnet.ie
>> >>
>> >>
>> >>
>> >>
>> >>
>> >> On Mon, May 25, 2015 at 5:04 PM, cogmission (David Ray)
>> >>
>> >> <[email protected]> wrote:
>> >>
>> >>
>> >>
>> >> This was someone's response to Jeff's interview (see here:
>> >>
>> >> https://www.facebook.com/fareedzakaria/posts/10152703985901330)
>> >>
>> >>
>> >>
>> >> Please read and comment if you feel the need...
>> >>
>> >>
>> >>
>> >> Cheers,
>> >>
>> >> David
>> >>
>> >>
>> >>
>> >> --
>> >>
>> >> With kind regards,
>> >>
>> >>
>> >>
>> >> David Ray
>> >>
>> >> Java Solutions Architect
>> >>
>> >>
>> >>
>> >> Cortical.io
>> >>
>> >> Sponsor of:  HTM.java
>> >>
>> >>
>> >>
>> >> [email protected]
>> >>
>> >> http://cortical.io
>> >>
>> >>
>> >>
>> >>
>> >>
>> >>
>> >>
>> >>
>> >>
>> >>
>> >>
>> >>
>> >>
>> >>
>> >>
>> >>
>> >>
>> >>
>> >>
>> >>
>> >>
>> >>
>> >>
>> >> --
>> >>
>> >> With kind regards,
>> >>
>> >>
>> >>
>> >> David Ray
>> >>
>> >> Java Solutions Architect
>> >>
>> >>
>> >>
>> >> Cortical.io
>> >>
>> >> Sponsor of:  HTM.java
>> >>
>> >>
>> >>
>> >> [email protected]
>> >>
>> >> http://cortical.io
>> >>
>> >>
>> >>
>> >>
>> >>
>> >>
>> >>
>> >>
>> >>
>> >>
>> >>
>> >> --
>> >>
>> >> With kind regards,
>> >>
>> >>
>> >>
>> >> David Ray
>> >>
>> >> Java Solutions Architect
>> >>
>> >>
>> >>
>> >> Cortical.io
>> >>
>> >> Sponsor of:  HTM.java
>> >>
>> >>
>> >>
>> >> [email protected]
>> >>
>> >> http://cortical.io
>> >>
>> >>
>> >>
>> >>
>> >>
>> >>
>> >>
>> >>
>> >
>>
>>
>
>
> --
> David Ragazzi
> Master in Sofware Engineering (University of Liverpool-UK)
> OS community commiter at Numenta.org
> --
> Have you tried *NuPIC Studio*? Just check out
> https://github.com/nupic-community/nupic.studio and enjoy it!
> --
> "I think James Connolly, the Irish revolutionary, is right when he says that
> the only prophets are those who make their future. So we're not
> anticipating, we're working for it."
>



-- 
*With kind regards,*

David Ray
Java Solutions Architect

*Cortical.io <http://cortical.io/>*
Sponsor of:  HTM.java <https://github.com/numenta/htm.java>

[email protected]
http://cortical.io

Reply via email to