Gorrell and Webb describe a neural implementation of LSA that seems more biologically plausible than the usual matrix factoring implementation. http://www.dcs.shef.ac.uk/~genevieve/gorrell_webb.pdf In the usual implementation, a word-word matrix A is factored to A = USV where S is diagonal (containing eigenvalues), and then the smaller elements of S are discarded. In the Gorrell model, U and V are the weights of a 3 layer neural network mapping words to words, and the nonzero elements of S represent the semantic space in the middle layer. As the network is trained, neurons are added to S. Thus, the network is trained online in a single pass, unlike factoring, which is offline.
-- Matt Mahoney, matmaho...@yahoo.com ________________________________ From: Gabriel Recchia <grecc...@gmail.com> To: agi <agi@v2.listbox.com> Sent: Wed, July 7, 2010 12:12:00 PM Subject: Re: [agi] Hutter - A fundamental misdirection? > In short, instead of a "pot of neurons", we might instead have a pot of > dozens >of types of > > neurons that each have their own complex rules regarding what other types of >neurons they > > can connect to, and how they process information... > ...there is plenty of evidence (from the slowness of evolution, the large >number (~200) > > of neuron types, etc.), that it is many-layered and quite complex... The disconnect between the low-level neural hardware and the implementation of algorithms that build conceptual spaces via dimensionality reduction--which generally ignore facts such as the existence of different types of neurons, the apparently hierarchical organization of neocortex, etc.--seems significant. Have there been attempts to develop computational models capable of LSA-style feats (e.g., constructing a vector space in which words with similar meanings tend to be relatively close to each other) that take into account basic facts about how neurons actually operate (ideally in a more sophisticated way than the nodes of early connectionist networks which, as we now know, are not particularly neuron-like at all)? If so, I would love to know about them. On Tue, Jun 29, 2010 at 3:02 PM, Ian Parker <ianpark...@gmail.com> wrote: The paper seems very similar in principle to LSA. What you need for a concept vector (or position) is the application of LSA followed by K-Means which will give you your concept clusters. > > >I would not knock Hutter too much. After all LSA reduces {primavera, >mamanthal, >salsa, resorte} to one word giving 2 bits saving on Hutter. > > > > > - Ian Parker > > > >On 29 June 2010 07:32, rob levy <r.p.l...@gmail.com> wrote: > >Sorry, the link I included was invalid, this is what I meant: >> >> >>http://www.geog.ucsb.edu/~raubal/Publications/RefConferences/ICSC_2009_AdamsRaubal_Camera-FINAL.pdf >> >> >> >> >>On Tue, Jun 29, 2010 at 2:28 AM, rob levy <r.p.l...@gmail.com> wrote: >> >>On Mon, Jun 28, 2010 at 5:23 PM, Steve Richfield <steve.richfi...@gmail.com> >>wrote: >>> >>>Rob, >>>> >>>>I just LOVE opaque postings, because they identify people who see things >>>>differently than I do. I'm not sure what you are saying here, so I'll make >>>>some >>>>"random" responses to exhibit my ignorance and elicit more explanation. >>>> >>>> >> >> >>>I think based on what you wrote, you understood (mostly) what I was trying >>>to >>>get across. So I'm glad it was at least quasi-intelligible. :) >>> >>> It sounds like this is a finer measure than the "dimensionality" that I was >>>referencing. However, I don't see how to reduce anything as quantized as >>>dimensionality into finer measures. Can you say some more about this? >>>> >>>> >> >> >>>I was just referencing Gardenfors' research program of "conceptual spaces" >>>(I >>>was intentionally vague about committing to this fully though because I >>>don't >>>necessarily think this is the whole answer). Page 2 of this article >>>summarizes >>>it pretty succinctly: >>>http://www.geog.ucsb.edu/.../ICSC_2009_AdamsRaubal_Camera-FINAL.pdf >>> >>> >>> >>>However, different people's brains, even the brains of identical twins, have >>>DIFFERENT mappings. This would seem to mandate experience-formed topology. >>>> >>>> >> >> >>>Yes definitely. >>> >>>Since these conceptual spaces that structure sensorimotor >>>expectation/prediction >>>(including in higher order embodied exploration of concepts I think) are >>>multidimensional spaces, it seems likely that some kind of neural >>>computation >>>over these spaces must occur, >>>> >>>>I agree. >>>> >>>> >>>>though I wonder what it actually would be in terms of neurons, (and if that >>>>matters). >>>> >>>>I don't see any route to the answer except via neurons. >> >> >>>I agree this is true of natural intelligence, though maybe in modeling, the >>>neural level can be shortcut to the topo map level without recourse to >>>neural >>>computation (use some more straightforward computation like matrix algebra >>>instead). >>> >>>Rob >> >>agi | Archives | Modify Your Subscription > >agi | Archives | Modify Your Subscription agi | Archives | Modify Your Subscription ------------------------------------------- agi Archives: https://www.listbox.com/member/archive/303/=now RSS Feed: https://www.listbox.com/member/archive/rss/303/ Modify Your Subscription: https://www.listbox.com/member/?member_id=8660244&id_secret=8660244-6e7fb59c Powered by Listbox: http://www.listbox.com