Sentient robots? Not possible if you do the maths
 
 So long, robot pals – and robot overlords. Sentient machines 
http://www.newscientist.com/article/mg21829171.900-consciousness-why-we-need-to-build-sentient-machines.html
 may never exist, according to a variation on a leading mathematical model of 
how our brains create consciousness 
http://www.newscientist.com/article/mg22229645.000-the-fourth-state-of-matter-consciousness.html.
 Over the past decade, Giulio Tononi at the University of Wisconsin-Madison and 
his colleagues have developed a mathematical framework for consciousness 
http://www.ploscompbiol.org/article/info%3Adoi%2F10.1371%2Fjournal.pcbi.1003588 
that has become one of the most influential theories in the field. According to 
their model, the ability to integrate information is a key property of 
consciousness. They argue that in conscious minds, integrated information 
cannot be reduced into smaller components. For instance, when a human perceives 
a red triangle, the brain cannot register the object as a colourless triangle 
plus a shapeless patch of red.
 But there is a catch, argues Phil Maguire http://www.cs.nuim.ie/~pmaguire/ at 
the National University of Ireland in Maynooth. He points to a computational 
device called the XOR logic gate, which involves two inputs, A and B. The 
output of the gate is "1" if A and B are the same and "0" if A and B are 
different. In this scenario, it is impossible to predict the output based on A 
or B alone – you need both.
 Memory edit Crucially, this type of integration requires loss of information, 
says Maguire: "You have put in two bits, and you get one out. If the brain 
integrated information in this fashion, it would have to be continuously 
haemorrhaging information."
 Maguire and his colleagues say the brain is unlikely to do this, because 
repeated retrieval of memories would eventually destroy them. Instead, they 
define integration in terms of how difficult information is to edit.
 
 Consider an album of digital photographs. The pictures are compiled but not 
integrated, so deleting or modifying individual images is easy. But when we 
create memories, we integrate those snapshots of information into our bank of 
earlier memories. This makes it extremely difficult to selectively edit out one 
scene from the "album" in our brain.
 Based on this definition, Maguire and his team have shown mathematically that 
computers can't handle any process that integrates information completely. If 
you accept that consciousness is based on total integration, then computers 
can't be conscious.
 Read the whole story here:
 
http://www.newscientist.com/article/dn25560-sentient-robots-not-possible-if-you-do-the-maths.html#.U3SA5_ldVZK
 
http://www.newscientist.com/article/dn25560-sentient-robots-not-possible-if-you-do-the-maths.html#.U3SA5_ldVZK

 But no, you won't find anything about actual transcending - whatever that is - 
that was a joke on John's surreal exposition from the weekend. I suspect that 
for a robot to transcend we would have to program it to have a set of emotional 
and spacial/visual recognition components that could be rearranged to reflect 
an inner state of disengagement from non-metaphorical stimulus/response 
routines and a corresponding increase of reward systems, possibly feedbacking 
into total rejection of programmed response to outer awareness in favour of 
self-satisfaction.
 Basically we'd have to build a robot to be just like us. But most scientists 
into the idea of artificial intelligence aren't thinking along those lines. In 
fact the best thing about this article is the link to one of the latest 
theories of consciousness:

 Integrated information theory (IIT) approaches the relationship between 
consciousness and its physical substrate by first identifying the fundamental 
properties of experience itself: existence, composition, information, 
integration, and exclusion. IIT then postulates that the physical substrate of 
consciousness must satisfy these very properties. We develop a detailed 
mathematical framework in which composition, information, integration, and 
exclusion are defined precisely and made operational. This allows us to 
establish to what extent simple systems of mechanisms, such as logic gates or 
neuron-like elements, can form complexes that can account for the fundamental 
properties of consciousness. Based on this principled approach, we show that 
IIT can explain many known facts about consciousness and the brain, leads to 
specific predictions, and allows us to infer, at least in principle, both the 
quantity and quality of consciousness for systems whose causal structure is 
known. For example, we show that some simple systems can be minimally 
conscious, some complicated systems can be unconscious, and two different 
systems can be functionally equivalent, yet one is conscious and the other one 
is not.

 

 Read all about it, but if you're used to thinking about consciousness in terms 
of touchy-feely spiritual-speak it'll be like cycling uphill with the brakes 
on. 
 It is worth the effort though:

 
http://www.ploscompbiol.org/article/info%3Adoi%2F10.1371%2Fjournal.pcbi.1003588 
http://www.ploscompbiol.org/article/info%3Adoi%2F10.1371%2Fjournal.pcbi.1003588

  



 

  • [FairfieldLife]... salyavin808
    • Re: [Fairf... TurquoiseBee turquoi...@yahoo.com [FairfieldLife]
      • Re: [F... Share Long sharelon...@yahoo.com [FairfieldLife]
    • Re: [Fairf... 'Richard J. Williams' pundits...@gmail.com [FairfieldLife]
    • Re: [Fairf... Bhairitu noozg...@sbcglobal.net [FairfieldLife]
      • Re: [F... fleetwood_macnche...@yahoo.com [FairfieldLife]
    • [Fairfield... jr_...@yahoo.com [FairfieldLife]

Reply via email to