To me it seemed like such a long series of implications (A--> B--> C-->D ...) that, in my opinion, it became less and less likely to be true, especially without evidence. It appears to me, that as we evolve as a species we develop more ability to work cooperatively. For example, the United Nations, the World Courtin the Haig, Financial Markets, the international space station. In fact, there is more coordination now as war is faught not solely through military might, but also through the media, and via financial markets, and economic sanctions. This requires increased coordination, which humans have risen to meet. Humans learn by error, so runway technological acceleration exceeding human capacity is the same as the errors of Three Mile Island, Chernobyl, Fukushima,the BP Gulf Spill, Global warming, and a host of other disasters that have been encountered. If the error is not critical or fatal, then the capacity to learn from and compensate for the error is usually within the realm of human capabiliities.If the human species fails to adapt, then the species will suffer the consequences.I don't think AGI is necessary at all. Carl Sagan did mention though that becoming a space faring species is necessary. I'd submit that a serious space program is more important than a serious AGI program. Besides, 100% error avoidance is an impossible goal anyway. We learn from mistakes.We need them. ~PM
Date: Thu, 11 Dec 2014 20:57:55 -0800 Subject: Re: [agi] Re: The Need for Cross-Transformational Compressions From: [email protected] To: [email protected] It is a believable postulate based in our own experience. It does not require direct evidence and I didn't say it did. Why did you take one thing out and one that is not even essential? On Thu, Dec 11, 2014 at 5:30 PM, Piaget Modeler via AGI <[email protected]> wrote: > > My view is that AGI is *required* if human beings are going to have any > vibrant future at all and required within the next few decades. Why? > Humans, and logically any technological species, evolve with certain > species characteristics including limitations on their effective > intelligence and ability to work cooperatively well together. As a > runaway technological acceleration occurs the species eventually hits a > point where its effective intelligence is ineffective in capacity and > speed for coming up with timely good enough solutions to more and more > complex and rapidly developing problems. Eventually, without either AGI > or rewriting its own nature significantly, the species fails to make > good enough choices followed through well enough to avoid calamity. > > > - samantha > Samantha, what evidence is there of other comparable human species evolutionfollowing this scenario you've outlined? Is this a conjecture on your part or has this happened before, and if so, when? Kindly advise, ~PM AGI | Archives | Modify Your Subscription AGI | Archives | Modify Your Subscription ------------------------------------------- AGI Archives: https://www.listbox.com/member/archive/303/=now RSS Feed: https://www.listbox.com/member/archive/rss/303/21088071-f452e424 Modify Your Subscription: https://www.listbox.com/member/?member_id=21088071&id_secret=21088071-58d57657 Powered by Listbox: http://www.listbox.com
