https://www.vice.com/en/article/pkadgm/man-dies-by-suicide-after-talking-with-ai-chatbot-widow-says

On Fri, Mar 31, 2023 at 1:59 PM Jed Rothwell <jedrothw...@gmail.com> wrote:

> Here is another article about this, written by someone who says he is an
> AI expert.
>
> https://time.com/6266923/ai-eliezer-yudkowsky-open-letter-not-enough/
>
> QUOTE:
>
> Pausing AI Developments Isn't Enough. We Need to Shut it All Down
>
> An open letter published today calls for “all AI labs to immediately pause
> for at least 6 months the training of AI systems more powerful than GPT-4.”
>
> This 6-month moratorium would be better than no moratorium. I have respect
> for everyone who stepped up and signed it. It’s an improvement on the
> margin. . . .
>
> The key issue is not “human-competitive” intelligence (as the open letter
> puts it); it’s what happens after AI gets to smarter-than-human
> intelligence. Key thresholds there may not be obvious, we definitely can’t
> calculate in advance what happens when, and it currently seems imaginable
> that a research lab would cross critical lines without noticing.
>
> Many researchers steeped in these issues, including myself, expect that
> the most likely result of building a superhumanly smart AI, under anything
> remotely like the current circumstances, is that literally everyone on
> Earth will die. Not as in “maybe possibly some remote chance,” but as in
> “that is the obvious thing that would happen.”
>

Reply via email to