Grazie Daniela per questo riferimento.
La citazione ricordata da Ezra Klein nell'intervista a Gary Marcus, sul fatto che una cosa può essere falsa ma, a parte il resto, non inferiore alla cosa reale, mi ha fatto venire in mente l'articolo di Daniel Dennet sul fatto che generare finzione attraverso l'IA è un atto immorale come produrre denaro falso. https://www.theatlantic.com/technology/archive/2023/05/problem-counterfeit-people/674075/
Ciao, Enrico

------ Messaggio originale ------
Da "Daniela Tafani" <daniela.taf...@unipi.it>
A "J.C. DE MARTIN" <juancarlos.demar...@polito.it>; "Nexa" <nexa@server-nexa.polito.it>
Data 19/06/2024 13:15:55
Oggetto Re: [nexa] "ChatGPT is bullshit"

Molti altri - nessuno dei quali mi pare sia citato nell'articolo - hanno scritto di 
chatGPT come di un "bullshit generator", in senso tecnico, con riferimento a 
Frankfurt.

Ad esempio, Dan McQuillan, https://www.danmcquillan.org/chatgpt.html

Anche Richard Stallman usa spesso l'espressione:
https://stallman.org/archives/2024-jan-apr.html

Per quanto ne so (ma non ho indagato), il primo è stato Ezra Klein, nel 2023, 
intervistando Gary Marcus:

ChatGPT and systems like it, what they’re going to do right now is they’re 
going to drive the cost of producing text and images and code and, soon enough, 
video and audio to basically zero. It’s all going to look and sound and read 
very, very convincing. That is what these systems are learning how to do. They 
are learning how to be convincing. They are learning how to sound and seem 
human.

But they have no actual idea what they are saying or doing. It is bullshit. And 
I don’t mean bullshit as slang. I mean it in the classic philosophical 
definition by Harry Frankfurt. It is content that has no real relationship to 
the truth.

[...]

EZRA KLEIN: Let’s sit on that word truthful for a minute because it gets to, I 
think, my motivation in the conversation. I’ve been interested — I’m not an 
A.I. professional the way you are, but I’ve been interested for a long time. 
I’ve had Sam on the show, had Brian Christian on the show. And I was surprised 
by my mix of sort of wonder and revulsion when I started using ChatGPT because 
it is a very, very cool program. And in many ways, I find that its answers are 
much better than Google for a lot of what I would ask it.

But I know enough about how it works to know that, as you were saying, 
truthfulness is not one of the dimensions of it. It’s synthesizing. It’s sort 
of copying. It’s pastiching. And I was trying to understand why I was so 
unnerved by it. And it got me thinking, have you ever read this great 
philosophy paper by Harry Frankfurt called “On Bullshit”?

GARY MARCUS: I know the paper.

EZRA KLEIN: So this is a — welcome to the podcast, everybody — this is a 
philosophy paper about what is bullshit. And he writes, quote, “The essence of 
bullshit is not that it is false but that it is phony. In order to appreciate 
this distinction, one must recognize that a fake or a phony need not be in any 
respect, apart from authenticity itself, inferior to the real thing. What is 
not genuine may not also be defective in some other way. It may be, after all, 
an exact copy. What is wrong with a counterfeit is not what it is like, but how 
it was made.”

And his point is that what’s different between bullshit and a lie is that a lie 
knows what the truth is and has had to move in the other direction. He has this 
great line where he says that people telling the truth and people telling lies 
are playing the same game but on different teams. But bullshit just has no 
relationship, really, to the truth.

And what unnerved me a bit about ChatGPT was the sense that we are going to 
drive the cost of bullshit to zero when we have not driven the cost of truthful 
or accurate or knowledge advancing information lower at all. And I’m curious 
how you see that concern.

GARY MARCUS: It’s exactly right. These systems have no conception of truth. 
Sometimes they land on it and sometimes they don’t, but they’re all 
fundamentally bullshitting in the sense that they’re just saying stuff that 
other people have said and trying to maximize the probability of that. It’s 
just auto complete, and auto complete just gives you bullshit.

And it is a very serious problem. I just wrote an essay called something like 
“The Jurassic Park Moment for A.I.” And that Jurassic Park moment is exactly 
that. It’s when the price of bullshit reaches zero and people who want to 
spread misinformation, either politically or maybe just to make a buck, start 
doing that so prolifically that we can’t tell the difference anymore in what we 
see between truth and bullshit.

https://www.nytimes.com/2023/01/06/podcasts/transcript-ezra-klein-interviews-gary-marcus.html
________________________________________
Da: nexa <nexa-boun...@server-nexa.polito.it> per conto di J.C. DE MARTIN 
<juancarlos.demar...@polito.it>
Inviato: mercoledì 19 giugno 2024 10:51
A: Nexa
Oggetto: [nexa] "ChatGPT is bullshit"

ChatGPT is bullshit

Michael Townsen Hicks1 · James Humphries1 · Joe Slater1

Abstract
Recently, there has been considerable interest in large language models: 
machine learning systems which produce human- like text and dialogue. 
Applications of these systems have been plagued by persistent inaccuracies in 
their output; these are often called “AI hallucinations”. We argue that these 
falsehoods, and the overall activity of large language models, is better 
understood as bullshit in the sense explored by Frankfurt (On Bullshit, 
Princeton, 2005): the models are in an important way indifferent to the truth 
of their outputs. We distinguish two ways in which the models can be said to be 
bullshitters, and argue that they clearly meet at least one of these 
definitions. We further argue that describing AI misrepresentations as bullshit 
is both a more useful and more accurate way of predicting and discussing the 
behaviour of these systems.

https://link.springer.com/article/10.1007/s10676-024-09775-5

Hicks, M.T., Humphries, J. & Slater, J. ChatGPT is bullshit. Ethics Inf Technol 
26, 38 (2024). https://doi.org/10.1007/s10676-024-09775-5

  *   Published 08 June 2024

  *   DOI https://doi.org/10.1007/s10676-024-09775-5


-- EN

https://www.hoepli.it/libro/la-rivoluzione-informatica/9788896069516.html
======================================================
Prof. Enrico Nardelli
Past President di "Informatics Europe"
Direttore del Laboratorio Nazionale "Informatica e Scuola" del CINI
Dipartimento di Matematica - Universit� di Roma "Tor Vergata"
Via della Ricerca Scientifica snc - 00133 Roma
home page: https://www.mat.uniroma2.it/~nardelli
blog: https://link-and-think.blogspot.it/
tel: +39 06 7259.4204 fax: +39 06 7259.4699
mobile: +39 335 590.2331 e-mail: narde...@mat.uniroma2.it
online meeting: https://blue.meet.garr.it/b/enr-y7f-t0q-ont
======================================================          

--

Reply via email to