Re: [FRIAM] ChatGPT is not very smart...

2023-02-12 Thread Santafe
Sorry for the last post.

I hadn’t read the link below in this one yet.

I will now get back to the work I was supposed to have been doing.

Eric


> On Feb 10, 2023, at 2:22 PM, glen  wrote:
> 
> This was laugh out loud funny for me. YMMV.
> 
> Arguing with AI: My first dispute with Microsoft’s brilliant and boneheaded 
> Bing search engine
> https://linkprotect.cudasvc.com/url?a=https%3a%2f%2fwww.geekwire.com%2f2023%2farguing-with-ai-my-first-dispute-with-microsofts-brilliant-and-boneheaded-new-bing-search-engine%2f%3fmc_cid%3de4cf83ee41%26mc_eid%3dd5c14b4c73=E,1,P1Xq0EwlrN6G3WM0A2WTH1H7AomK__NRgyniiFYF0Rz_D4I0KuzYnL9EJlwOg8N1_yIPRafbdPnl4gf5U5NdwCTpUEKCUcjSpM4JRAbx=1
> 
> I quoted my comment about choosing your use case wisely. In this article, I 
> think it's clear that mixing the 2 use cases doesn't work out that well ... 
> yet, anyway.
> 
> On 2/10/23 10:14, glen wrote:
>> [⛧] I wanted to write something about how the use case of ChatGPT severely 
>> biases these perspectives on what GPT3.5 is or can do. But the post was too 
>> long already. It's sufficient to say there are many many *other* use cases 
>> for GPT3.5. ChatGPT ≠ GPT3.5.
> 
> -- 
> ꙮ Mɥǝu ǝlǝdɥɐuʇs ɟᴉƃɥʇ' ʇɥǝ ƃɹɐss snɟɟǝɹs˙ ꙮ
> 
> -. --- - / ...- .- .-.. .. -.. / -- --- .-. ... . / -.-. --- -.. .
> FRIAM Applied Complexity Group listserv
> Fridays 9a-12p Friday St. Johns Cafe   /   Thursdays 9a-12p Zoom 
> https://linkprotect.cudasvc.com/url?a=https%3a%2f%2fbit.ly%2fvirtualfriam=E,1,ir6m4SkmXohdtGP9Y3caumZaSTghHBR3r1ZpJp0VpbRqdr5TCHQj3E-hAR-6Qa-d1H7jiiBv2iUD661w4t3Rt57yWIhxmqiw0A0CH5_w=1
> to (un)subscribe 
> https://linkprotect.cudasvc.com/url?a=http%3a%2f%2fredfish.com%2fmailman%2flistinfo%2ffriam_redfish.com=E,1,VPEyN5poYN3a8hZyVNSsIeCcTmy0J-SiW0UDXicLHD9MhnOGERYwqjHe5h1UL-TmO7h-vQlShyVJOsSk-QknkoWKmzv6WLhwIXnjyC4AJ5PSe9BeUtkQ2g,,=1
> FRIAM-COMIC 
> https://linkprotect.cudasvc.com/url?a=http%3a%2f%2ffriam-comic.blogspot.com%2f=E,1,zMPnLB1bnITVYbabHbpKaUE1Zaum6_rZstvYRGMkYZE1xKWwjs32_grxsub07xDK0qTypXTKyxU_Xv1M4FCM4ELytb0eHR0wXqn7l58QtXFklrLpFw,,=1
> archives:  5/2017 thru present 
> https://linkprotect.cudasvc.com/url?a=https%3a%2f%2fredfish.com%2fpipermail%2ffriam_redfish.com%2f=E,1,wU-ASxBH_kmg4JPXUkyMIgcjQZMecZxiujPY7fT7ZCYhHI1frGWP-YXtGSZDr2oLs1fSvOMTcUtW0xvdJB4Ph1D7B9Y4V8IudlIpODwOrkyNUr8,=1
> 1/2003 thru 6/2021  http://friam.383.s1.nabble.com/


-. --- - / ...- .- .-.. .. -.. / -- --- .-. ... . / -.-. --- -.. .
FRIAM Applied Complexity Group listserv
Fridays 9a-12p Friday St. Johns Cafe   /   Thursdays 9a-12p Zoom 
https://bit.ly/virtualfriam
to (un)subscribe http://redfish.com/mailman/listinfo/friam_redfish.com
FRIAM-COMIC http://friam-comic.blogspot.com/
archives:  5/2017 thru present https://redfish.com/pipermail/friam_redfish.com/
  1/2003 thru 6/2021  http://friam.383.s1.nabble.com/


Re: [FRIAM] ChatGPT is not very smart...

2023-02-12 Thread Santafe
> On Feb 10, 2023, at 2:31 PM, Marcus Daniels  wrote:
> 
> I don't think it is necessarily the case a very large neural net would need 
> to backtrack.   It could devote more and more resources to different 
> hypotheses so long as it was not energy constrained.

In some sense, I understand that this must be right, because your language is 
more faithful to how the algorithm works.

I have another colleague who has commented that one needs to think differently 
about large neural-net learners because in very high dimensions, there are not 
the same kinds of snags as those encountered in lower dimensions.  There is 
always “a way around”, and therefore gradient descent works better than one has 
come to expect.

Yet in the output, there still seems to be a tree somehow; maybe from things 
that aren’t flexible by a design logic (?):

Time runs forward, for the machine as for the rest of us.

Having said something, the chatbot can’t un-say it.  

So then there is a design decision: Do we take what we have said as a 
constraint on what we can say next?  I can well imagine that there is strong 
reinforcement for some version of that, because it is inherent in fluidity, and 
even the notion of a “topic” in the pragmatics of a string of sentences.  If 
topicalization somehow means strict retention (doubling down), then our speech 
lives on a tree, and there seem to be certain ways of putting “more resources 
on different hypotheses” that become closed to us.  Glen’s characterization as 
mansplaining encapsulates this route nicely in one word.

Or, does the chatbot have a way to actually say Wow, shit.  I was confused; 
start over.  Then more of the dimensionality would be available.

Are there any chatbots that operate in this general space that people would 
call “admitting one’s mistakes”?  If not, why not?  Is it a design 
consideration in what to reward?  I can’t imagine it’s a property of the 
training set in the abstract, as there’s all sorts of literature on admitting 
mistakes (the wedding guest in the Ancient Mariner “a sadder but a wiser man”). 
 Are those patterns somehow “harder” to learn, even though they are there?  
What would make them harder, as a category of composition patterns?

A “third way”, I guess, is what one could call the “dumb pathological liar” (a 
sort of trump-like character), who simply rambles along in mutually 
contradicting utterances and doesn’t “structurally” acknowledge their 
existence.  Yet at some level of speech, that isn’t happening, because there is 
local fluidity, topics, and so forth.  At that lower level, the preference 
given to continuity seems very strong.


A fun aside about topicalization.  Our friend, the historical linguist Sergei 
Starostin, whose English was to my ear essentially blank and perfect, was once 
telling me that for a Russian speaker, the last latent point of fear in 
speaking English was the use of the definite article.  He said that no matter 
how many years on, he stil had this nagging stress about whether he was using 
it in the right places.  I don’t remember whether it was in that conversation — 
I think it was, and that the conversation was also about topicalization in 
Japanese with ga and wa, and in Chinese with topic-comment phrase structure 
(which French also uses in many constructions, but less structured around a 
“pivot” as it would be called for Chinese) — but either then or subsequently I 
came to appreciate what a problem topicalization is.  I would say it lives in 
speech at the level of pragmatics, in that one can almost see the “attention” 
as a kind of searchlight that is moving around, and that pragmatics is supposed 
to respect in the unfolding of a conversation for the conversation to be 
coherent.  The challenge of marking topic — one form of “definiteness” of the 
definite article versus an indefinite one — is that it involves this 
ever-negotiated problem of how much either from the discourse or from presumed 
shared knowledge the listener has in primed-awareness at any given moment.  
“The” drifts back and forth between implicit definiteness (I can just say “the 
moon”, without a further specifying clause, presuming that we both know there 
is only one), versus definiteness that demans a specifier (the senator from 
Wisconsin, when first introduced in the discourse).  I guess “the” in English 
is unusually fraught, in that its insertion or omission also modulates category 
terms versus literal instances (the AI chatbots say silly things, versus AI 
chatbots say silly things), and all these functional roles are in tension with 
each other at the same time.

So it’s all very attention-scope semantic.  Yet it can fail to be semantic at 
other levels.  What it would be about the encoding of speech that makes them so 
different is still hard for me to see.

Eric




> -Original Message-
> From: Friam  On Behalf Of Santafe
> Sent: Friday, February 10, 2023 3:11 AM
> To: The Friday Morning Applied Complexity Coffee