On Mon, Mar 20, 2023 at 10:15 AM Jason Resch <[email protected]> wrote:

Jason, that was a very interesting and insightful post, thanks for posting
it.



John K Clark    See what's on my new list at  Extropolis
<https://groups.google.com/g/extropolis>

i70


>
> On Mon, Mar 20, 2023, 9:51 AM Telmo Menezes <[email protected]>
> wrote:
>
>>
>>
>> Am Mo, 20. Mär 2023, um 14:28, schrieb Jason Resch:
>>
>> The video John shared is worth watching. This is significant. It is now
>> solving complex math problems which requires a long sequence of steps.
>>
>>
>> I agree that it is significant and extremely impressive. I never said the
>> opposite. What baffles me is that John is now requiring religious reverence
>> towards a scientific result, and criticizing when I ask questions that are
>> part of the same standard machine learning methodology that got us here.
>>
>
> I see, I appreciate that clarification.
>
>
>
>> Over-fitting is less of an issue here because it's trivial to write a
>> sentence that's never before been written by any human in history.
>>
>>
>> That is not enough. A small variation on a standard IQ test is still the
>> same IQ test for a super powerful pattern detector such as GPT-4.
>>
>> I have no doubt that GPT-4 can generalize in its domain. It was
>> rigorously designed and tested for that by people who know what they are
>> doing. My doubt is that you can give it an IQ test and claim OMG GPT-4 IQ >
>> 140. This is just silly and it is junk science.
>>
>
> It's true that once one learns a way to solve problems it becomes easier
> to reapply that method when you next encounter a related problem.
>
> But isn't that partly what intelligence is? If a system has read the whole
> Internet and seen every type of problem we know how to solve, and it can
> generalize to know what method to use in any situation, that's an
> incredible level of intelligence which until now, we haven't had in machine
> form before.
>
>
>
>> You can tweak the parameters of the problem to guarantee it's a problem
>> it has never before been seen, and it can still solve it.
>>
>>
>> Some yes, some no. Almost one century of computer science still applies.
>>
>> You can choose to wait for the academic write ups to come out a few
>> months down the line but by then things will have advanced another few
>> levels from where we are today.
>>
>>
>> I am not wanting to wait for anything, I am asking questions that can be
>> addressed right now:
>>
>> - Are there IQ tests in the training data of GPT-4. Yes or no?
>> - Can we conceive of human-level intelligence without recurrent
>> connections or some form of ongoing recursivity / Turing completeness? Yes
>> or no?
>>
>
>
> I've been thinking about this a lot.
>
>
> My friend with access to GPT-4 asked it: "Does your neural network contain
> such reflexive loops, or is it strictly feed forward?", below is its reply:
>
> --------
>
> "As an AI language model, my underlying architecture is based on the
> Transformer model, which is primarily feedforward in nature but includes
> some elements of recurrence or reflexivity, specifically in the
> self-attention mechanism.
>
> The Transformer model consists of a multi-layer feedforward neural network
> with self-attention mechanisms that allow the model to learn complex
> relationships between input and output sequences. The self-attention
> mechanism weighs the importance of different input elements relative to
> each other, effectively capturing long-range dependencies and relationships
> within the input data. This mechanism introduces a form of reflexivity or
> recurrence, as it allows the model to consider the context of the entire
> input sequence when generating outputs.
>
> However, it is important to note that the self-attention mechanism in the
> Transformer model is not equivalent to traditional recurrent neural
> networks (RNNs), which maintain a hidden state that is updated at each time
> step. The Transformer model processes input sequences in parallel rather
> than sequentially, which makes it fundamentally different from RNNs.
>
> In summary, while my neural network architecture is primarily feedforward,
> it includes some elements of reflexivity in the form of self-attention
> mechanisms that allow the model to capture complex relationships within
> input sequences."
>
> --------
>
> Is this enough to meet Hofstadter's requirements of recursion? I do not
> have the expertise to say. But I do see recursion exist in a way no one
> seems to ever mention:
>
> The output of the LLM is fed back in, as input to the LLM that produced
> it. So all the high level processing and operation of the network at the
> highest level, used to produce a few characters of output, then reaches
> back down to the lowest level to effect the lowest level of the input
> layers of the network.
>
> If you asked the network, where did that input that it sees come from, it
> would have no other choice but to refer back to itself, as "I". "I
> generated that text."
>
> Loops are needed to maintain and modify a persistent state or memory, to
> create a strange loop of self-reference, and to achieve Turing
> completeness. But a loop may not exist entirely in the "brain" of an
> entity, it might offload part of the loop into the environment in which it
> is operating. I think that is the case for things like thermostats, guided
> missiles, AlphaGo, and perhaps even ourselves.
>
> We observe our own actions, they become part of our sensory awareness and
> input. We cannot say exactly where they came from or how they were done,
> aside from modeling an "I" who seems to intercede in physics itself, but
> this is a consequence of being a strange loop. In a sense, our actions do
> come in from "on high", a higher level of abstraction in the hierarchy of
> processing, and this seems as if it is a dualistic interaction by a soul in
> heaven as Descartes described.
>
> In the case of GPT-4, its own output buffer can act as a scratch pad
> memory buffer, to which it continuously appends it's thoughts to. Is this
> not a form of memory and recursion?
>
> For one of the problems in John's video, it looked like it solved the
> Chinese remainder theorem in a series of discrete steps. Each step is
> written to and saved in it's output buffer, which becomes readable as it's
> input buffer.
>
> Given this, I am not sure we can say that GPT-4, in its current
> architecture and implementation, is entirely devoid of a memory, or a
> loop/recursion.
>
> I am anxious to hear your opinion though.
>
> Jason
>
> -
>

-- 
You received this message because you are subscribed to the Google Groups 
"Everything List" group.
To unsubscribe from this group and stop receiving emails from it, send an email 
to [email protected].
To view this discussion on the web visit 
https://groups.google.com/d/msgid/everything-list/CAJPayv3hypQ%3DwoCdNmNXMEeQRPfJwmfOc_2539xsG%2BESvr%2Bqvw%40mail.gmail.com.

Reply via email to