Proprietary code from within the M$ ecosystem is uninspired and bad code by
comparison. Open source code is the gold mine so M$ will not like being
told they cannot use open source to compile codex. It's a complete r*pe of
open source. GPT is trained on public language and language belongs to
people generally, not some select group. It's not meant to be a tool for
controlling people. GPT is literally the soul of a billion people and
should be public domain and not feared by GNU but instead rescued. Sorry
for the rhetoric!

On Sat, Jul 24, 2021 at 2:34 PM Shane Mulligan <[email protected]> wrote:

> This is why the technology is a bit like a
> personal Google search, Stackoverflow, which
> you can store offline because it's an index of the internet that is
> capable of reconstruction.
>
> But it's not limited to code generation. Codex
> is nothing. Emacs + GPT would carve a large
> piece out of M$.
>
> Codex is a model trained for the purpose of
> generating code, but GPT models will become
> abundant for all tasks, including image and
> audio synthesis and understanding.
>
> Emacs is a complete operating system.
> VSCode is geared towards programming.
>
> Emacs can do infinitely more things with GPT
> than VSCode can because it's holistic.
>
> Even the 'eliza' in emacs can pass the turing
> test with GPT. GPT can run sequences of commands in emacs to automate
> entire workflows with natural language.
>
> But the future is in collaborative GPT.
>
> The basis/base truth would become versions of
> LMs or ontologies.
>
> Right now that's EleutherAI.
>
> Shane Mulligan
>
> How to contact me:
> 🇦🇺 00 61 421 641 250
> 🇳🇿 00 64 21 1462 759 <+64-21-1462-759>
> [email protected]
>
>
> On Sat, Jul 24, 2021 at 2:10 PM Shane Mulligan <[email protected]>
> wrote:
>
>> It's a bit like whitewashing because it's
>> reconstructing generatively by finding
>> artificial/contrived associations between
>> different works that the author had not
>> intended but may have been part of their
>> inspiration inspiration, and it compresses the
>> information based on these assocations.
>>
>> It's a bit like running a lossy 'zip' on the
>> internet and then decompressing
>> probabilistically.
>>
>> When run deterministically (set the temperature of GPT to 0), you may
>> actually
>> see 'snippets' from various places, every time, with the same input
>> generating
>> the same snippets.
>>
>> So the source material is important.
>>
>> What GitHub did was very, very bad but they
>> did it anyway.
>>
>> That doesn't mean GPT is bad, it just means
>> they zipped up content they should not have
>> and created this language 'index' or ('codex'
>> is what they call it).
>>
>> What they really should do, if they are honest
>> people, is train the model on subsets of
>> GitHub code by separate licence and release
>> the models with the same license.
>>
>> Shane Mulligan
>>
>> How to contact me:
>> 🇦🇺 00 61 421 641 250
>> 🇳🇿 00 64 21 1462 759 <+64-21-1462-759>
>> [email protected]
>>
>>
>> On Sat, Jul 24, 2021 at 1:14 PM Richard Stallman <[email protected]> wrote:
>>
>>> [[[ To any NSA and FBI agents reading my email: please consider    ]]]
>>> [[[ whether defending the US Constitution against all enemies,     ]]]
>>> [[[ foreign or domestic, requires you to follow Snowden's example. ]]]
>>>
>>>   > > That's not what happens with these services: they don't _copy_ code
>>>   > > from other software (that won't work, because the probability of
>>> the
>>>   > > variables being called by other names is 100%, and thus such code,
>>> if
>>>   > > pasted into your program, will not compile).  What they do, they
>>>   > > extract ideas and algorithms from those other places, and express
>>> them
>>>   > > in terms of your variables and your data types.  So licenses are
>>> not
>>>   > > relevant here.
>>>
>>>   > According to online reviews chunks of code is copied even verbatim
>>> and
>>>   > people find from where. Even if modified, it still requires licensing
>>>   > compliance.
>>>
>>> From what I have read, it seems that the behavior of copilot runs on a
>>> spectrum from the first description to the second description.  I
>>> expect that in many cases, nothing copyrightable has been copied, but
>>> in some cases copilot does copy a substantial amount from a
>>> copyrighted work.
>>>
>>> --
>>> Dr Richard Stallman (https://stallman.org)
>>> Chief GNUisance of the GNU Project (https://gnu.org)
>>> Founder, Free Software Foundation (https://fsf.org)
>>> Internet Hall-of-Famer (https://internethalloffame.org)
>>>
>>>
>>> --

Shane Mulligan

How to contact me:
🇦🇺 00 61 421 641 250
🇳🇿 00 64 21 1462 759 <+64-21-1462-759>
[email protected]

Reply via email to