Dear Luke, dear All
Interesting essay Francis, and always appreciate Brian's thoughtful comments. I think the historical angle Brian is pointing towards is important as a way to push against the claims of AI models as somehow entirely new or revolutionary.

In particular, I want to push back against this idea that this is the last 'pure' cultural snapshot available to AI models, that future harvesting will be 'tainted' by automated content.

At no point did I allude to the 'pureness' of a cultural snapshot, as you suggest. Why should I? I was discussing this from a material perspective, where data for training diffusion models becomes the statistical material to inform these models. This data has never been 'pure'. I used the distinction of uncontaminated/contaminated to show the difference between a training process for machine learning which builds on an snapshot, that is still uncontaminated by the outputs of CLIP or GPT and one which includes generated text and images using this techique on a large scale.

It is obvious, but maybe I should have made it more clear, that the training data in itself is already far from pure. Honestly I'm a bit shocked, you would suggest I'd come up with a nostalgic argument about purity.

Francis' examples of hip hop and dnb culture, with sampling at their heart, already starts to point to the problems with this statement. Culture has always been a project of cutting and splicing, appropriating, transforming, and remaking existing material. It's funny that AI commentators like Gary Marcus talk about GPT-3 as the 'king of pastiche'. Pastiche is what culture does. Indeed, we have whole genres (the romance novel, the murder mystery, etc) that are about reproducing certain elements in slightly different permutations, over and over again.
Maybe it is no coincidence that I included exactly this example.
Unspoken in this claim of machines 'tainting' or 'corrupting' culture is the idea of authenticity.
I didn't claim 'tainting' or 'corrupting' culture, not even unspoken. Who am I to argue against the productive forces?
It really reminds me of the moral panic surrounding algorithmic news and platform-driven disinformation, where pundits lamented the shift from truth to 'post-truth.'  This is not to suggest that misinformation is not an issue, nor that veracity doesn't matter (i.e. Rohingya and Facebook). But the premise of some halcyon age of truth prior to the digital needs to get wrecked.
I agree. Only, I never equaled 'uncontaminated' to a "truth prior to the digital", I equaled it to a snapshot that doesn't contain material created by transformer models.
Yes, Large language models and other AI technologies do introduce new conditions, generating truth claims rapidly and at scale. But rather than hand-wringing about 'fake news,' it's more productive to see how they splice together several truth theories (coherence, consensus, social construction, etc) into new formations.

I was more interested in two points:

1.) Subversion: What I called in my original text the 'data space' (created through cultural snapshots as suggested by Eva Cetinic) is an already biased, largely uncurated information space where image data and language data are scaped and then mathemtically-statistically merged together. The focus point here is the sheer scale on which this happens. GPT-3 and CLIP are techniques that both build on massive datascraping (compared for instance to GANs) so that it is only possible for well funded organizations such as Open-AI or LAION to build these datasets. This dataspace could be spammed a) if you want to subvert it and b) if you'd want to advertise. The spam would need to be on a large scale in order to influence the next (contaminated) iteration of a cultural snapshot. In that sense only I used the un/contaminated distinction.

2). In response to Brian I evoked a scenario that builds on what we already experience when it comes to information spamming. We all know, that mis-information is a social and _not_ a machinic function. Maybe I should have made this more clear (I simply assumed it). I ignored Brians comment on the decline of culture, whatever this would mean, and could have been more precise in this regards. I don't assume culture declines. Beyond this, there have been discussions about deepfakes for instance and we saw that deepfakes are not needed at all to create mis-information, when one can just cut any video using standard video editing practices towards 'make-believe'. I wasn't 'hand-wringing' about fake news, in my comment to Brian, instead I was quoting Langlois with the concept of 'real fakes'. Further I'm suggesting that CLIP and GPT make it more easy to automate large scale spamming, making online communities uninhabitable or moderation more difficult. Maybe I'm overestimating the effect. We can already observe GPT-3 automated comments appearing on twitter or the ban of GPTChat posts on Stackoverflow (https://meta.stackoverflow.com/questions/421831/temporary-policy-chatgpt-is-banned), the latter already being a Berghain-no-photo-policy.

Finally, I'm interested in the question of bias and representation, and how a cultural snapshot, that builds on a biased dataset (and no, I'm not saying there are unbiased datasets at all), can further deepen these biases with each future interation, when these bias get statistically reproduced through 'AI' and the become basis for the next dataset.

best

Francis


nga mihi / best,
Luke


On Tue, 20 Dec 2022 at 22:20, Francis Hunger <francis.hun...@irmielin.org> wrote:

    Hi Brian,
    On Mon, Dec 19, 2022 at 3:55 AM Francis Hunger
    <francis.hun...@irmielin.org> wrote:

        While some may argue that generated text and images will save
        time and money for businesses, a data ecological view
        immediately recognizes a major problem: AI feeds into AI. To
        rephrase it: statistical computing feeds into statistical
        computing. In using these models and publishing the results
        online we are beginning to create a loop of prompts and
        results, with the results being fed into the next iteration
        of the cultural snapshots. That’s why I call the early
        cultural snapshots still uncontaminated, and I expect the
        next iterations of cultural snapshots will be contaminated.


    Francis, thanks for your work, it's always totally interesting.

    Your argumentation is impeccable and one can easily see how
    positive feedback loops will form around elements of AI-generated
    (or perhaps "recombined") images. I agree, this will become
    untenable, though I'd be interested in your ideas as to why. What
    kind of effects do you foresee, both on the level of the images
    themselves and their reception?

    Foresight is a difficult field, as most estimates can extrapolate
    maximum 7 year into the future and there are a lot of independent
    factors (such as e.g. OpenAI, the producer of CLIP could go
    bankrupt etc.).

    It's worth considering that similar loops have been in place for
    decades, in the area of market research, product design and
    advertising. Now, all of neoclassical economics is based on the
    concept of "consumer preferences," and discovering what consumers
    prefer is the official justification for market research; but
    it's clear that advertising has attempted, and in many cases
    succeeded, in shaping those preferences over generations. The
    preferences that people express today are, at least in part,
    artifacts of past advertising campaigns. Product design in the
    present reflects the influence of earlier products and associated
    advertising.

    That's an great and interesting argument. Because it plays into
    the cultural snapshot idea.

    Obviously Language wise, people already use translation tools,
    such as Deepl and translate Text from German to English and back
    to German in order to profit off the "clarity" and "orthographic
    correction" brought by the statistical analysis that feeds into
    the translator and seems to straighten the German text. We see the
    same stuff appearing for products like text editors and thus
    widely employed for cultural production. That's one example.
    Automated forum posts using GPT-3, for instance on Reddit are
    another, because we know that the CLIP Model also partly build on
    Reddit posts.

    Another example is images generated using diffusion models and
    prompts building on cultural snapshots and being used as _cheap_
    illustrations for editorial products, feeding off stock
    photography and to a certain extend replacing stock photography.
    This is more or less an economic motivation with cultural
    consequences. The question is what changes, when there is not
    sufficiently 'original' stock photography circulating, but the
    majority is syntheticly generated? Maybe others want to join in,
    to speculate about it.

    We could further look into 1980s HipHop or 1990s Drum'n Bass
    sample culture, which for instance took (and some argue: stole)
    one particular sound break, the Amen Break, from an obscure 1969
    Soul music record by The Winston Brothers and build a whole
    cultural genre from it. Cf.
    https://en.wikipedia.org/wiki/Amen_break Here the sample was
    refined over time, with generations of musicians cleaning the
    sample (compression, frequencies, deverbing, etc.) and providing
    many variations of it, then reusing it, because later generation
    did not build on the original sample, but on the published
    versions of it.

    We can maybe distinguish two modi operandi where a) "the cultural
    snapshot" is understood as an automated feedback loop, operating
    on a large scale, mainly through automated scraping and
    publication of the derivates of data, amplifying the already most
    visible representations of culture and b) "the cultural snapshot"
    is a feedback loop with many creative human interventions, be it
    through curatorial selection, prompt engineering or intended data
    manipulation.

    Blade Runner vividly demonstrated this cultural condition in the
    early 1980s, through the figure of the replicants with their
    implanted memories.
    I dont know if I get your point. I'd always say that Blade Runner
    is a cultural imaginary, one of the many phantasms about the
    machinisation of humans since at least 1900 if not earlier, and
    that's an entirely different discussion then. I would avoid this
    as an metaphor.
    The intensely targeted production of postmodern culture ensued,
    and has been carried on since then with the increasingly granular
    market research of surveillance capitalism, where the calculation
    of statistically probable behavior becomes a good deal more
    precise. The effect across the neoliberal period has been, not
    increasing standardization or authoritarian control, but instead,
    the rationalized proliferation of customizable products, whose
    patterns of use and modification, however divergent or "deviant"
    they may be, are then fed back into the design process. Not only
    the "quality of the image" seems to degrade in this process.
    Instead, culture in general seems to degrade, even though it also
    becomes more inclusive and more diverse at the same time.

    When looking for a plausible scenario regarding synthetic text and
    synthetic images, Steve Bannons “The real opposition is the media.
    And the way to deal with them is to flood the zone with shit.” is
    sadly a good candidate. This ties in with what Ganaele Langlois
    posits:

        „Therefore: communicative fascism posts that what is real is
        the opposite of social justice, and we now see the armies of
        ‚Social Injustice Warriors‘ as Sarah Sharma (2019) calls them,
        busy typing away at their keyboards to defend the rights to
        keep their fear of Others unchallenged and to protect their
        bigotry, misogyny, and racism from being debunked as inept
        constructions of themselves“ Langlois 2021:3

        „The first aspect of this new communicative fascism is related
        to what can be called ‚real fakes_ that is to say, the
        construction of a fictional and alternative reality where the
        paranoid position of fear and rage can find some validation …
        Real fakes are about what reality ought to be: they are
        virtual backgrounds on which fascists can find their validity
        and raising’être.“ Langlois 2021:3f

    So this is to be expected both for political or consumer marketing
    purposes.

    AI is poised to do a lot of things - but one of them is to
    further accelerate the continual remaking of generational
    preferences for the needs of capitalist marketing. Do you think
    that's right, Francis?

    That's one possible reading. I would insist, to not use an active
    verb with AI however, rephrasing your point towards "AI may be
    used for a lot of things". Better even replace 'AI' with the term
    'statistical computation'.

    Currently I would read 'AI' as a mixture of imaginations and
    phantasms about automation, of which some may become true – just
    in another way from what was expected or promoted. For certain,
    the inner logics of capital circulation command to deploy
    statistical computation to replace living, human labor. We already
    see how the job description of translators changes towards an
    human–statistical_computation entanglement and how the repetetive
    parts of the illustrator job, like coloring get automated away and
    put people out of jobs and it is plausible to expect the
    consolidation of jobs like photo editor, news editor, author with
    prompt-engineering. Since we are concentrating on the cultural
    sphere here, I'll limit the examples to this field. Human Labor in
    production, logistics, care labor would need their own thoughts.

    What other consequences do you see? And above all, what to do in
    the face of a seemingly inevitable trend?

    We are going to create separate data ecologies, which prohibit
    spamming the data space. These would be spaces, comparable to the
    no-photo-policy in clubs like Berghain or IFZ with a no-synthetics
    policy. While vast areas of the information space may be indeed
    flooded, these would be valuable zones of cultural exchange. (The
    answer would be much longer indeed, but we're not writing a book
    here).



    best, Brian

-- Researcher at Training The Archive, HMKV Dortmund

    Artistic Practicehttp://www.irmielin.org
    Ph.D. at Bauhaus University Weimarhttp://databasecultures.irmielin.org

    Daily Tweetshttps://twitter.com/databaseculture


    Peter and Irene Ludwig guest professorship at the Hungarian University of 
Fine Arts in Budapest 2022/23

    #  distributed via <nettime>: no commercial use without permission
    #  <nettime>  is a moderated mailing list for net criticism,
    #  collaborative text filtering and cultural politics of the nets
    #  more info: http://mx.kein.org/mailman/listinfo/nettime-l
    #  archive: http://www.nettime.org contact: nett...@kein.org
    #  @nettime_bot tweets mail w/ sender unless #ANON is in Subject:

--
Researcher at Training The Archive, HMKV Dortmund

Artistic Practicehttp://www.irmielin.org
Ph.D. at Bauhaus University Weimarhttp://databasecultures.irmielin.org

Daily Tweetshttps://twitter.com/databaseculture


Peter and Irene Ludwig guest professorship at the Hungarian University of Fine 
Arts in Budapest 2022/23
#  distributed via <nettime>: no commercial use without permission
#  <nettime>  is a moderated mailing list for net criticism,
#  collaborative text filtering and cultural politics of the nets
#  more info: http://mx.kein.org/mailman/listinfo/nettime-l
#  archive: http://www.nettime.org contact: nett...@kein.org
#  @nettime_bot tweets mail w/ sender unless #ANON is in Subject:

Reply via email to