None of this is surprising. I’m scared of criminals using it

From my iPad

Betty



> On Apr 2, 2023, at 6:52 PM, ghe2001 <ghe2...@pm.me> wrote:
> 
> -----BEGIN PGP SIGNED MESSAGE-----
> Hash: SHA256
> 
> This is from a few days ago -- all of them are doing it.  I'm assuming this 
> story wasn't written by some AI software.  This email wasn't.  Which is 
> exactly what it'd say if it was...
> 
> 
> By Shane Goldmacher
> 
> The Democratic Party has begun testing the use of artificial intelligence to 
> write first drafts of some fund-raising messages, appeals that often perform 
> better than those written entirely by human beings.
> 
> Fake A.I. images of Donald J. Trump getting arrested in New York spread 
> faster than they could be fact-checked last week.
> 
> And voice-cloning tools are producing vividly lifelike audio of President 
> Biden — and many others — saying things they did not actually say.
> 
> Artificial intelligence isn’t just coming soon to the 2024 campaign trail. 
> It’s already here.
> 
> The swift advance of A.I. promises to be as disruptive to the political 
> sphere as to broader society. Now any amateur with a laptop can manufacture 
> the kinds of convincing sounds and images that were once the domain of the 
> most sophisticated digital players. This democratization of disinformation is 
> blurring the boundaries between fact and fake at a moment when the acceptance 
> of universal truths — that Mr. Biden beat Mr. Trump in 2020, for example — is 
> already being strained.
> 
> And as synthetic media gets more believable, the question becomes: What 
> happens when people can no longer trust their own eyes and ears?
> 
> Inside campaigns, artificial intelligence is expected to soon help perform 
> mundane tasks that previously required fleets of interns. Republican and 
> Democratic engineers alike are racing to develop tools to harness A.I. to 
> make advertising more efficient, to engage in predictive analysis of public 
> behavior, to write more and more personalized copy and to discover new 
> patterns in mountains of voter data. The technology is evolving so fast that 
> most predict a profound impact, even if specific ways in which it will upend 
> the political system are more speculation than science.
> 
> “It’s an iPhone moment — that’s the only corollary that everybody will 
> appreciate,” said Dan Woods, the chief technology officer on Mr. Biden’s 2020 
> campaign. “It’s going to take pressure testing to figure out whether it’s 
> good or bad — and it’s probably both.”
> 
> OpenAI, whose ChatGPT chatbot ushered in the generative-text gold rush, has 
> already released a more advanced model. Google has announced plans to expand 
> A.I. offerings inside popular apps like Google Docs and Gmail, and is rolling 
> out its own chatbot. Microsoft has raced a version to market, too. A smaller 
> firm, ElevenLabs, has developed a text-to-audio tool that can mimic anyone’s 
> voice in minutes. Midjourney, a popular A.I. art generator, can conjure 
> hyper-realistic images with a few lines of text that are compelling enough to 
> win art contests.
> 
> “A.I. is about to make a significant change in the 2024 election because of 
> machine learning’s predictive ability,” said Brad Parscale, Mr. Trump’s first 
> 2020 campaign manager, who has since founded a digital firm that advertises 
> some A.I. capabilities.
> 
> Disinformation and “deepfakes” are the dominant fear. While forgeries are 
> nothing new to politics — a photoshopped image of John Kerry and Jane Fonda 
> was widely shared in 2004 — the ability to produce and share them has 
> accelerated, with viral A.I. images of Mr. Trump being restrained by the 
> police only the latest example. A fake image of Pope Francis in a white puffy 
> coat went viral in recent days, as well.
> 
> Many are particularly worried about local races, which receive far less 
> scrutiny. Ahead of the recent primary in the Chicago mayoral race, a fake 
> video briefly sprung up on a Twitter account called “Chicago Lakefront News” 
> that impersonated one candidate, Paul Vallas.
> 
> “Unfortunately, I think people are going to figure out how to use this for 
> evil faster than for improving civic life,” said Joe Rospars, who was chief 
> strategist on Senator Elizabeth Warren’s 2020 campaign and is now the chief 
> executive of a digital consultancy.
> 
> Those who work at the intersection of politics and technology return 
> repeatedly to the same historical hypothetical: If the infamous “Access 
> Hollywood” tape broke today — the one in which Mr. Trump is heard bragging 
> about assaulting women and getting away with it — would Mr. Trump acknowledge 
> it was him, as he did in 2016?
> 
> The nearly universal answer was no.
> 
> “I think about that example all the time,” said Matt Hodges, who was the 
> engineering director on Mr. Biden’s 2020 campaign and is now executive 
> director of Zinc Labs, which invests in Democratic technology. Republicans, 
> he said, “may not use ‘fake news’ anymore. It may be ‘Woke A.I.’”
> 
> For now, the frontline function of A.I. on campaigns is expected to be 
> writing first drafts of the unending email and text cash solicitations.
> 
> “Given the amount of rote, asinine verbiage that gets produced in politics, 
> people will put it to work,” said Luke Thompson, a Republican political 
> strategist.
> 
> As an experiment, The New York Times asked ChatGPT to produce a fund-raising 
> email for Mr. Trump. The app initially said, “I cannot take political sides 
> or promote any political agenda.” But then it immediately provided a template 
> of a potential Trump-like email.
> 
> The chatbot denied a request to make the message “angrier” but complied when 
> asked to “give it more edge,” to better reflect the often apocalyptic tone of 
> Mr. Trump’s pleas. “We need your help to send a message to the radical left 
> that we will not back down,” the revised A.I. message said. “Donate now and 
> help us make America great again.”
> 
> Among the prominent groups that have experimented with this tool is the 
> Democratic National Committee, according to three people briefed on the 
> efforts. In tests, the A.I.-generated content the D.N.C. has used has, as 
> often as not, performed as well or better than copy drafted entirely by 
> humans, in terms of generating engagement and donations.
> 
> Party officials still make edits to the A.I. drafts, the people familiar with 
> the efforts said, and no A.I. messages have yet been written under the name 
> of Mr. Biden or any other person, two people said. The D.N.C. declined to 
> comment.
> 
> Higher Ground Labs, a small venture capital firm that invests in political 
> technology for progressives, is currently working on a project, called 
> Quiller, to more systematically use A.I. to write, send and test the 
> effectiveness of fund-raising emails — all at once.
> 
> “A.I. has mostly been marketing gobbledygook for the last three cycles,” said 
> Betsy Hoover, a founding partner at Higher Ground Labs who was the director 
> of digital organizing for President Barack Obama’s 2012 campaign. “We are at 
> a moment now where there are things people can do that are actually helpful.”
> 
> Political operatives, several of whom were granted anonymity to discuss 
> potentially unsavory uses of artificial intelligence they are concerned about 
> or planning to deploy, raised a raft of possibilities.
> 
> Some feared bad actors could leverage A.I. chatbots to distract or waste a 
> campaign’s precious staff time by pretending to be potential voters. Others 
> floated producing deepfakes of their own candidate to generate personalized 
> videos — thanking supporters for their donations, for example. In India, one 
> candidate in 2020 produced a deepfake to disseminate a video of himself 
> speaking in different languages; the technology is far superior now.
> 
> Mr. Trump himself shared an A.I. image in recent days that appeared to show 
> him kneeling in prayer. He posted it on Truth Social, his social media site, 
> with no explanation.
> 
> One strategist predicted that the next generation of dirty tricks could be 
> direct-to-voter misinformation that skips social media sites entirely. What 
> if, this strategist said, an A.I. audio recording of a candidate was sent 
> straight to the voice mail of voters on the eve of an election?
> 
> Synthetic audio and video are already swirling online, much of it as parody.
> 
> On TikTok, there is an entire genre of videos featuring Mr. Biden, Mr. Obama 
> and Mr. Trump profanely bantering, with the A.I.-generated audio overlaid as 
> commentary during imaginary online video gaming sessions.
> 
> On “The Late Show,” Stephen Colbert recently used A.I. audio to have the Fox 
> News host Tucker Carlson “read” aloud his text messages slamming Mr. Trump. 
> Mr. Colbert labeled the audio as A.I. and the image on-screen showed a blend 
> of Mr. Carlson’s face and a Terminator cyborg for emphasis.
> 
> The right-wing provocateur Jack Posobiec pushed out a “deepfake” video last 
> month of Mr. Biden announcing a national draft because of the conflict in 
> Ukraine. It was quickly seen by millions.
> 
> “The videos we’ve seen in the last few weeks are really the canary in the 
> coal mine,” said Hany Farid, a professor of computer science at University of 
> California at Berkeley, who specializes in digital forensics. “We measure 
> advances now not in years but in months, and there are many months before the 
> election.”
> 
> Some A.I. tools were deployed in 2020. The Biden campaign created a program, 
> code-named Couch Potato, that linked facial recognition, voice-to-text and 
> other tools to automate the transcription of live events, including debates. 
> It replaced the work of a host of interns and aides, and was immediately 
> searchable through an internal portal.
> 
> The technology has improved so quickly, Mr. Woods said, that off-the-shelf 
> tools are “1,000 times better” than what had to be built from scratch four 
> years ago.
> 
> One looming question is what campaigns can and cannot do with OpenAI’s 
> powerful tools. One list of prohibited uses last fall lumped together 
> “political campaigns, adult content, spam, hateful content.”
> 
> Kim Malfacini, who helped create the OpenAI’s rules and is on the company’s 
> trust and safety team, said in an interview that “political campaigns can use 
> our tools for campaigning purposes. But it’s the scaled use that we are 
> trying to disallow here.” OpenAI revised its usage rules after being 
> contacted by The Times, specifying now that “generating high volumes of 
> campaign materials” is prohibited.
> 
> Tommy Vietor, a former spokesman for Mr. Obama, dabbled with the A.I. tool 
> from ElevenLabs to create a faux recording of Mr. Biden calling into the 
> popular “Pod Save America” podcast that Mr. Vietor co-hosts. He paid a few 
> dollars and uploaded real audio of Mr. Biden, and out came an audio likeness.
> 
> “The accuracy was just uncanny,” Mr. Vietor said in an interview.
> 
> The show labeled it clearly as A.I. But Mr. Vietor could not help noticing 
> that some online commenters nonetheless seemed confused. “I started playing 
> with the software thinking this is so much fun, this will be a great vehicle 
> for jokes,” he said, “and finished thinking, ‘Oh God, this is going to be a 
> big problem.’”
> 
> 
> 
> --
> Glenn English
> -----BEGIN PGP SIGNATURE-----
> Version: ProtonMail
> 
> wsBzBAEBCAAnBQJkKhT7CRDmyitW0WqrdxYhBMRG7aHlwXhPX67r9+bKK1bR
> aqt3AABsmwgA5CwRAP/ByfW9G25XWcVr6vQGU12vIoiKAHWB3C7QNy+d/B7Y
> 1pc706TAuvw5BPpugmiY26EghxL7pDhtCTQ6Af7Wo5qbXL1bzdJPrEs7znwM
> 1JzFYtZTR/Vjd6l4U7l86i/i5sdVMeACgrxyzDhG9IO1nyhNpWFi0nRLzpE4
> yGFzq4UmttKTRGoqx2dXG5vrO/TYbJHw/nHx8maZYtEtvvjTgLIgldEmbidi
> b/+xJDO3t1Kj5Xb1j0txqqoINXB8hsQNdSj+Sy3PlSBPA0vaV1v4IJrfnSjL
> ugDRs7FVqENohLfczKl6hozIF1dhv+VEHlaotgcQxrjoybbZ28q47Q==
> =QXTH
> -----END PGP SIGNATURE-----
> 

Reply via email to