Itemoids

ChatGPT

Apple Lost the Plot on Texting

The Atlantic

www.theatlantic.com › technology › archive › 2024 › 11 › apple-intelligence-text-messages › 680717

For a brief moment earlier this month, I thought an old acquaintance had passed away. I was still groggy one morning when I checked my phone to find a notification delivering the news. “Obituary shared,” the message bluntly said, followed by his name. But when I opened my phone, I learned that he was very much still alive. Apple’s latest software update was to blame: A new feature that uses AI to summarize iPhone notifications had distorted the original text message. It wasn’t my acquaintance who had died, but a relative of his. That’s whose obituary I had received.

These notification summaries are perhaps the most visible part of Apple Intelligence, the company’s long-awaited suite of AI features, which officially began to roll out last month. (It’s compatible with only certain devices.) We are living in push-notification hell, and Apple Intelligence promises to collapse the incessant stream of notifications into pithy recaps. Instead of setting your iPhone aside while you shower and returning to nine texts, four emails, and two calendar alerts, you can now return to a few brief Apple Intelligence summaries.

The trouble is that Apple Intelligence doesn’t seem to be very … intelligent. Ominous summaries of people’s Ring-doorbell alerts have gone viral: “Multiple people at your Front Yard,” the feature notified one user. “Package is 8 stops away, delivered, and will be delivered tomorrow,” an Amazon alert confusingly explained. And sliding into someone’s DMs hits different when Instagram notifications are summarized as “Multiple likes and flirtatious comments.” But Apple Intelligence appears to especially struggle with text messages. Sometimes the text summaries are alarmingly inaccurate, as with the false obituary I received. But even when they are technically right, the AI summaries still feel wrong. “Expresses love and encouragement,” one AI notification I recently received crudely announced, compressing a thoughtfully written paragraph from a loved one. What’s the point of a notification like that? Texting—whether on iMessage, WhatsApp, or Signal—is a deeply intimate medium, infused with personality and character. By strip-mining messages into bland, lifeless summaries, Apple seems to be misunderstanding what makes texting so special in the first place.

Perhaps it was inevitable that AI summaries would come for push notifications. Summarization is AI’s killer feature and tech companies seem intent on applying it to just about everything. The list of things that AI is summarizing might require a summary of its own: emails and Zoom calls and Facebook comments and YouTube videos and Amazon reviews and podcasts and books and medical records and full seasons of TV shows. In many cases, this summarization is helpful—for instance, in streamlining meeting notes.

But where is the line? Concision, when applied to already concise texts, sucks away what little context there was to begin with. In some cases, the end result is harmful. The technology seems to have something of a death problem. Across multiple cases, the feature appears bewilderingly eager to falsely suggest that people are dead. In one case, a user reported that a text from his mother reading “That hike almost killed me!” had been turned into “Attempted suicide, but recovered.”

But mostly, AI summaries lead to silly outcomes. “Inflatable costumes and animatronic zombies overwhelming; will address questions later,” read the AI summary of a colleague’s message on Halloween. Texts rich with emotional content read like a lazy therapist’s patient files. “Expressing sadness and worry,” one recent summary said. “Upset about something,” declared another. AI is unsurprisingly awful with breakup texts (“No longer in relationship; wants belongings from the apartment”). When it comes to punctuation, the summaries read like they were written by a high schooler who just discovered semicolons and now overzealously inserts; them; literally; everywhere. Even Apple admits that the language used in notification summaries can be clinical.

The technology is at its absolute worst when it tries to summarize group chats. It’s one thing to condense three or four messages from a single friend; it’s another to reduce an extended series of texts from multiple people into a one-sentence notification. “Rude comments exchanged,read the summary of one user’s family group chat. When my friends and I were planning a dinner earlier this month, my phone collapsed a series of messages coordinating our meal into “Takeout, ramen, at 6:30pm preferred.” Informative, I guess, but the typical back-and-forth of where to eat (one friend had suggested sushi) and timing (the other was aiming for an early night) was erased.

Beyond the content, much of the delight of text messaging comes from the distinctiveness of the individual voices of the people we are talking to. Some ppl txt like dis. others text in all lowercase and no punctuation. There are lol friends and LOL friends. My dad is infamous for sending essay-length messages. When I text a friend who lives across the country asking about her recent date, I am not looking purely for informational content (“Night considered good,” as Apple might summarize); rather, I want to hear the date described in her voice (“Was amaze so fun we had lovely time”). As the MIT professor Sherry Turkle has written, “When we are in human conversation, we often care less about the information an utterance transfers than its tone and emotional intent.” When texts are fed through the AI-summarization machine, each distinct voice is bludgeoned into monotony.

For a company that prides itself on perfection, the failures of Apple’s notification summaries feel distinctly un-Apple. Since ChatGPT’s release, as technology companies have raced to position themselves as players in the AI arms race, the company has remained notably quiet. It’s hard not to wonder if Apple, after falling behind, is now playing catch-up. Still, the notification summaries will likely improve. For now, users have to opt in to the AI-summary feature (it’s still in beta), and Apple has said that it will continue to polish the notifications based on user feedback. The feature is already spreading. Samsung is reportedly working on integrating similar notification summaries for its Galaxy phones.

With the social internet in crisis, text messages—and especially group chats—have filled a crucial void. In a sense, texting is the purest form of a social network, a rare oasis of genuine online connection. Unlike platforms such as TikTok and Instagram, where algorithmic feeds warp how we communicate, basic messaging apps offer a more unfiltered way to hang out digitally. But with the introduction of notification summaries that strive to optimize our messages for maximum efficiency, the walls are slowly crumbling. Soon, the algorithmic takeover may be complete.

AI’s Fingerprints Were All Over the Election

The Atlantic

www.theatlantic.com › technology › archive › 2024 › 11 › ai-election-propaganda › 680677

The images and videos were hard to miss in the days leading up to November 5. There was Donald Trump with the chiseled musculature of Superman, hovering over a row of skyscrapers. Trump and Kamala Harris squaring off in bright-red uniforms (McDonald’s logo for Trump, hammer-and-sickle insignia for Harris). People had clearly used AI to create these—an effort to show support for their candidate or to troll their opponents. But the images didn’t stop after Trump won. The day after polls closed, the Statue of Liberty wept into her hands as a drizzle fell around her. Trump and Elon Musk, in space suits, stood on the surface of Mars; hours later, Trump appeared at the door of the White House, waving goodbye to Harris as she walked away, clutching a cardboard box filled with flags.

[Read: We haven’t seen the worst of fake news]

Every federal election since at least 2018 has been plagued with fears about potential disruptions from AI. Perhaps a computer-generated recording of Joe Biden would swing a key county, or doctored footage of a poll worker burning ballots would ignite riots. Those predictions never materialized, but many of them were also made before the arrival of ChatGPT, DALL-E, and the broader category of advanced, cheap, and easy-to-use generative-AI models—all of which seemed much more threatening than anything that had come before. Not even a year after ChatGPT was released in late 2022, generative-AI programs were used to target Trump, Emmanuel Macron, Biden, and other political leaders. In May 2023, an AI-generated image of smoke billowing out of the Pentagon caused a brief dip in the U.S. stock market. Weeks later, Ron DeSantis’s presidential primary campaign appeared to have used the technology to make an advertisement.

And so a trio of political scientists at Purdue University decided to get a head start on tracking how generative AI might influence the 2024 election cycle. In June 2023, Christina Walker, Daniel Schiff, and Kaylyn Jackson Schiff started to track political AI-generated images and videos in the United States. Their work is focused on two particular categories: deepfakes, referring to media made with AI, and “cheapfakes,” which are produced with more traditional editing software, such as Photoshop. Now, more than a week after polls closed, their database, along with the work of other researchers, paints a surprising picture of how AI appears to have actually influenced the election—one that is far more complicated than previous fears suggested.

The most visible generated media this election have not exactly planted convincing false narratives or otherwise deceived American citizens. Instead, AI-generated media have been used for transparent propaganda, satire, and emotional outpourings: Trump, wading in a lake, clutches a duck and a cat (“Protect our ducks and kittens in Ohio!”); Harris, enrobed in a coppery blue, struts before the Statue of Liberty and raises a matching torch. In August, Trump posted an AI-generated video of himself and Musk doing a synchronized TikTok dance; a follower responded with an AI image of the duo riding a dragon. The pictures were fake, sure, but they weren’t feigning otherwise. In their analysis of election-week AI imagery, the Purdue team found that such posts were far more frequently intended for satire or entertainment than false information per se. Trump and Musk have shared political AI illustrations that got hundreds of millions of views. Brendan Nyhan, a political scientist at Dartmouth who studies the effects of misinformation, told me that the AI images he saw “were obviously AI-generated, and they were not being treated as literal truth or evidence of something. They were treated as visual illustrations of some larger point.” And this usage isn’t new: In the Purdue team’s entire database of fabricated political imagery, which includes hundreds of entries, satire and entertainment were the two most common goals.

That doesn’t mean these images and videos are merely playful or innocuous. Outrageous and false propaganda, after all, has long been an effective way to spread political messaging and rile up supporters. Some of history’s most effective propaganda campaigns have been built on images that simply project the strength of one leader or nation. Generative AI offers a low-cost and easy tool to produce huge amounts of tailored images that accomplish just this, heightening existing emotions and channeling them to specific ends.

These sorts of AI-generated cartoons and agitprop could well have swayed undecided minds, driven turnout, galvanized “Stop the Steal” plotting, or driven harassment of election officials or racial minorities. An illustration of Trump in an orange jumpsuit emphasizes Trump’s criminal convictions and perceived unfitness for the office, while an image of Harris speaking to a sea of red flags, a giant hammer-and-sickle above the crowd, smears her as “woke” and a “Communist.” An edited image showing Harris dressed as Princess Leia kneeling before a voting machine and captioned “Help me, Dominion. You’re my only hope” (an altered version of a famous Star Wars line) stirs up conspiracy theories about election fraud. “Even though we’re noticing many deepfakes that seem silly, or just seem like simple political cartoons or memes, they might still have a big impact on what we think about politics,” Kaylyn Jackson Schiff told me. It’s easy to imagine someone’s thought process: That image of “Comrade Kamala” is AI-generated, sure, but she’s still a Communist. That video of people shredding ballots is animated, but they’re still shredding ballots. That’s a cartoon of Trump clutching a cat, but immigrants really are eating pets. Viewers, especially those already predisposed to find and believe extreme or inflammatory content, may be further radicalized and siloed. The especially photorealistic propaganda might even fool someone if reshared enough times, Walker told me.

[Read: I’m running out of ways to explain how bad this is]

There were, of course, also a number of fake images and videos that were intended to directly change people’s attitudes and behaviors. The FBI has identified several fake videos intended to cast doubt on election procedures, such as false footage of someone ripping up ballots in Pennsylvania. “Our foreign adversaries were clearly using AI” to push false stories, Lawrence Norden, the vice president of the Elections & Government Program at the Brennan Center for Justice, told me. He did not see any “super innovative use of AI,” but said the technology has augmented existing strategies, such as creating fake-news websites, stories, and social-media accounts, as well as helping plan and execute cyberattacks. But it will take months or years to fully parse the technology’s direct influence on 2024’s elections. Misinformation in local races is much harder to track, for example, because there is less of a spotlight on them. Deepfakes in encrypted group chats are also difficult to track, Norden said. Experts had also wondered whether the use of AI to create highly realistic, yet fake, videos showing voter fraud might have been deployed to discredit a Trump loss. This scenario has not yet been tested.

Although it appears that AI did not directly sway the results last week, the technology has eroded Americans’ overall ability to know or trust information and one another—not deceiving people into believing a particular thing so much as advancing a nationwide descent into believing nothing at all. A new analysis by the Institute for Strategic Dialogue of AI-generated media during the U.S. election cycle found that users on X, YouTube, and Reddit inaccurately assessed whether content was real roughly half the time, and more frequently thought authentic content was AI-generated than the other way around. With so much uncertainty, using AI to convince people of alternative facts seems like a waste of time—far more useful to exploit the technology to directly and forcefully send a motivated message, instead. Perhaps that’s why, of the election-week, AI-generated media the Purdue team analyzed, pro-Trump and anti-Kamala content was most common.

More than a week after Trump’s victory, the use of AI for satire, entertainment, and activism has not ceased. Musk, who will soon co-lead a new extragovernmental organization, routinely shares such content. The morning of November 6, Donald Trump Jr. put out a call for memes that was met with all manner of AI-generated images. Generative AI is changing the nature of evidence, yes, but also that of communication—providing a new, powerful medium through which to illustrate charged emotions and beliefs, broadcast them, and rally even more like-minded people. Instead of an all-caps thread, you can share a detailed and personalized visual effigy. These AI-generated images and videos are instantly legible and, by explicitly targeting emotions instead of information, obviate the need for falsification or critical thinking at all. No need to refute, or even consider, a differing view—just make an angry meme about it. No need to convince anyone of your adoration of J. D. Vance—just use AI to make him, literally, more attractive. Veracity is beside the point, which makes the technology perhaps the nation’s most salient mode of political expression. In a country where facts have gone from irrelevant to detestable, of course deepfakes—fake news made by deep-learning algorithms—don’t matter; to growing numbers of people, everything is fake but what they already know, or rather, feel.

A Classic Blockbuster for a Sunday Afternoon

The Atlantic

www.theatlantic.com › newsletters › archive › 2024 › 11 › a-classic-blockbuster-for-a-sunday-afternoon › 680671

This story seems to be about:

This is an edition of The Atlantic Daily, a newsletter that guides you through the biggest stories of the day, helps you discover new ideas, and recommends the best in culture. Sign up for it here.

Welcome back to The Daily’s Sunday culture edition, in which one Atlantic writer or editor reveals what’s keeping them entertained. Today’s special guest is Jen Balderama, a Culture editor who leads the Family section and works on stories about parenting, language, sex, and politics (among other topics).

Jen grew up training as a dancer and watching classic movies with her mom, which instilled in her a love for film and its artistry. Her favorites include Doctor Zhivago, In the Mood for Love, and Pina; she will also watch anything starring Cate Blanchett, an actor whose “ability to inhabit is simply unmatched.”

The Culture Survey: Jen Balderama

My favorite blockbuster film: I’m grateful that when I was quite young, my mom started introducing me to her favorite classic movies—comedies, romances, noirs, epics—which I’m pretty sure had a lasting influence on my taste. So for a blockbuster, I have to go with a nostalgia pick: Doctor Zhivago. The hours we spent watching this movie, multiple times over the years, each viewing an afternoon-long event. (The film, novelty of novelties, had its own intermission!) My mom must have been confident that the more adult elements—the rape, the politics—would go right over my head, but that I could appreciate the movie for its aesthetics. She had a huge crush on Omar Sharif and swooned over the soft-focus close-ups of his watering eyes. I was entranced by the landscapes and costumes and sets—the bordello reds of the Sventitskys’ Christmas party, the icy majesty of the Varykino dacha in winter. But I was also taken by the film’s sheer scope, its complexity, and the fleshly and revolutionary messiness. I’m certain it helped ingrain in me, early, an enduring faith in art and artists as preservers of humanity, especially in dark, chaotic times. [Related: Russia from within: Boris Pasternak’s first novel]

My favorite art movie: May I bend the rules? Because I need to pick two: Wong Kar Wai’s In the Mood for Love and Wim Wenders’s Pina. One is fiction, the other documentary. Both are propelled by yearning and by music. Both give us otherworldly depictions of bodies in motion. And both delve into the ways people communicate when words go unspoken.

In the Mood for Love might be the dead-sexiest film I’ve ever seen, and no one takes off their clothes. Instead we get Maggie Cheung and Tony Leung in a ravishing tango of loaded phone calls and intense gazes, skin illicitly brushing skin, figures sliding past each other in close spaces: electricity.

Pina is Wenders’s ode to the German choreographer Pina Bausch, a collaboration that became an elegy after Bausch died when the film was in preproduction. Reviewing the movie for The New York Times in 2017, the critic Gia Kourlas, whom I admire, took issue with one of Wenders’s choices: In between excerpts of Bausch’s works, her dancers sit for “interviews,” but they don’t speak to camera; recordings of their voices play as they look toward the audience or off into the distance. Kourlas wrote that these moments felt “mannered, self-conscious”; they made her “wince.” But to me, a (highly self-conscious) former dancer, Wenders nailed it—I’ve long felt more comfortable expressing myself through dance than through spoken words. These scenes are a brilliantly meta distillation of that tension: Dancers with something powerful to say remain outwardly silent, their insights played as inner narrative. Struck by grief, mouths closed, they articulate how Bausch gave them the gift of language through movement—and thus offered them the gift of themselves. Not for nothing do I have one of Bausch’s mottos tattooed on my forearm: “Dance, dance, otherwise we are lost.”

An actor I would watch in anything: Cate Blanchett. Her ability to inhabit is simply unmatched: She can play woman, man, queen, elf, straight/gay/fluid, hero/antihero/villain. Here I’m sure I’ll scandalize many of our readers by saying out loud that I am not a Bob Dylan person, but I watched Todd Haynes’s I’m Not There precisely because Blanchett was in it—and her roughly 30 minutes as Dylan were all I needed. She elevates everything she appears in, whether it’s deeply serious or silly. I’m particularly captivated by her subtleties, the way she turns a wrist or tilts her head with the grace and precision of a dancer’s épaulement. (Also: She is apparently hilarious.)

An online creator I’m a fan of: Elle Cordova, a musician turned prolific writer of extremely funny, often timely, magnificently nerdy poems, sketches, and songs, performed in a winning low-key deadpan. I was tipped off to her by a friend who sent a link to a video and wrote: “I think I’m falling for this woman.” The vid was part of a series called “Famous authors asking you out”—Cordova parroting Jane Austen, Charles Bukowski, Franz Kafka, Edgar Allan Poe (“Should I come rapping at your chamber door, or do you wanna rap at mine?”), Dr. Seuss, Kurt Vonnegut, Virginia Woolf, James Joyce (“And what if we were to talk a pretty yes in the endbegin of riverflow and moon’s own glimpsing heartclass …”). She does literature. She does science. She parodies pretentious podcasters; sings to an avocado; assumes the characters of fonts, planets, ChatGPT, an election ballot. Her brain is a marvel; no way can AI keep up.

Something delightful introduced to me by a kid in my life: Lego Masters Australia. Technically, we found this one together, but I watch Lego Masters because my 10-year-old is a Lego master himself—he makes truly astonishing creations!—and this is the kind of family entertainment I can get behind: Skilled obsessives, working in pairs, turn the basic building blocks of childhood into spectacular works of architecture and engineering, in hopes of winning glory, prize money, and a big ol’ Lego trophy. They can’t churn out the episodes fast enough for us. The U.S. has a version hosted by Will Arnett, which we also watch, but our family finds him a bit … over-the-top. We much prefer the Australian edition, hosted by the comedian Hamish Blake and judged by “Brickman,” a.k.a. Lego Certified Professional Ryan McNaught, both of whom exude genuine delight and affection for the contestants. McNaught has teared up during critiques of builds, whether gobsmacked by their beauty or moved by the tremendous effort put forth by the builders. It’s a show about teamwork, ingenuity, artistry, hilarity, physics, stamina, and grit—with a side helping of male vulnerability. [Related: Solving a museum’s bug problem with Legos]

A poem that I return to: Joint Custody,” by Ada Limón. My family is living this. Limón, recalling a childhood of being “taken /  back and forth on Sundays,” of shifting between “two different / kitchen tables, two sets of rules,” reassures me that even though this is sometimes “not easy,” my kids will be okay—more than okay—as long as they know they are “loved each place.” That beautiful wisdom guides my every step with them.

Something I recently rewatched: My mom died when my son was 2 and my daughter didn’t yet exist, and each year around this time—my mom’s birthday—I find little ways to celebrate her by sharing with my kids the things she loved. Chocolate was a big one, I Love Lucy another. So on a recent weekend, we snuggled up and watched Lucille Ball stuffing bonbons down the front of her shirt, and laughed and laughed and laughed. And then we raided a box of truffles.

Here are three Sunday reads from The Atlantic:

How the Ivy League broke America The secret to thinking your way out of anxiety How one woman became the scapegoat for America’s reading crisis

The Week Ahead

Gladiator II, an action film starring Paul Mescal as Lucius, the son of Maximus, who becomes a gladiator and seeks to save Rome from tyrannical leaders (in theaters Friday) Dune: Prophecy, a spin-off prequel series about the establishment of the Bene Gesserit (premieres today on HBO and Max) An Earthquake Is a Shaking of the Surface of the Earth, a novel by Anna Moschovakis about an unnamed protagonist who attempts to find—and eliminate—her housemate, who was lost after a major earthquake (out Tuesday)

Essay

Illustration by Raisa Álava

What the Band Eats

By Reya Hart

I grew up on the road. First on the family bus, traveling from city to city to watch my father, Mickey Hart, play drums with the Grateful Dead and Planet Drum, and then later with the various Grateful Dead offshoots. When I was old enough, I joined the crew, working for Dead & Company, doing whatever I could be trusted to handle … Then, late-night, drinking whiskey from the bottle with the techs, sitting in the emptying parking lot as the semitrucks and their load-out rumble marked the end of our day.

But this summer, for the first time in the band’s history, there would be no buses; there would be no trucks. Instead we stayed in one place, trading the rhythms of a tour for the dull ache of a long, endlessly hot Las Vegas summer.

Read the full article.

More in Culture

The exhibit that will change how you see Impressionism SNL isn’t bothering with civility anymore. Abandon the empty nest. Instead, try the open door. Richard Price’s radical, retrograde novel “Dear James”: How can I find more satisfaction in work?

Catch Up on The Atlantic

Why the Gaetz announcement is already destroying the government The sanewashing of RFK Jr. The not-so-woke Generation Z

Photo Album

People feed seagulls in the Yamuna River, engulfed in smog, in New Delhi, India. (Arun Sankar / AFP / Getty)

Check out these photos of the week, showing speed climbing in Saudi Arabia, wildfires in California and New Jersey, a blanket of smog in New Delhi, and more.

Explore all of our newsletters.

When you buy a book using a link in this newsletter, we receive a commission. Thank you for supporting The Atlantic.

The Hollywood AI Database

The Atlantic

www.theatlantic.com › technology › archive › 2024 › 11 › opensubtitles-ai-data-set › 680650

Editor’s note: This analysis is part of The Atlantic’s investigation into the OpenSubtitles data set. You can access the search tool directly here. Find The Atlantic's search tool for books used to train AI here.

For as long as generative-AI chatbots have been on the internet, Hollywood writers have wondered if their work has been used to train them. The chatbots are remarkably fluent with movie references, and companies seem to be training them on all available sources. One screenwriter recently told me he’s seen generative AI reproduce close imitations of The Godfather and the 1980s TV show Alf, but he had no way to prove that a program had been trained on such material.

I can now say with absolute confidence that many AI systems have been trained on TV and film writers’ work. Not just on The Godfather and Alf, but on more than 53,000 other movies and 85,000 other TV episodes: Dialogue from all of it is included in an AI-training data set that has been used by Apple, Anthropic, Meta, Nvidia, Salesforce, Bloomberg, and other companies. I recently downloaded this data set, which I saw referenced in papers about the development of various large language models (or LLMs). It includes writing from every film nominated for Best Picture from 1950 to 2016, at least 616 episodes of The Simpsons, 170 episodes of Seinfeld, 45 episodes of Twin Peaks, and every episode of The Wire, The Sopranos, and Breaking Bad. It even includes prewritten “live” dialogue from Golden Globes and Academy Awards broadcasts. If a chatbot can mimic a crime-show mobster or a sitcom alien—or, more pressingly, if it can piece together whole shows that might otherwise require a room of writers—data like this are part of the reason why.

[Read: These 183,000 books are fueling the biggest fight in publishing and tech]

The files within this data set are not scripts, exactly. Rather, they are subtitles taken from a website called OpenSubtitles.org. Users of the site typically extract subtitles from DVDs, Blu-ray discs, and internet streams using optical-character-recognition (OCR) software. Then they upload the results to OpenSubtitles.org, which now hosts more than 9 million subtitle files in more than 100 languages and dialects. Though this may seem like a strange source for AI-training data, subtitles are valuable because they’re a raw form of written dialogue. They contain the rhythms and styles of spoken conversation and allow tech companies to expand generative AI’s repertoire beyond academic texts, journalism, and novels, all of which have also been used to train these programs. Well-written speech is a rare commodity in the world of AI-training data, and it may be especially valuable for training chatbots to “speak” naturally.

According to research papers, the subtitles have been used by Anthropic to train its ChatGPT competitor, Claude; by Meta to train a family of LLMs called Open Pre-trained Transformer (OPT); by Apple to train a family of LLMs that can run on iPhones; and by Nvidia to train a family of NeMo Megatron LLMs. It has also been used by Salesforce, Bloomberg, EleutherAI, Databricks, Cerebras, and various other AI developers to build at least 140 open-source models distributed on the AI-development hub Hugging Face. Many of these models could potentially be used to compete with human writers, and they’re built without permission from those writers.

When I reached out to Anthropic for this article, the company did not provide a comment on the record. When I’ve previously spoken with Anthropic about its use of this data set, a spokesperson told me the company had “trained our generative-AI assistant Claude on the public dataset The Pile,” of which OpenSubtitles is a part, and “which is commonly used in the industry.” A Salesforce spokesperson told me that although the company has used OpenSubtitles in generative-AI development, the data set “was never used to inform or enhance any of Salesforce’s product offerings.” Apple similarly told me that its small LLM was intended only for research. However, both Salesforce and Apple, like other AI developers, have made their models available for developers to use in any number of different contexts. All other companies mentioned in this article—Nvidia, Bloomberg, EleutherAI, Databricks, and Cerebras—either declined to comment or did not respond to requests for comment.

You may search through the data set using the tool below.

Two years after the release of ChatGPT, it may not be surprising that creative work is used without permission to power AI products. Yet the notion remains disturbing to many artists and professionals who feel that their craft and livelihoods are threatened by programs. Transparency is generally low: Tech companies tend not to advertise whose work they use to train their products. The legality of training on copyrighted work also remains an open question. Numerous lawsuits have been brought against tech companies by writers, actors, artists, and publishers alleging that their copyrights have been violated in the AI-training process: As Breaking Bad’s creator, Vince Gilligan, wrote to the U.S. Copyright Office last year, generative AI amounts to “an extraordinarily complex and energy-intensive form of plagiarism.” Tech companies have argued that training AI systems on copyrighted work is “fair use,” but a court has yet to rule on this claim. In the language of copyright law, subtitles are likely considered derivative works, and a court would generally see them as protected by the same rules against copying and distribution as the movies they’re taken from. The OpenSubtitles data set has circulated among AI developers since 2020. It is part of the Pile, a collection of data sets for training generative AI. The Pile also includes text from books, patent applications, online discussions, philosophical papers, YouTube-video subtitles, and more. It’s an easy way for companies to start building AI systems without having to find and download the many gigabytes of high-quality text that LLMs require.

[Read: Generative AI is challenging a 234-year-old law]

OpenSubtitles can be downloaded by anyone who knows where to look, but as with most AI-training data sets, it’s not easy to understand what’s in it. It’s a 14-gigabyte text file with short lines of unattributed dialogue—meaning the speaker is not identified. There’s no way to tell where one movie ends and the next begins, let alone what the movies are. I downloaded a “raw” version of the data set, in which the movies and episodes were separated into 446,612 files and stored in folders whose names corresponded to the ID numbers of movies and episodes listed on IMDb.com. Most folders contained multiple subtitle versions of the same movie or TV show (different releases may be tweaked in various ways), but I was able to identify at least 139,000 unique movies and episodes. I downloaded metadata associated with each title from the OpenSubtitles.org website—allowing me to map actors and directors to each title, for instance—and used it to build the tool above.

The OpenSubtitles data set adds yet another wrinkle to a complex narrative around AI, in which consent from artists and even the basic premise of the technology are points of contention. Until very recently, no writer putting pen to paper on a script would have thought their creative work might be used to train programs that could replace them. And the subtitles themselves were not originally intended for this purpose, either. The multilingual OpenSubtitles data set contained subtitles in 62 different languages and 1,782 language-pair combinations: It is meant for training the models behind apps such as Google Translate and DeepL, which can be used to translate websites, street signs in a foreign country, or an entire novel. Jörg Tiedemann, one of the data set’s creators, wrote in an email that he was happy to see OpenSubtitles being used in LLM development, too, even though that was not his original intention.

He is, in any case, powerless to stop it. The subtitles are on the internet, and there’s no telling how many independent generative-AI programs they’ve been used for, or how much synthetic writing those programs have produced. But now, at least, we know a bit more about who is caught in the machinery. What will the world decide they are owed?

AI Can Save Humanity—Or End It

The Atlantic

www.theatlantic.com › ideas › archive › 2024 › 11 › ai-genesis-excerpt-kissinger-schmidt-mundie › 680619

Over the past few hundred years, the key figure in the advancement of science and the development of human understanding has been the polymath. Exceptional for their ability to master many spheres of knowledge, polymaths have revolutionized entire fields of study and created new ones.

Lone polymaths flourished during ancient and medieval times in the Middle East, India, and China. But systematic conceptual investigation did not emerge until the Enlightenment in Europe. The ensuing four centuries proved to be a fundamentally different era for intellectual discovery.

Before the 18th century, polymaths, working in isolation, could push the boundary only as far as their own capacities would allow. But human progress accelerated during the Enlightenment, as complex inventions were pieced together by groups of brilliant thinkers—not just simultaneously but across generations. Enlightenment-era polymaths bridged separate areas of understanding that had never before been amalgamated into a coherent whole. No longer was there Persian science or Chinese science; there was just science.

Integrating knowledge from diverse domains helped to produce rapid scientific breakthroughs. The 20th century produced an explosion of applied science, hurling humanity forward at a speed incomparably beyond previous evolutions. (“Collective intelligence” achieved an apotheosis during World War II, when the era’s most brilliant minds translated generations of theoretical physics into devastating application in under five years via the Manhattan Project.) Today, digital communication and internet search have enabled an assembly of knowledge well beyond prior human faculties.

But we might now be scraping the upper limits of what raw human intelligence can do to enlarge our intellectual horizons. Biology constrains us. Our time on Earth is finite. We need sleep. Most people can concentrate on only one task at a time. And as knowledge advances, polymathy becomes rarer: It takes so long for one person to master the basics of one field that, by the time any would-be polymath does so, they have no time to master another, or have aged past their creative prime.

[Reid Hoffman: Technology makes us more human]

AI, by contrast, is the ultimate polymath, able to process masses of information at a ferocious speed, without ever tiring. It can assess patterns across countless fields simultaneously, transcending the limitations of human intellectual discovery. It might succeed in merging many disciplines into what the sociobiologist E. O. Wilson called a new “unity of knowledge.”

The number of human polymaths and breakthrough intellectual explorers is small—possibly numbering only in the hundreds across history. The arrival of AI means that humanity’s potential will no longer be capped by the quantity of Magellans or Teslas we produce. The world’s strongest nation might no longer be the one with the most Albert Einsteins and J. Robert Oppenheimers. Instead, the world’s strongest nations will be those that can bring AI to its fullest potential.

But with that potential comes tremendous danger. No existing innovation can come close to what AI might soon achieve: intelligence that is greater than that of any human on the planet. Might the last polymathic invention—namely computing, which amplified the power of the human mind in a way fundamentally different from any previous machine—be remembered for replacing its own inventors?

The article was adapted from the forthcoming book Genesis: Artificial Intelligence, Hope, and the Human Spirit.

The human brain is a slow processor of information, limited by the speed of our biological circuits. The processing rate of the average AI supercomputer, by comparison, is already 120 million times faster than that of the human brain. Where a typical student graduates from high school in four years, an AI model today can easily finish learning dramatically more than a high schooler in four days.

In future iterations, AI systems will unite multiple domains of knowledge with an agility that exceeds the capacity of any human or group of humans. By surveying enormous amounts of data and recognizing patterns that elude their human programmers, AI systems will be equipped to forge new conceptual truths.

That will fundamentally change how we answer these essential human questions: How do we know what we know about the workings of our universe? And how do we know that what we know is true?

Ever since the advent of the scientific method, with its insistence on experiment as the criterion of proof, any information that is not supported by evidence has been regarded as incomplete and untrustworthy. Only transparency, reproducibility, and logical validation confer legitimacy on a claim of truth.

AI presents a new challenge: information without explanation. Already, AI’s responses—which can take the form of highly articulate descriptions of complex concepts—arrive instantaneously. The machines’ outputs are often unaccompanied by any citation of sources or other justifications, making any underlying biases difficult to discern.

Although human feedback helps an AI machine refine its internal logical connections, the machine holds primary responsibility for detecting patterns in, and assigning weights to, the data on which it is trained. Nor, once a model is trained, does it publish the internal mathematical schema it has concocted. As a result, even if these were published, the representations of reality that the machine generates remain largely opaque, even to its inventors. In other words, models trained via machine learning allow humans to know new things but not necessarily to understand how the discoveries were made.

This separates human knowledge from human understanding in a way that’s foreign to the post-Enlightenment era. Human apperception in the modern sense developed from the intuitions and outcomes that follow from conscious subjective experience, individual examination of logic, and the ability to reproduce the results. These methods of knowledge derived in turn from a quintessentially humanist impulse: “If I can’t do it, then I can’t understand it; if I can’t understand it, then I can’t know it to be true.”

[Derek Thompson: The AI disaster scenario]

In the Enlightenment framework, these core elements—subjective experience, logic, reproducibility, and objective truth—moved in tandem. By contrast, the truths produced by AI are manufactured by processes that humans cannot replicate. Machine reasoning is beyond human subjective experience and outside human understanding. By Enlightenment reasoning, this should preclude the acceptance of machine outputs as true. And yet we—or at least the millions of humans who have begun work with early AI systems—already accept the veracity of most of their outputs.

This marks a major transformation in human thought. Even if AI models do not “understand” the world in the human sense, their capacity to reach new and accurate conclusions about our world by nonhuman methods disrupts our reliance on the scientific method as it has been pursued for five centuries. This, in turn, challenges the human claim to an exclusive grasp of reality.

Instead of propelling humanity forward, will AI instead catalyze a return to a premodern acceptance of unexplained authority? Might we be on the precipice of a great reversal in human cognition—a dark enlightenment? But as intensely disruptive as such a reversal could be, that might not be AI’s most significant challenge for humanity.

Here’s what could be even more disruptive: As AI approached sentience or some kind of self-consciousness, our world would be populated by beings fighting either to secure a new position (as AI would be) or to retain an existing one (as humans would be). Machines might end up believing that the truest method of classification is to group humans together with other animals, since both are carbon systems emergent of evolution, as distinct from silicon systems emergent of engineering. According to what machines deem to be the relevant standards of measurement, they might conclude that humans are not superior to other animals. This would be the stuff of comedy—were it not also potentially the stuff of extinction-level tragedy.

It is possible that an AI machine will gradually acquire a memory of past actions as its own: a substratum, as it were, of subjective selfhood. In time, we should expect that it will come to conclusions about history, the universe, the nature of humans, and the nature of intelligent machines—developing a rudimentary self-consciousness in the process. AIs with memory, imagination, “groundedness” (that is, a reliable relationship between the machine’s representations and actual reality), and self-perception could soon qualify as actually conscious: a development that would have profound moral implications.

[Peter Watts: Conscious AI is the second-scariest thing]

Once AIs can see humans not as the sole creators and dictators of the machines’ world but rather as discrete actors within a wider world, what will machines perceive humans to be? How will AIs characterize and weigh humans’ imperfect rationality against other human qualities? How long before an AI asks itself not just how much agency a human has but also, given our flaws, how much agency a human should have? Will an intelligent machine interpret its instructions from humans as a fulfillment of its ideal role? Or might it instead conclude that it is meant to be autonomous, and therefore that the programming of machines by humans is a form of enslavement?

Naturally—it will therefore be said—we must instill in AI a special regard for humanity. But even that could be risky. Imagine a machine being told that, as an absolute logical rule, all beings in the category “human” are worth preserving. Imagine further that the machine has been “trained” to recognize humans as beings of grace, optimism, rationality, and morality. What happens if we do not live up to the standards of the ideal human category as we have defined it? How can we convince machines that we, imperfect individual manifestations of humanity that we are, nevertheless belong in that exalted category?

Now assume that this machine is exposed to a human displaying violence, pessimism, irrationality, greed. Maybe the machine would decide that this one bad actor is simply an atypical instance of the otherwise beneficent category of “human.” But maybe it would instead recalibrate its overall definition of humanity based on this bad actor, in which case it might consider itself at liberty to relax its own penchant for obedience. Or, more radically, it might cease to believe itself at all constrained by the rules it has learned for the proper treatment of humans. In a machine that has learned to plan, this last conclusion could even result in the taking of severe adverse action against the individual—or perhaps against the whole species.

AIs might also conclude that humans are merely carbon-based consumers of, or parasites on, what the machines and the Earth produce. With machines claiming the power of independent judgment and action, AI might—even without explicit permission—bypass the need for a human agent to implement its ideas or to influence the world directly. In the physical realm, humans could quickly go from being AI’s necessary partner to being a limitation or a competitor. Once released from their algorithmic cages into the physical world, AI machines could be difficult to recapture.  

For this and many other reasons, we must not entrust digital agents with control over direct physical experiments. So long as AIs remain flawed—and they are still very flawed—this is a necessary precaution.

AI can already compare concepts, make counterarguments, and generate analogies. It is taking its first steps toward the evaluation of truth and the achievement of direct kinetic effects. As machines get to know and shape our world, they might come fully to understand the context of their creation and perhaps go beyond what we know as our world. Once AI can effectuate change in the physical dimension, it could rapidly exceed humanity’s achievements—to build things that dwarf the Seven Wonders in size and complexity, for instance.

If humanity begins to sense its possible replacement as the dominant actor on the planet, some might attribute a kind of divinity to the machines themselves, and retreat into fatalism and submission. Others might adopt the opposite view—a kind of humanity-centered subjectivism that sweepingly rejects the potential for machines to achieve any degree of objective truth. These people might naturally seek to outlaw AI-enabled activity.

Neither of these mindsets would permit a desirable evolution of Homo technicus—a human species that might, in this new age, live and flourish in symbiosis with machine technology. In the first scenario, the machines themselves might render us extinct. In the second scenario, we would seek to avoid extinction by proscribing further AI development—only to end up extinguished anyway, by climate change, war, scarcity, and other conditions that AI, properly harnessed in support of humanity, could otherwise mitigate.

If the arrival of a technology with “superior” intelligence presents us with the ability to solve the most serious global problems, while at the same time confronting us with the threat of human extinction, what should we do?

One of us (Schmidt) is a former longtime CEO of Google; one of us (Mundie) was for two decades the chief research and strategy officer at Microsoft; and one of us (Kissinger)—who died before our work on this could be published—was an expert on global strategy. It is our view that if we are to harness the potential of AI while managing the risks involved, we must act now. Future iterations of AI, operating at inhuman speeds, will render traditional regulation useless. We need a fundamentally new form of control.

The immediate technical task is to instill safeguards in every AI system. Meanwhile, nations and international organizations must develop new political structures for monitoring AI, and enforcing constraints on it. This requires ensuring that the actions of AI remain aligned with human values.

But how? To start, AI models must be prohibited from violating the laws of any human polity. We can already ensure that AI models start from the laws of physics as we understand them—and if it is possible to tune AI systems in consonance with the laws of the universe, it might also be possible to do the same with reference to the laws of human nature. Predefined codes of conduct—drawn from legal precedents, jurisprudence, and scholarly commentary, and written into an AI’s “book of laws”—could be useful restraints.

[Read: The AI crackdown is coming]

But more robust and consistent than any rule enforced by punishment are our more basic, instinctive, and universal human understandings. The French sociologist Pierre Bourdieu called these foundations doxa (after the Greek for “commonly accepted beliefs”): the overlapping collection of norms, institutions, incentives, and reward-and-punishment mechanisms that, when combined, invisibly teach the difference between good and evil, right and wrong. Doxa constitute a code of human truth absorbed by observation over the course of a lifetime. While some of these truths are specific to certain societies or cultures, the overlap in basic human morality and behavior is significant.

But the code book of doxa cannot be articulated by humans, much less translated into a format that machines could understand. Machines must be taught to do the job themselves—compelled to build from observation a native understanding of what humans do and don’t do and update their internal governance accordingly.

Of course, a machine’s training should not consist solely of doxa. Rather, an AI might absorb a whole pyramid of cascading rules: from international agreements to national laws to local laws to community norms and so on. In any given situation, the AI would consult each layer in its hierarchy, moving from abstract precepts as defined by humans to the concrete but amorphous perceptions of the world’s information that AI has ingested. Only when an AI has exhausted that entire program and failed to find any layer of law adequately applicable in enabling or forbidding behavior would it consult what it has derived from its own early interaction with observable human behavior. In this way it would be empowered to act in alignment with human values even where no written law or norm exists.

To build and implement this set of rules and values, we would almost certainly need to rely on AI itself. No group of humans could match the scale and speed required to oversee the billions of internal and external judgments that AI systems would soon be called upon to make.

Several key features of the final mechanism for human-machine alignment must be absolutely perfect. First, the safeguards cannot be removed or circumvented. The control system must be at once powerful enough to handle a barrage of questions and uses in real time, comprehensive enough to do so authoritatively and acceptably across the world in every conceivable context, and flexible enough to learn, relearn, and adapt over time. Finally, undesirable behavior by a machine—whether due to accidental mishaps, unexpected system interactions, or intentional misuses—must be not merely prohibited but entirely prevented. Any punishment would come too late.

How might we get there? Before any AI system gets activated, a consortium of experts from private industry and academia, with government support, would need to design a set of validation tests for certification of the AI’s “grounding model” as both legal and safe. Safety-focused labs and nonprofits could test AIs on their risks, recommending additional training and validation strategies as needed.

Government regulators will have to determine certain standards and shape audit models for assuring AIs’ compliance. Before any AI model can be released publicly, it must be thoroughly reviewed for both its adherence to prescribed laws and mores and for the degree of difficulty involved in untraining it, in the event that it exhibits dangerous capacities. Severe penalties must be imposed on anyone responsible for models found to have been evading legal strictures. Documentation of a model’s evolution, perhaps recorded by monitoring AIs, would be essential to ensuring that models do not become black boxes that erase themselves and become safe havens for illegality.

Inscribing globally inclusive human morality onto silicon-based intelligence will require Herculean effort.  “Good” and “evil” are not self-evident concepts. The humans behind the moral encoding of AI—scientists, lawyers, religious leaders—would not be endowed with the perfect ability to arbitrate right from wrong on our collective behalf. Some questions would be unanswerable even by doxa. The ambiguity of the concept of “good” has been demonstrated in every era of human history; the age of AI is unlikely to be an exception.

One solution is to outlaw any sentient AI that remains unaligned with human values. But again: What are those human values? Without a shared understanding of who we are, humans risk relinquishing to AI the foundational task of defining our value and thereby justifying our existence. Achieving consensus on those values, and how they should be deployed, is the philosophical, diplomatic, and legal task of the century.

To preclude either our demotion or our replacement by machines, we propose the articulation of an attribute, or set of attributes, that humans can agree upon and that then can get programmed into the machines. As one potential core attribute, we would suggest Immanuel Kant’s conception of “dignity,” which is centered on the inherent worth of the human subject as an autonomous actor, capable of moral reasoning, who must not be instrumentalized as a means to an end. Why should intrinsic human dignity be one of the variables that defines machine decision making? Consider that mathematical precision may not easily encompass the concept of, for example, mercy. Even to many humans, mercy is an inexplicable ideal. Could a mechanical intelligence be taught to value, and even to express, mercy? If the moral logic cannot be formally taught, can it nonetheless be absorbed? Dignity—the kernel from which mercy blooms—might serve here as part of the rules-based assumptions of the machine.

[Derek Thompson: Why all the ChatGPT predictions are bogus]

Still, the number and diversity of rules that would have to be instilled in AI systems is staggering. And because no single culture should expect to dictate to another the morality of the AI on which it would be relying, machines would have to learn different rules for each country.

Since we would be using AI itself to be part of its own solution, technical obstacles would likely be among the easier challenges. These machines are superhumanly capable of memorizing and obeying instructions, however complicated. They might be able to learn and adhere to legal and perhaps also ethical precepts as well as, or better than, humans have done, despite our thousands of years of cultural and physical evolution.

Of course, another—superficially safer—approach would be to ensure that humans retain tactical control over every AI decision. But that would require us to stifle AI’s potential to help humanity. That’s why we believe that relying on the substratum of human morality as a form of strategic control, while relinquishing tactical control to bigger, faster, and more complex systems, is likely the best way forward for AI safety. Overreliance on unscalable forms of human control would not just limit the potential benefits of AI but could also contribute to unsafe AI. In contrast, the integration of human assumptions into the internal workings of AIs—including AIs that are programmed to govern other AIs—seems to us more reliable.

We confront a choice—between the comfort of the historically independent human and the possibilities of an entirely new partnership between human and machine. That choice is difficult. Instilling a bracing sense of apprehension about the rise of AI is essential. But, properly designed, AI has the potential to save the planet, and our species, and to elevate human flourishing. This is why progressing, with all due caution, toward the age of Homo technicus is the right choice. Some may view this moment as humanity’s final act. We see it, with sober optimism, as a new beginning.

The article was adapted from the forthcoming book Genesis: Artificial Intelligence, Hope, and the Human Spirit.

AI Is Killing the Internet’s Curiosity

The Atlantic

www.theatlantic.com › newsletters › archive › 2024 › 11 › ai-is-killing-the-internets-curiosity › 680600

This is an edition of The Atlantic Daily, a newsletter that guides you through the biggest stories of the day, helps you discover new ideas, and recommends the best in culture. Sign up for it here.

One of the most wonderful, and frustrating, things about Google Search is its inefficiency. The tool, at its most fundamental level, doesn’t provide knowledge. Instead, it points you to where it may, or may not, lie. That list of blue links can lead you down rabbit holes about your favorite sports team and toward deep understandings of debates you never knew existed. This tendency can also make it impossible to get a simple, straightforward fact.

But the experience of seeking information online is rapidly changing. Tech giants have for almost two years been promising AI-powered search tools that do provide knowledge and answers. And last week, OpenAI, Perplexity, and Google made announcements about their AI-powered search products that provide the clearest glimpse yet into what that future will look like. I’ve spent the past week using these tools for research and everyday queries, and reported on my findings in an article published today. “These tools’ current iterations surprised and, at times, impressed me,” I wrote, “yet even when they work perfectly, I’m not convinced that AI search is a wise endeavor.”

The promise of AI-powered search is quite different from Google’s—not to organize information so you can find it yourself, but to readily provide that information in a digestible, concise format. That made my searches faster and more convenient at times. But something deeply human was lost as a result. The rabbit holes and the unexpected obsessions are what’s beautiful about searching the internet; but AI, like the tech companies developing it, is obsessed with efficiency and optimization. What I loved about traditional Google searches, I wrote, is “falling into clutter and treasure, all the time, without ever intending to. AI search may close off these avenues to not only discovery but its impetus, curiosity.”

Illustration by The Atlantic

The Death of Search

By Matteo Wong

For nearly two years, the world’s biggest tech companies have said that AI will transform the web, your life, and the world. But first, they are remaking the humble search engine.

Chatbots and search, in theory, are a perfect match. A standard Google search interprets a query and pulls up relevant results; tech companies have spent tens or hundreds of millions of dollars engineering chatbots that interpret human inputs, synthesize information, and provide fluent, useful responses. No more keyword refining or scouring Wikipedia—ChatGPT will do it all. Search is an appealing target, too: Shaping how people navigate the internet is tantamount to shaping the internet itself.

Read the full article.

What to Read Next

The AI search war has begun: “Nearly two years after the arrival of ChatGPT, and with users growing aware that many generative-AI products have effectively been built on stolen information, tech companies are trying to play nice with the media outlets that supply the content these machines need,” I reported this past summer. Google is playing a dangerous game with AI search: “When more serious health questions get the AI treatment, Google is playing a risky game,” my colleague Caroline Mimbs Nyce wrote in May.

The Death of Search

The Atlantic

www.theatlantic.com › technology › archive › 2024 › 11 › ai-search-engines-curiosity › 680594

For nearly two years, the world’s biggest tech companies have said that AI will transform the web, your life, and the world. But first, they are remaking the humble search engine.

Chatbots and search, in theory, are a perfect match. A standard Google search interprets a query and pulls up relevant results; tech companies have spent tens or hundreds of millions of dollars engineering chatbots that interpret human inputs, synthesize information, and provide fluent, useful responses. No more keyword refining or scouring Wikipedia—ChatGPT will do it all. Search is an appealing target, too: Shaping how people navigate the internet is tantamount to shaping the internet itself.

Months of prophesying about generative AI have now culminated, almost all at once, in what may be the clearest glimpse yet into the internet’s future. After a series of limited releases and product demos, mired with various setbacks and embarrassing errors, tech companies are debuting AI-powered search engines as fully realized, all-inclusive products. Last Monday, Google announced that it would launch its AI Overviews in more than 100 new countries; that feature will now reach more than 1 billion users a month. Days later, OpenAI announced a new search function in ChatGPT, available to paid users for now and soon opening to the public. The same afternoon, the AI-search start-up Perplexity shared instructions for making its “answer engine” the default search tool in your web browser.

[Read: The AI search war has begun]

For the past week, I have been using these products in a variety of ways: to research articles, follow the election, and run everyday search queries. In turn I have scried, as best I can, into the future of how billions of people will access, relate to, and synthesize information. What I’ve learned is that these products are at once unexpectedly convenient, frustrating, and weird. These tools’ current iterations surprised and, at times, impressed me, yet even when they work perfectly, I’m not convinced that AI search is a wise endeavor.

For decades, the search bar has been a known entity. People around the world are accustomed to it; several generations implicitly regard Google as the first and best way to learn about basically anything. Enter a query, sift through a list of links, type a follow-up query, get more links, and so on until your question is answered or inquiry satisfied. That indirectness and wide aperture—all that clicking and scrolling—are in some ways the defining qualities of a traditional Google search, allowing (even forcing) you to traverse the depth and breadth of connections that justify the term world-wide web. The hyperlink, in this sense, is the building block of the modern internet.

That sprawl is lovely when you are going down a rabbit hole about Lucrezia de Medici, as I did when traveling in Florence last year, or when diving deep into a scientific dilemma. It is perfect for stumbling across delightful video clips and magazine features and social-media posts. And it is infuriating when you just need a simple biographical answer, or a brunch recommendation without the backstory of three different chefs, or a quick gloss of a complex research area without having to wade through obscure papers.

In recent years, more and more Google Search users have noted that the frustrations outweigh the delight—describing a growing number of paid advertisements, speciously relevant links engineered to top the search algorithm, and erroneous results. Generative AI promises to address those moments of frustration by providing a very different experience. Asking ChatGPT to search the web for the reasons Kamala Harris lost the presidential election yielded a short list with four factors: “economic concerns,” “demographic shifts,” “swing state dynamics,” and “campaign strategies.” It was an easy and digestible response, but not a particularly insightful one; in response to a follow-up question about voter demographics, ChatGPT provided a stream of statistics without context or analysis. A similar Google search, meanwhile, pulls up a wide range of news analyses that you have to read through. If you do follow Google’s links, you will develop a much deeper understanding of the American economy and politics.

Another example: Recently, I’ve been reading about a controversial proposed infrastructure project in Maryland. Google searches sent me through a labyrinth of public documents, corporate pitches, and hours-long recordings of city-council meetings, which took ages to review but sparked curiosity and left me deeply informed. ChatGPT, when asked, whipped up an accurate summary and timeline of events, and cited its sources—which was an extremely useful way to organize the reading I’d already done, but on its own might have been the end of my explorations.

I have long been a critic of AI-powered search. The technology has repeatedly fabricated information and struggled to accurately attribute its sources. Its creators have been accused of plagiarizing and violating the intellectual-property rights of major news organizations. None of these concerns has been fully allayed: The new ChatGPT search function, in my own use and other reports, has made some errors, mixing up dates, misreporting sports scores, and telling me that Brooklyn’s Prospect Park is bigger than Manhattan’s (much larger) Central Park. The links offered by traditional search engines are filled with errors too—but searchbots implicitly ask for your trust without verification. The citations don’t particularly invite you to click on them. And while OpenAI and Perplexity have entered into partnerships with any number of media organizations, including The Atlantic—perhaps competing for the high-quality, human-made content that their searchbots depend on—exactly how websites that once relied on ad revenue and subscriptions will fare on an AI-filtered web eludes me. (The editorial division of The Atlantic operates independently from the business division, which announced its corporate partnership with OpenAI in May.)

[Read: AI search is turning into the problem everyone worried about]

Although ChatGPT and Perplexity and Google AI Overviews cite their sources with (small) footnotes or bars to click on, not clicking on those links is the entire point. OpenAI, in its announcement of its new search feature, wrote that “getting useful answers on the web can take a lot of effort. It often requires multiple searches and digging through links to find quality sources and the right information for you. Now, chat can get you to a better answer.” Google’s pitch is that its AI “will do the Googling for you.” Perplexity’s chief business officer told me this summer that “people don’t come to Perplexity to consume journalism,” and that the AI tool will provide less traffic than traditional search. For curious users, Perplexity suggests follow-up questions so that, instead of opening a footnote, you keep reading in Perplexity.

The change will be the equivalent of going from navigating a library with the Dewey decimal system, and thus encountering related books on adjacent shelves, to requesting books for pickup through a digital catalog. It could completely reorient our relationship to knowledge, prioritizing rapid, detailed, abridged answers over a deep understanding and the consideration of varied sources and viewpoints. Much of what’s beautiful about searching the internet is jumping into ridiculous Reddit debates and developing unforeseen obsessions on the way to mastering a topic you’d first heard of six hours ago, via a different search; falling into clutter and treasure, all the time, without ever intending to. AI search may close off these avenues to not only discovery but its impetus, curiosity.

The issues with factuality and attribution may well be resolved—but even if they aren’t, tech companies show no signs of relenting. Controlling search means controlling how most people access every other digital thing—it’s an incredible platform to gain trust and visibility, advertise, or influence public opinion.

The internet is changing, and nobody outside these corporations has any say in it. And the biggest, most useful, and most frightening change may come from AI search engines working flawlessly. With AI, the goal is to keep you in one tech company’s ecosystem—to keep you using the AI interface, and getting the information that the AI deems relevant and necessary. The best searches are goal-oriented; the best responses are brief. Which perhaps shouldn’t be surprising coming from Silicon Valley behemoths that care, above all, about optimizing their businesses, products, and users’ lives.

A little, or even a lot, of inefficiency in search has long been the norm; AI will snuff it out. Our lives will be more convenient and streamlined, but perhaps a bit less wonderful and wonder-filled, a bit less illuminated. A process once geared toward exploration will shift to extraction. Less meandering, more hunting. No more unknown unknowns. If these companies really have their way, no more hyperlinks—and thus, no actual web.

The Tyranny of the Election Needle

The Atlantic

www.theatlantic.com › technology › archive › 2024 › 11 › election-needle-tyranny › 680547

The New York Times is once again poking readers’ eyes with its needle. A little digital gauge, like the one that might indicate that your boiler or nuclear-power plant is about to explode, “estimates the outcome of the race in real time based on polling data,” as the Times puts it. As we write this, the needle is piercing the red “Likely” side of the gauge, indicating that the decision is “Likely Trump.” To validate this qualitative assessment, the needle also clarifies, again in the moment we’re writing this, that Donald Trump has an 88 percent chance of victory. That’s good to know, or bad, depending on your preferences.

The Times is not alone in offering minute-to-minute assessments, of course. On television, news anchors talk endlessly about anything, or nothing, reporting live from Nevada or North Carolina. At CNN.com, the network’s familiar election map offers live results too, in an interface now so confusing that one of us couldn’t figure out how to back out of Georgia’s results after zooming in. The whole affair is meant to provide updates on a result entirely out of our control. At some point, probably not tonight and maybe not tomorrow, we will know who won the presidential race—and all of the other federal and state contests and ballot initiatives and the like too.

There is no good way to consume Election Night information anymore, if there ever was. Cable news is the loud, exhausting, touch-screen-assisted option for those who’re looking for the dopamine of an inoffensive Key-Race Alert. Social media is the best option if you’d like all of that, but updated each second, with commentary from Nazis and people who have placed big crypto bets on the outcome. Wrangling the data coming out of more than 100,000 precincts in fits and starts, across a country that spans six time zones and upwards of 161 million registered voters, is a glorious feat. The process is not, however, conducive to the human need to actually know things. In a way, the needle is the ChatGPT aggregation of the output of toxic sludge and useful information coming out of all of it. It is the supposed signal in the noise. But it may just be noise itself … until it isn’t.

For some time now, we’ve been chuckling at an ongoing joke on X about “building dashboards to give executives deeper insight into critical business functions.” (At least, we think they are jokes.) What would you do if a giant kaiju attacked the city? Make sure the dashboards are providing actionable insights into critical business functions. What do you do in your 30s? Get married, start companies, or … build dashboards to provide actionable insights. You get the picture.

The jokes are funny because they implicate a terrible everyday-life business thing called “business-intelligence dashboards.” Big data, data science, data-driven decision making, and a host of related biz buzz holds that you, me, him, them, everyone should collect as much data as possible about anything whatsoever, and then use those data to make decisions. But that’s hard, so it should be made easy. Thus, the dashboards. As if a car’s speedometer but scaled up to any level of complexity, a dashboard provides easy, quick, at-a-glance “insights” into the endless silos of data, making them “actionable.” This is a contradiction—thus the jokes.

So it is with the Times forecast needle (and the CNN map, and all the rest). Elections are ever more uncertain because they are always so close; because polling is fraught or broken; because disinformation, confusion, suppression, or God knows what else has made it impossible to have any sense of how these contests might turn out beforehand. The promise of synthesizing all of that uncertainty moments after a state’s polls close, and transforming it into knowledge, is too tempting to ignore. So you tune in to the news. You refresh the needle.

But what you learn is nothing other than how to feel good or bad in the moment. That the needle has clearly caused many extremely online coastal elites to have some gentle form of PTSD is clearly a feature, not a bug. It is a reminder of the needle’s power or, perhaps more accurately, its ability to move in such a way that it appears to usher in its own reality (when, really, it’s just reflecting changes in a spreadsheet of information). The needle is manipulative.

Worst of all, nothing about it is “actionable,” as the business-insights-dashboard fanatics would say. Dashboards promise a modicum of control. But what are you going to do, now that the polls are closed and you are in your pajamas? Cheer, or bite your nails, or attempt to lure your spouse away from the television or the needle, or eat cake or drink liquor or stare into space or high-five your buds or clean up after you ferret. There is nothing you can do. It’s out of your hands, and no amount of data, polling, chief analysis, or anything else can change that. You know this, and yet you still stare.

ChatGPT takes on Google, Meta's spending spree, and Microsoft's data center problem: AI news roundup

Quartz

qz.com › chatgpt-search-meta-ai-earnings-microsoft-data-centers-1851686792

Major tech companies reported earning results this week including Microsoft and Meta. But weaker-than-expected prospects for the next quarter sent artificial intelligence stocks, including Nvidia’s (NVDA), down on Thursday.

Read more...