Itemoids

DeepMind

The Big Questions About AI in 2024

The Atlantic

www.theatlantic.com › technology › archive › 2023 › 12 › ai-chatbot-llm-questions-2024 › 676942

Let us be thankful for the AI industry. Its leaders may be nudging humans closer to extinction, but this year, they provided us with a gloriously messy spectacle of progress. When I say “year,” I mean the long year that began late last November, when OpenAI released ChatGPT and, in doing so, launched generative AI into the cultural mainstream. In the months that followed, politicians, teachers, Hollywood screenwriters, and just about everyone else tried to understand what this means for their future. Cash fire-hosed into AI companies, and their executives, now glowed up into international celebrities, fell into Succession-style infighting. The year to come could be just as tumultuous, as the technology continues to evolve and its implications become clearer. Here are five of the most important questions about AI that might be answered in 2024.

Is the corporate drama over?

OpenAI’s Greg Brockman is the president of the world’s most celebrated AI company and the golden-retriever boyfriend of tech executives. Since last month, when Sam Altman was fired from his position as CEO and then reinstated shortly thereafter, Brockman has appeared to play a dual role—part cheerleader, part glue guy—for the company. As of this writing, he has posted no fewer than five group selfies from the OpenAI office to show how happy and nonmutinous the staffers are. (I leave to you to judge whether and to what degree these smiles are forced.) He described this year’s holiday party as the company’s best ever. He keeps saying how focused, how energized, how united everyone is. Reading his posts is like going to dinner with a couple after an infidelity has been revealed: No, seriously, we’re closer than ever. Maybe it’s true. The rank and file at OpenAI are an ambitious and mission-oriented lot. They were almost unanimous in calling for Altman’s return (although some have since reportedly said that they felt pressured to do so). And they may have trauma-bonded during the whole ordeal. But will it last? And what does all of this drama mean for the company’s approach to safety in the year ahead?

An independent review of the circumstances of Altman’s ouster is ongoing, and some relationships within the company are clearly strained. Brockman has posted a picture of himself with Ilya Sutskever, OpenAI’s safety-obsessed chief scientist, adorned with a heart emoji, but Altman’s feelings toward the latter have been harder to read. In his post-return statement, Altman noted that the company was discussing how Sutskever, who had played a central role in Altman’s ouster, “can continue his work at OpenAI.” (The implication: Maybe he can’t.) If Sutskever is forced out of the company or otherwise stripped of his authority, that may change how OpenAI weighs danger against speed of progress.

Is OpenAI sitting on another breakthrough?

During a panel discussion just days before Altman lost his job as CEO, he told a tantalizing story about the current state of the company’s AI research. A couple of weeks earlier, he had been in the room when members of his technical staff had pushed “the frontier of discovery forward,” he said. Altman declined to offer more details, unless you count additional metaphors, but he did mention that only four times since the company’s founding had he witnessed an advance of such magnitude.

During the feverish weekend of speculation that followed Altman’s firing, it was natural to wonder whether this discovery had spooked OpenAI’s safety-minded board members. We do know that in the weeks preceding Altman’s firing, company researchers raised concerns about a new “Q*” algorithm. Had the AI spontaneously figured out quantum gravity? Not exactly. According to reports, it had only solved simple mathematical problems, but it may have accomplished this by reasoning from first principles. OpenAI hasn’t yet released any official information about this discovery, if it is even right to think of it as a discovery. “As you can imagine, I can’t really talk about that,” Altman told me recently when I asked him about Q*. Perhaps the company will have more to say, or show, in the new year.

Does Google have an ace in the hole?

When OpenAI released its large-language-model chatbot in November 2022, Google was caught flat-footed. The company had invented the transformer architecture that makes LLMs possible, but its engineers had clearly fallen behind. Bard, Google’s answer to ChatGPT, was second-rate.

Many expected OpenAI’s leapfrog to be temporary. Google has a war chest that is surpassed only by Apple’s and Microsoft’s, world-class computing infrastructure, and storehouses of potential training data. It also has DeepMind, a London-based AI lab that the company acquired in 2014. The lab developed the AIs that bested world champions at chess and Go and intuited protein-folding secrets that nature had previously concealed from scientists. Its researchers recently claimed that another AI they developed is suggesting novel solutions to long-standing problems of mathematical theory. Google had at first allowed DeepMind to operate relatively independently, but earlier this year, it merged the lab with Google Brain, its homegrown AI group. People expected big things.

Then months and months went by without Google so much as announcing a release date for its next-generation LLM, Gemini. The delays could be taken as a sign that the company’s culture of innovation has stagnated. Or maybe Google’s slowness is a sign of its ambition? The latter possibility seems less likely now that Gemini has finally been released and does not appear to be revolutionary. Barring a surprise breakthrough in 2024, doubts about the company—and the LLM paradigm—will continue.

Are large language models already topping out?

Some of the novelty has worn off LLM-powered software in the mold of ChatGPT. That’s partly because of our own psychology. “We adapt quite quickly,” OpenAI’s Sutskever once told me. He asked me to think about how rapidly the field has changed. “If you go back four or five or six years, the things we are doing right now are utterly unimaginable,” he said. Maybe he’s right. A decade ago, many of us dreaded our every interaction with Siri, with its halting, interruptive style. Now we have bots that converse fluidly about almost any subject, and we struggle to remain impressed.

AI researchers have told us that these tools will only get smarter; they’ve evangelized about the raw power of scale. They’ve said that as we pump more data into LLMs, fresh wonders will emerge from them, unbidden. We were told to prepare to worship a new sand god, so named because its cognition would run on silicon, which is made of melted-down sand.

ChatGPT has certainly improved since it was first released. It can talk now, and analyze images. Its answers are sharper, and its user interface feels more organic. But it’s not improving at a rate that suggests that it will morph into a deity. Altman has said that OpenAI has begun developing its GPT-5 model. That may not come out in 2024, but if it does, we should have a better sense of how much more intelligent language models can become.

How will AI affect the 2024 election?

Our political culture hasn’t yet fully sorted AI issues into neatly polarized categories. A majority of adults profess to worry about AI’s impact on their daily life, but those worries aren’t coded red or blue. That’s not to say the generative-AI moment has been entirely innocent of American politics. Earlier this year, executives from companies that make chatbots and image generators testified before Congress and participated in tedious White House roundtables. Many AI products are also now subject to an expansive executive order.

But we haven’t had a big national election since these technologies went mainstream, much less one involving Donald Trump. Many blamed the spread of lies through social media for enabling Trump’s victory in 2016, and for helping him gin up a conspiratorial insurrection following his 2020 defeat. But the tools of misinformation that were used in those elections were crude compared with those that will be available next year.

A shady campaign operative could, for instance, quickly and easily conjure a convincing picture of a rival candidate sharing a laugh with Jeffrey Epstein. If that doesn’t do the trick, they could whip up images of poll workers stuffing ballot boxes on Election Night, perhaps from an angle that obscures their glitchy, six-fingered hands. There are reasons to believe that these technologies won’t have a material effect on the election. Earlier this year, my colleague Charlie Warzel argued that people may be fooled by low-stakes AI images—the pope in a puffer coat, for example—but they tend to be more skeptical of highly sensitive political images. Let’s hope he’s right.

Soundfakes, too, could be in the mix. A politician’s voice can now be cloned by AI and used to generate offensive clips. President Joe Biden and former President Trump have been public figures for so long—and voters’ perceptions of them are so fixed—that they may be resistant to such an attack. But a lesser-known candidate could be vulnerable to a fake audio recording. Imagine if during Barack Obama’s first run for the presidency, cloned audio of him criticizing white people in colorful language had emerged just days before the vote. Until bad actors experiment with these image and audio generators in the heat of a hotly contested election, we won’t know exactly how they’ll be misused, and whether their misuses will be effective. A year from now, we’ll have our answer.

The UK is investigating the Microsoft and OpenAI partnership

Quartz

qz.com › the-uk-is-investigating-the-microsoft-and-openai-partne-1851089181

The Competition and Markets Authority (CMA) is gathering information on Microsoft and OpenAI to decide whether their partnership threatens competition in the UK, which is home to Google’s DeepMind.

Read more...

Science Is Becoming Less Human

The Atlantic

www.theatlantic.com › technology › archive › 2023 › 12 › ai-scientific-research › 676304

This summer, a pill intended to treat a chronic, incurable lung disease entered mid-phase human trials. Previous studies have demonstrated that the drug is safe to swallow, although whether it will improve symptoms of the painful fibrosis that it targets remains unknown; this is what the current trial will determine, perhaps by next year. Such a tentative advance would hardly be newsworthy, except for a wrinkle in the medicine’s genesis: It is likely the first drug fully designed by artificial intelligence to come this far in the development pipeline.

The pill’s maker, the biotech company Insilico Medicine, used hundreds of AI models to discover both a new target in the body that could treat the fibrosis and which molecules might be synthesized for the drug itself. Those programs allowed Insilico to go from scratch to putting this drug through the first phase of human trials in two and a half years, rather than the typical five or so. Even if the pill proves useless, a real possibility, plenty of other drugs designed with the help of AI are in the wings. Scientists and companies alike hope that these will reach pharmacies far faster than traditionally designed medicine—bringing a drug to market typically takes well over a decade, and failure rates are high.

Medicine is just one aspect of a broader transformation in science. In only the past few months, AI has appeared to predict tropical storms with similar accuracy and much more speed than conventional models; Meta has released a model that can analyze brain scans to reproduce what a person is looking at; Google recently used AI to propose millions of new materials that could enhance supercomputers, electric vehicles, and more. Just as the technology has blurred the line between human-created and computer-generated text and images—upending how people work, learn, and socialize—AI tools are accelerating and refashioning some of the basic elements of science. “We can really make discoveries that would not be possible without the use of AI,” Marinka Zitnik, a biomedical and AI researcher at Harvard, told me.

Science has never been faster than it is today. But the introduction of AI is also, in some ways, making science less human. For centuries, knowledge of the world has been rooted in observing and explaining it. Many of today’s AI models twist this endeavor, providing answers without justifications and leading scientists to study their own algorithms as much as they study nature. In doing so, AI may be challenging the very nature of discovery.

AI exists to derive impossibly intricate patterns from data sets that are too large for any person to fathom, a mystifying phenomenon that has grown more familiar since ChatGPT was released last year. The chatbot—a tool suddenly at everyone’s fingertips that appears to synthesize the entire internet—changed how we can access and apply knowledge, but it simultaneously tainted much of our thinking with doubt. We do not understand exactly how generative-AI chatbots determine their responses, only that they sound remarkably human, making it hard to parse what is real, logical, or trustworthy, and whether writing, even our own, is fully human or bears a silicon touch. When a response does make sense, it can seem to offer a shortcut rather than any true understanding of how or why the answer came to be.

AI may be doing something similar in a broad range of scientific disciplines. Among the most notable scientific advances achieved via AI may be those in molecular biology from DeepMind, a leading AI research lab now based at Google. After DeepMind’s programs conquered the game of Go in 2016—a game so much more complex than chess that many thought that computers could never master it—Demis Hassabis, DeepMind’s CEO, told me that he began considering how to build an AI program for the decades-old challenge of protein folding. All sorts of biological processes depend on proteins, and every protein is made of a sequence of amino acids. How those molecules fold into a three-dimensional shape determines a protein’s function, and mapping those structures could help scientists develop new vaccines, kill antibiotics-resistant bacteria, and explore new cancer treatments. Without a protein’s 3-D shape, scientists have little more than a bunch of Lego bricks without instructions for putting them together.

Figuring out a single protein structure from a sequence of amino acids used to take years. But in 2022, DeepMind’s flagship scientific model, AlphaFold, found the most likely structure of almost every protein known to science—some 200 million of them. Much like the company’s chess- and Go-playing programs, which search for the best possible move, AlphaFold searches the number of possible structures for an amino-acid sequence to find the most probable one. The program compresses what could have been an entire Ph.D.’s worth of work into seconds, and it has been widely lauded for its “revolutionary impact” on basic biology and the development of novel treatments alike. Still, independent researchers have noted that despite inhuman speed, the model does not fully explain why a specific structure is likely. As a result, scientists are trying to demystify AlphaFold’s predictions, and Hassabis noted that those efforts are making good progress.

[Read: Welcome to the big blur]

AI allows researchers to study complex systems in “the world of bits” at a much faster pace than in the “world of atoms,” Hassabis said, and then physically test their hypotheses as a final step. The technology is pushing forward advances in numerous other disciplines—not just improving speed and scale, but changing what kind of research is thought possible. Neuroscientists at Meta and elsewhere, for instance, are turning artificial neural networks trained to “see” photographs or “read” text into hypotheses for how the brain processes both images and language. Biologists are using AI trained on genetic data to study rare diseases, improve immunotherapies, and better understand SARS-CoV-2 variants of concern. “Now we have viable hypotheses, where before we had mysteries,” Jim DiCarlo, a cognitive scientist at MIT who has pioneered the use of AI to study vision in the brain, told me.

Astronomers and physicists are using machine learning to process data sets from the universe that were too immense to touch before, Brice Ménard, an astrophysicist at Johns Hopkins, told me. Some experiments, such as the CERN particle collider, produce too much information to physically store. Researchers rely on AI to throw out familiar observations while keeping unknowns for analysis. “We don’t know what the needle looks like, because these are undetected physics events, but we know what the hay looks like,” Alexander Szalay, the director of the Institute for Data Intensive Science at Johns Hopkins, told me. “So computers are trained to recognize the hay and basically throw it away.”

[Read: Computers are learning to smell]

The long-term vision could even involve combining AI models and physical experiments in a sort of “self-driving lab,” Zitnik said, wherein computer programs and robots generate hypotheses, plan experiments to test them, and analyze the results. Such labs are a ways off, although prototypes do exist, such as the Scientific Autonomous Reasoning Agent, a robotic system that has already discovered new materials for renewable energy. SARA uses a laser to analyze and alter materials iteratively, with each loop lasting a few seconds, Carla Gomes, a computer scientist at Cornell, told me—reducing days of research to hours. This future, if it comes to pass, will elevate software and robots from tools to collaborators, even co-creators of knowledge.

Quantum observations too numerous for humans to store, experiments too rapid for humans to run, neuroscientific hypotheses too complex for humans to derive—even as AI enables scientific work never before thought possible, those same tools pose an epistemic dilemma. They will produce groundbreaking knowledge while breaking apart what it means to know in the first place.

“The holy grail of science is understanding,” Zitnik said. “To be able to understand a phenomenon, whether that’s the behavior of a cell or a planetary system, requires being able to identify causes and effects.” But AI models are famously opaque. They detect patterns based on gargantuan data sets via software architectures whose inner workings baffle human intuition and reasoning. Experts have taken to calling them “black boxes.”

This presents obvious problems for the scientific method. “We have to understand what is going on inside this black box so we can see where this discovery is coming from,” Szalay told me. To predict events without understanding why those predictions are accurate might gesture toward a different type of science, in which knowledge and the resulting actions are not always accompanied by an explanation. An AI model might predict a thunderstorm’s arrival but struggle to explain the underlying physics and atmospheric changes that triggered it, analyze an X-ray without showing how it arrived at its diagnosis, or propose abstract mathematical conjectures without proving them. Such shifts from observations and grounded reasoning to mathematical probability have happened in the sciences before: The equations of quantum mechanics, which emerged in the 20th century, accurately predict subatomic phenomena that physicists still don’t fully understand—leading Albert Einstein himself to doubt quantum theory.

[From the January 1951 issue: Faith in science]

Science itself may offer a solution to this conundrum. Physical experiments have uncovered a great deal in the past century about the quantum world, and similarly, AI tools might appear inscrutable partly because researchers haven’t spent enough time probing them. “You have to build the artifact first before you can pull it apart and scientifically analyze it,” Hassabis told me, and scientists have only recently begun to build AI models worthy of study. Even older numerical simulations, although far less complex than today’s AI models, are hard to interpret in an intuitive way, but they have nevertheless informed new discoveries for decades.

If researchers understand how artificial neurons respond to an image, they might be able to translate those predictions to biological neurons; if researchers understand which parts of an AI model link a mutation to a disease, scientists could gain new insights into the human genome. Such models are “fully observable systems. You can measure all the parts,” DiCarlo said. Whereas he cannot measure every neuron and synapse in a monkey brain during surgery, he can do that for an AI model. With the right access, AI programs might present scientists not with black boxes so much as with a new type of object requiring a new sort of inquiry—not “models” of the natural world so much as addendums to it. Some scientists even hope to build “digital twins” to simulate cells, organs, and the planet.

AI is not a silver bullet, though. AlphaFold may be revolutionary, and perhaps Insilico will indeed drastically reduce the time it takes to develop new medicine. But the technology has significant limitations. For instance, AI models need to train on large amounts of relevant data. AlphaFold is a “spectacular success,” Jennifer Listgarten, a computational biologist and computer scientist at UC Berkeley, told me—but it also “relied on very expensive, highly curated data that was generated over decades in the laboratory on a very crisply defined problem that could be evaluated extremely cleanly.” The lack of high-quality data in other disciplines can prevent or limit the use of AI.

Even with those data, the real world can be more complex and dynamic than a silicon simulation. Translating the static structure of a molecule into its interactions with various systems in the body, for instance, is a problem that researchers are still working on, Andreas Bender, who studies molecular informatics at the University of Cambridge, told me. AI can propose new medicines quickly, but “you still need to run the drug-discovery process, which is, of course, quite long,” John Jumper, a researcher at DeepMind who led the development of AlphaFold, told me.

Clinical trials take years and many are unsuccessful; plenty of AI drug start-ups and initiatives have scaled back. Those failures are, in some sense, evidence of science working. Experimental results, along with known physical laws, allow scientists to prevent their models from hallucinating, Anima Anandkumar, a computer scientist at Caltech, told me. No analogous laws of linguistic accuracy exist for chatbots—consumers have to trust Big Tech.

In a lab, novel predictions can be physically and safely tested in an isolated setting. But when developing drugs or treating patients, the stakes are much higher. Existing maps of the human genome, for instance, skew toward white Europeans, but the expression of many conditions, such as diabetes, varies significantly by race and ethnicity. Just as biased data sets produce racist chatbots, skewed biological data might mean that “models are not applicable to people of non-caucasian origin,” Bender told me, or those of different ages, or with existing diseases and on co-medications. A cancer-diagnosis program or treatment designed by AI might be especially effective only on a small slice of the population.

AI models might transform not just how we understand the world but how we understand understanding itself. If so, we must build out new models of knowledge as well—what we can trust, why, and when. Otherwise, our reliance on a chatbot, drug-discovery tool, or AI hurricane forecast might depart from the realm of science. It might be more akin to faith.