Tools
Change country:
The Atlantic
  1. SNL Has No Idea What to Do With the College Protests The Saturday Night Live cold open is usually a place for the series to do its most topical, often political, material. But an awkward sense of obligation hung over last night’s sketch, about campus protests surrounding the conflict in Gaza. The activism at colleges across the U.S. has been dominating the news, especially as the university and police responses have led to arrests. SNL seemed compelled to acknowledge this in some way, but all it gave its audience was uncomfortable, limp material that failed to make any real point about the urgent subjects animating protesters.The show opened with a fake NY1 community-affairs panel featuring parents of New York City college students. Even the cast seemed ill at ease. Heidi Gardner’s Hunter College mom spoke of the strain the protests had put on her relationship with her daughter. Mikey Day’s New School dad said, “I want to let my son make his own choices, but to be honest, it’s a little scary.” Kenan Thompson, playing a Columbia dad named Alphonse Roberts, appeared to be fully supportive of the protests—“Nothing makes me prouder than young people using their voices to fight for what they believe in”—until it was implied that his daughter might be out there. “I am supportive of y’all’s kids protesting,” he said; “not my kids.”Thompson’s delivery is routinely one of the most delightful things SNL has to offer, and that was the case here. Yet the sketch was underdeveloped, with little discussion of the reason students are demonstrating, and that tension hung in the air. There was some loose commentary on class and race in the divide between the concerned white parents and Alphonse, a Black man who works multiple jobs to pay Columbia’s exorbitant tuition. It turned out that Alphonse didn’t really care about the protests, as long as his daughter “had her butt in class,” pursuing the degree he was paying for. By the time Thompson got to the close and yelled, “Live from New York, it’s Saturday night!,” he looked both surprised and relieved the moment had come.It was as if the writers felt moved to say something but ended up reaching for the most potentially inoffensive angle. The ultimate joke was less about the protests and more about how expensive it is to attend college—a fact probably anyone in the audience would agree with. The sketch certainly didn’t have the boldness of Ramy Youssef’s opening monologue earlier this season, a deft stand-up set about how “complicated” his prayers are these days, in which he also said, “Please free the people of Palestine, please,” and “Please free the hostages, all the hostages, please.”In general lately, SNL seems to be struggling with the most newsworthy material. Regarding Israel and Gaza, that makes sense. The war is nearly impossible to joke about, especially for a program trying to be broadly appealing. But even in a sketch last night on a lower-stakes topic—the ongoing beef between Kendrick Lamar and Drake, featuring host Dua Lipa as a clueless southern morning-show reporter with limited knowledge of Black culture—the gags were labored.The sketches that hit were the most bizarre and absurdist: Sarah Sherman’s riff on The Elephant Man, titled “The Anomalous Man,” in which a 19th-century woman, played by Lipa, falls for Sherman’s monstrous playwright, who turns out to be a major player and is cheating on her; and “Sonny Angel,” in which Bowen Yang played a tiny, naked doll hooking up with Lipa’s character, a woman with a fixation on her “little boyfriend” toys.One sketch late in the night perhaps unintentionally captured SNL’s predicament. In a fake ad for the “Teeny Tiny Statement Pin,” the writers mocked celebrities who were too afraid of taking a stand to wear a normal-size pin on the red carpet. “It’s wrong to stay silent, but it’s also wrong to say too much,” Gardner said. “I just wish there was a way to split the difference.” The joke was supposed to be on wishy-washy famous people—but SNL might as well have been sending up itself.
    theatlantic.com
  2. Oh Great, Spiders Can Swim This article was originally published in Knowable Magazine.Shrubbery, toolsheds, basements—these are places one might expect to find spiders. But what about the beach? Or in a stream? Some spiders make their homes near or, more rarely, in water: tucking into the base of kelp stalks, spinning watertight cocoons in ponds or lakes, hiding under pebbles at the seaside or along a creek bank.“Spiders are surprisingly adaptable, which is one of the reasons they can inhabit this environment,” says Ximena Nelson, a behavioral biologist at the University of Canterbury in Christchurch, New Zealand.Finding aquatic or semiaquatic spiders is difficult work, Nelson says: She and a student have spent four years chasing a jumping spider known as Marpissa marina around the pebbly seaside beaches it likes, but too often, as soon as they manage to find one, it disappears under rocks. And sadly, some aquatic spiders may disappear altogether before they come to scientists’ attention, as their watery habitats shrivel because of climate change and other human activities.What scientists do know is that dozens of described spider species spend at least some of their time in or near the water, and more are almost surely awaiting discovery, says Sarah Crews, an arachnologist at the California Academy of Sciences, in San Francisco. It also appears that spiders evolved aquatic preferences on several distinct occasions throughout the history of this arthropod order. Crews and colleagues surveyed spiders and reported in 2019 that 21 taxonomic families are associated with aquatic habitats, suggesting that the evolutionary event occurred multiple independent times. Only a swashbuckling few—not even 0.3 percent of described spider species—are seashore spiders; many more have been found near fresh water, Nelson says.It’s not clear what would induce successful land-dwelling critters to move to watery habitats. Spiders, as a group, probably evolved about 400 million years ago from chunkier creatures that had recently left the water. These arthropods lacked the skinny waist sported by modern spiders. Presumably, the spiders that later returned to a life aquatic were strongly drawn by something to eat there, or driven by unsafe conditions on land, says Geerat Vermeij, a paleobiologist and distinguished professor at UC Davis—because water would have presented major survival challenges.“Since they depend on air so much, they are severely limited in whether they can do anything at all when they are submerged, other than just toughing it out,” Vermeij says. Newly aquatic spiders would have had to compete with predators better adapted to watery conditions, such as crustaceans, with competition particularly fierce in the oceans, Vermeij says. And if water floods a spider’s air-circulation system, it will die, so adaptations were obviously needed.But spiders as a group already possess several water-friendly features, Crews suggests. They have waxy, water-repellent exteriors, often covered in hairs that conveniently trap air bubbles. Even having eight legs is helpful, Nelson says: Spiders can distribute their weight nicely while they skitter across a water surface, or use their octet of appendages to row along.[Read: The spiders that choose death]Some spiders take their aquatic adaptations to the next level, though. Consider the diving-bell spider, Argyroneta aquatica, an overachieving arachnid that is the only one known to do it all underwater: breathe, hunt, dine on insects and their larvae, and make spiderlings. Found in fresh water in Europe and parts of Asia, it spins a silken underwater canopy and brings air bubbles from the surface to its submerged home via its body hairs. When it goes out, it carries a smaller air bubble, like a little scuba tank, on its back.Seashore spiders face particularly daunting conditions, says Nelson, who co-authored an article about adaptations of marine spiders for the 2024 Annual Review of Entomology. “There’s a splash zone,” she says. “It’s kind of a wild environment.” A spider might be baking in the hot sun one minute and drenched in chilly salt water the next. Some spiders migrate up and down their beaches with the tides; Nelson speculates that they monitor lunar cycles to anticipate when to move.Other seashore spiders spin watertight nests where they hide out for hours while the tide is in. M. marina, for example, seeks seashells with nice, concave spaces in which to spin safe tents. Another spider, Desis marina, hides in holdfasts where bull kelp attaches to rocks, lining the holdfast’s interior with silk to create an air-filled pocket and staying submerged for as long as 19 days. D. marina emerges only when the tide is going out, to hunt for invertebrates like shrimp.A spider that’s even occasionally submerged in salt water or that eats briny seafood will also have to maintain proper internal salt levels. “Presumably, they will be able to concentrate the salt somehow and then poo it out,” Nelson says. Scientists don’t know how marine spiders pull this off. And at least one intertidal-zone spider, Desis formidabilis of South Africa’s cape, comfortably maintains an interior salt concentration much like the crustaceans it eats, according to a 1984 study. (Freshwater species also probably require adaptations because their insides must stay saltier than their surroundings or food, Vermeij speculates.)When a spider hides out with a limited air supply for days or weeks at a time, oxygen levels also may become a crucial issue. Intriguingly, researchers have identified gene variants within the oxygen-guzzling, energy-making mitochondria of aquatic spiders that may help them cope with low-oxygen environments. These changes mirror beneficial changes to mitochondrial genes in birds that live in high-altitude, low-oxygen environments.In another study, researchers investigated the genes used in the silk glands of aquatic and land spiders. They found that water-spider silk seems to have a high proportion of water-repelling amino acids—which might also be an adaptation, they suggest.But all the adaptations in the world might not be enough to save some water spiders. Nelson’s M. marina, for example, seems to be very particular about the beaches it occupies. The pebbles must be just right, not too big or small. If sea-level rise inundates M. marina’s beaches, it’s possible the spiders will have nowhere else to go, Nelson says. “So those spiders will be lost.”Marco Isaia, an arachnologist at the University of Turin, in Italy, investigated the wetland habitats of the diving-bell spider and the fen raft spider, Dolomedes plantarius. As wetlands continue to disappear, the habitats available to each species will contract by more than 25 percent over a decade, and their ideal ranges will move northward, Isaia and colleagues predicted in a 2022 study. It would be difficult for the spiders to cross dry land for new wetlands, and Northern European winters might prove too cold anyway. “The loss and degradation of wetland habitats is expected to have serious impacts on their survival,” Isaia says, “and an increase in their extinction risk.”Given these risks, some aquatic spiders might go the way of the dodo before science gets a handle on them. “I suspect in every rocky bed of beach or river, there are probably spiders that we just don’t know exist there,” Nelson says. “Because they’re hiding.”
    theatlantic.com
  3. ElevenLabs Is Building an Army of Voice Clones My voice was ready. I’d been waiting, compulsively checking my inbox. I opened the email and scrolled until I saw a button that said, plainly, “Use voice.” I considered saying something aloud to mark the occasion, but that felt wrong. The computer would now speak for me.I had thought it’d be fun, and uncanny, to clone my voice. I’d sought out the AI start-up ElevenLabs, paid $22 for a “creator” account, and uploaded some recordings of myself. A few hours later, I typed some words into a text box, hit “Enter,” and there I was: all the nasal lilts, hesitations, pauses, and mid-Atlantic-by-way-of-Ohio vowels that make my voice mine.It was me, only more pompous. My voice clone speaks with the cadence of a pundit, no matter the subject. I type I like to eat pickles, and the voice spits it out as if I’m on Meet the Press. That’s not my voice’s fault; it is trained on just a few hours of me speaking into a microphone for various podcast appearances. The model likes to insert ums and ahs: In the recordings I gave it, I’m thinking through answers in real time and choosing my words carefully. It’s uncanny, yes, but also quite convincing—a part of my essence that’s been stripped, decoded, and reassembled by a little algorithmic model so as to no longer need my pesky brain and body. Listen to the author's AI voice: Using ElevenLabs, you can clone your voice like I did, or type in some words and hear them spoken by “Freya,” “Giovanni,” “Domi,” or hundreds of other fake voices, each with a different accent or intonation. Or you can dub a clip into any one of 29 languages while preserving the speaker’s voice. In each case, the technology is unnervingly good. The voice bots don’t just sound far more human than voice assistants such as Siri; they also sound better than any other widely available AI audio software right now. What’s different about the best ElevenLabs voices, trained on far more audio than what I fed into the machine, isn’t so much the quality of the voice but the way the software uses context clues to modulate delivery. If you feed it a news report, it speaks in a serious, declarative tone. Paste in a few paragraphs of Hamlet, and an ElevenLabs voice reads it with a dramatic storybook flare. Listen to ElevenLabs read Hamlet: ElevenLabs launched an early version of its product a little over a year ago, but you might have listened to one of its voices without even knowing it. Nike used the software to create a clone of the NBA star Luka Dončić’s voice for a recent shoe campaign. New York City Mayor Eric Adams’s office cloned the politician’s voice so that it could deliver robocall messages in Spanish, Yiddish, Mandarin, Cantonese, and Haitian Creole. The technology has been used to re-create the voices of children killed in the Parkland school shooting, to lobby for gun reform. An ElevenLabs voice might be reading this article to you: The Atlantic uses the software to auto-generate audio versions of some stories, as does The Washington Post.It’s easy, when you play around with the ElevenLabs software, to envision a world in which you can listen to all the text on the internet in voices as rich as those in any audiobook. But it’s just as easy to imagine the potential carnage: scammers targeting parents by using their children’s voice to ask for money, a nefarious October surprise from a dirty political trickster. I tested the tool to see how convincingly it could replicate my voice saying outrageous things. Soon, I had high-quality audio of my voice clone urging people not to vote, blaming “the globalists” for COVID, and confessing to all kinds of journalistic malpractice. It was enough to make me check with my bank to make sure any potential voice-authentication features were disabled.I went to visit the ElevenLabs office and meet the people responsible for bringing this technology into the world. I wanted to better understand the AI revolution as it’s currently unfolding. But the more time I spent—with the company and the product—the less I found myself in the present. Perhaps more than any other AI company, ElevenLabs offers a window into the near future of this disruptive technology. The threat of deepfakes is real, but what ElevenLabs heralds may be far weirder. And nobody, not even its creators, seems ready for it.In mid-November, I buzzed into a brick building on a London side street and walked up to the second floor. The corporate headquarters of ElevenLabs—a $1 billion company—is a single room with a few tables. No ping-pong or beanbag chairs—just a sad mini fridge and the din of dutiful typing from seven employees packed shoulder to shoulder. Mati Staniszewski, ElevenLabs’ 29-year-old CEO, got up from his seat in the corner to greet me. He beckoned for me to follow him back down the stairs to a windowless conference room ElevenLabs shares with a company that, I presume, is not worth $1 billion.Staniszewski is tall, with a well-coiffed head of blond hair, and he speaks quickly in a Polish accent. Talking with him sometimes feels like trying to engage in conversation with an earnest chatbot trained on press releases. I started our conversation with a few broad questions: What is it like to work on AI during this moment of breathless hype, investor interest, and genuine technological progress? What’s it like to come in each day and try to manipulate such nascent technology? He said that it’s exciting.We moved on to what Staniszewski called his “investor story.” He and the company’s co-founder, Piotr Dabkowski, grew up together in Poland watching foreign movies that were all clumsily dubbed into a flat Polish voice. Man, woman, child—whoever was speaking, all of the dialogue was voiced in the same droning, affectless tone by male actors known as lektors.They both left Poland for university in the U.K. and then settled into tech jobs (Staniszewski at Palantir and Dabkowski at Google). Then, in 2021, Dabkowski was watching a film with his girlfriend and realized that Polish films were still dubbed in the same monotone lektor style. He and Staniszewski did some research and discovered that markets outside Poland were also relying on lektor-esque dubbing. Mati Staniszewski’s “investor story” as CEO of ElevenLabs begins in Poland, where he grew up watching foreign films clumsily dubbed into a flat voice. (Daniel Stier for The Atlantic) The next year, they founded ElevenLabs. AI voices were everywhere—think Alexa, or a car’s GPS—but actually good AI voices, they thought, would finally put an end to lektors. The tech giants have hundreds or thousands of employees working on AI, yet ElevenLabs, with a research team of just seven people, built a voice tool that’s arguably better than anything its competitors have released. The company poached researchers from top AI companies, yes, but it also hired a college dropout who’d won coding competitions, and another “who worked in call centers while exploring audio research as a side gig,” Staniszewski told me. “The audio space is still in its breakthrough stage,” Alex Holt, the company’s vice president of engineering, told me. “Having more people doesn’t necessarily help. You need those few people that are incredible.”ElevenLabs knew its model was special when it started spitting out audio that accurately represented the relationships between words, Staniszewski told me—pronunciation that changed based on the context (minute, the unit of time, instead of minute, the description of size) and emotion (an exclamatory phrase spoken with excitement or anger).Much of what the model produces is unexpected—sometimes delightfully so. Early on, ElevenLabs’ model began randomly inserting applause breaks after pauses in its speech: It had been training on audio clips from people giving presentations in front of live audiences. Quickly, the model began to improve, becoming capable of ums and ahs. “We started seeing some of those human elements being replicated,” Staniszewski said. The big leap was when the model began to laugh like a person. (My voice clone, I should note, struggles to laugh, offering a machine-gun burst of “haha”s that sound jarringly inhuman.)Compared with OpenAI and other major companies, which are trying to wrap their large language models around the entire world and ultimately build an artificial human intelligence, ElevenLabs has ambitions that are easier to grasp: a future in which ALS patients can still communicate in their voice after they lose their speech. Audiobooks that are ginned up in seconds by self-published authors, video games in which every character is capable of carrying on a dynamic conversation, movies and videos instantly dubbed into any language. A sort of Spotify of voices, where anyone can license clones of their voice for others to use—to the dismay of professional voice actors. The gig-ification of our vocal cords.What Staniszewski also described when talking about ElevenLabs is a company that wants to eliminate language barriers entirely. The dubbing tool, he argued, is its first step toward that goal. A user can upload a video, and the model will translate the speaker’s voice into a different language. When we spoke, Staniszewski twice referred to the Babel fish from the science-fiction book The Hitchhiker’s Guide to the Galaxy—he described making a tool that immediately translates every sound around a person into a language they can understand.Every ElevenLabs employee I spoke with perked up at the mention of this moonshot idea. Although ElevenLabs’ current product might be exciting, the people building it view current dubbing and voice cloning as a prelude to something much bigger. I struggled to separate the scope of Staniszewski’s ambition from the modesty of our surroundings: a shared conference room one floor beneath the company’s sparse office space. ElevenLabs may not achieve its lofty goals, but I was still left unmoored by the reality that such a small collection of people could build something so genuinely powerful and release it into the world, where the rest of us have to make sense of it.ElevenLabs’ voice bots launched in beta in late January 2023. It took very little time for people to start abusing them. Trolls on 4chan used the tool to make deepfakes of celebrities saying awful things. They had Emma Watson reading Mein Kampf and the right-wing podcaster Ben Shapiro making racist comments about Representative Alexandria Ocasio-Cortez. In the tool’s first days, there appeared to be virtually no guardrails. “Crazy weekend,” the company tweeted, promising to crack down on misuse.ElevenLabs added a verification process for cloning; when I uploaded recordings of my voice, I had to complete multiple voice CAPTCHAs, speaking phrases into my computer in a short window of time to confirm that the voice I was duplicating was my own. The company also decided to limit its voice cloning strictly to paid accounts and announced a tool that lets people upload audio to see if it is AI generated. But the safeguards from ElevenLabs were “half-assed,” Hany Farid, a deepfake expert at UC Berkeley, told me—an attempt to retroactively focus on safety only after the harm was done. And they left glaring holes. Over the past year, the deepfakes have not been rampant, but they also haven’t stopped.I first started reporting on deepfakes in 2017, after a researcher came to me with a warning of a terrifying future where AI-generated audio and video would bring about an “infocalypse” of impersonation, spam, nonconsensual sexual imagery, and political chaos, where we would all fall into what he called “reality apathy.” Voice cloning already existed, but it was crude: I used an AI voice tool to try to fool my mom, and it worked only because I had the halting, robotic voice pretend I was losing cell service. Since then, fears of an infocalypse have lagged behind the technology’s ability to distort reality. But ElevenLabs has closed the gap.The best deepfake I’ve seen was from the filmmaker Kenneth Lurt, who used ElevenLabs to clone Jill Biden’s voice for a fake advertisement where she’s made to look as if she’s criticizing her husband over his handling of the Israel-Gaza conflict. The footage, which deftly stitches video of the first lady giving a speech with an ElevenLabs voice-over, is incredibly convincing and has been viewed hundreds of thousands of times. The ElevenLabs technology on its own isn’t perfect. “It’s the creative filmmaking that actually makes it feel believable,” Lurt said in an interview in October, noting that it took him a week to make the clip.“It will totally change how everyone interacts with the internet, and what is possible,” Nathan Lambert, a researcher at the Allen Institute for AI, told me in January. “It’s super easy to see how this will be used for nefarious purposes.” When I asked him if he was worried about the 2024 elections, he offered a warning: “People aren’t ready for how good this stuff is and what it could mean.” When I pressed him for hypothetical scenarios, he demurred, not wanting to give anyone ideas. Daniel Stier for The Atlantic A few days after Lambert and I spoke, his intuitions became reality. The Sunday before the New Hampshire presidential primary, a deepfaked, AI-generated robocall went out to registered Democrats in the state. “What a bunch of malarkey,” the robocall began. The voice was grainy, its cadence stilted, but it was still immediately recognizable as Joe Biden’s drawl. “Voting this Tuesday only enables the Republicans in their quest to elect Donald Trump again,” it said, telling voters to stay home. In terms of political sabotage, this particular deepfake was relatively low stakes, with limited potential to disrupt electoral outcomes (Biden still won in a landslide). But it was a trial run for an election season that could be flooded with reality-blurring synthetic information.Researchers and government officials scrambled to locate the origin of the call. Weeks later, a New Orleans–based magician confessed that he’d been paid by a Democratic operative to create the robocall. Using ElevenLabs, he claimed, it took him less than 20 minutes and cost $1.Afterward, ElevenLabs introduced a “no go”–voices policy, preventing users from uploading or cloning the voice of certain celebrities and politicians. But this safeguard, too, had holes. In March, a reporter for 404 Media managed to bypass the system and clone both Donald Trump’s and Joe Biden’s voices simply by adding a minute of silence to the beginning of the upload file. Last month, I tried to clone Biden’s voice, with varying results. ElevenLabs didn’t catch my first attempt, for which I uploaded low-quality sound files from YouTube videos of the president speaking. But the cloned voice sounded nothing like the president’s—more like a hoarse teenager’s. On my second attempt, ElevenLabs blocked the upload, suggesting that I was about to violate the company’s terms of service.For Farid, the UC Berkeley researcher, ElevenLabs’ inability to control how people might abuse its technology is proof that voice cloning causes more harm than good. “They were reckless in the way they deployed the technology,” Farid said, “and I think they could have done it much safer, but I think it would have been less effective for them.”The core problem of ElevenLabs—and the generative-AI revolution writ large—is that there is no way for this technology to exist and not be misused. Meta and OpenAI have built synthetic voice tools, too, but have so far declined to make them broadly available. Their rationale: They aren’t yet sure how to unleash their products responsibly. As a start-up, though, ElevenLabs doesn’t have the luxury of time. “The time that we have to get ahead of the big players is short,” Staniszewski said. “If we don’t do it in the next two to three years, it’s going to be very hard to compete.” Despite the new safeguards, ElevenLabs’ name is probably going to show up in the news again as the election season wears on. There are simply too many motivated people constantly searching for ways to use these tools in strange, unexpected, even dangerous ways.In the basement of a Sri Lankan restaurant on a soggy afternoon in London, I pressed Staniszewski about what I’d been obliquely referring to as “the bad stuff.” He didn’t avert his gaze as I rattled off the ways ElevenLabs’ technology could be and has been abused. When it was his time to speak, he did so thoughtfully, not dismissively; he appears to understand the risks of his products. “It’s going to be a cat-and-mouse game,” he said. “We need to be quick.”Later, over email, he cited the “no go”–voices initiative and told me that ElevenLabs is “testing new ways to counteract the creation of political content,” adding more human moderation and upgrading its detection software. The most important thing ElevenLabs is working on, Staniszewski said—what he called “the true solution”—is digitally watermarking synthetic voices at the point of creation so civilians can identify them. That will require cooperation across dozens of companies: ElevenLabs recently signed an accord with other AI companies, including Anthropic and OpenAI, to combat deepfakes in the upcoming elections, but so far, the partnership is mostly theoretical.The uncomfortable reality is that there aren’t a lot of options to ensure bad actors don’t hijack these tools. “We need to brace the general public that the technology for this exists,” Staniszewski said. He’s right, yet my stomach sinks when I hear him say it. Mentioning media literacy, at a time when trolls on Telegram channels can flood social media with deepfakes, is a bit like showing up to an armed conflict in 2024 with only a musket.The conversation went on like this for a half hour, followed by another session a few weeks later over the phone. A hard question, a genuine answer, my own palpable feeling of dissatisfaction. I can’t look at ElevenLabs and see beyond the risk: How can you build toward this future? Staniszewski seems unable to see beyond the opportunities: How can’t you build toward this future? I left our conversations with a distinct sense that the people behind ElevenLabs don’t want to watch the world burn. The question is whether, in an industry where everyone is racing to build AI tools with similar potential for harm, intentions matter at all.To focus only on deepfakes elides how ElevenLabs and synthetic audio might reshape the internet in unpredictable ways. A few weeks before my visit, ElevenLabs held a hackathon, where programmers fused the company’s tech with hardware and other generative-AI tools. Staniszewski said that one team took an image-recognition AI model and connected it to both an Android device with a camera and ElevenLabs’ text-to-speech model. The result was a camera that could narrate what it was looking at. “If you’re a tourist, if you’re a blind person and want to see the world, you just find a camera,” Staniszewski said. “They deployed that in a weekend.”Repeatedly during my visit, ElevenLabs employees described these types of hybrid projects—enough that I began to see them as a helpful way to imagine the next few years of technology. Products that all hook into one another herald a future that’s a lot less recognizable. More machines talking to machines; an internet that writes itself; an exhausting, boundless comingling of human art and human speech with AI art and AI speech until, perhaps, the provenance ceases to matter.I came to London to try to wrap my mind around the AI revolution. By staring at one piece of it, I thought, I would get at least a sliver of certainty about what we’re barreling toward. Turns out, you can travel across the world, meet the people building the future, find them to be kind and introspective, ask them all of your questions, and still experience a profound sense of disorientation about this new technological frontier. Disorientation. That’s the main sense of this era—that something is looming just over the horizon, but you can’t see it. You can only feel the pit in your stomach. People build because they can. The rest of us are forced to adapt.
    theatlantic.com