Power
Platformer

OpenAI vs. the world

By Casey Newton
Tue 05.21.24
US-APEC-SUMMIT
ANDREW CABALLERO-REYNOLDS/AFP via Getty Images

How worried should we be about OpenAI’s leadership?

The company hasn’t been the same since Sam Altman’s return — and its treatment of Scarlett Johansson should concern everyone

Casey Newton
5/21/24 11:57AM

OpenAI hasn't been the same company since Sam Altman returned as its CEO.

It’s a thought that has occurred to me repeatedly over the past week as the company has found mounting reasons to stop, reverse course, and apologize over product and personnel issues. And it’s one that feels increasingly urgent in a world where the company’s technology is poised to serve as the artificial intelligence backbone on both Microsoft and Apple devices. 

On Monday, actress Scarlett Johansson revealed that the OpenAI CEO had been negotiating with her for the better part of a year to lend her voice to ChatGPT. “He told me that he felt that by my voicing the system, I could bridge the gap between tech companies and creatives and help consumers to feel comfortable with the seismic shift concerning humans and AI,” Johansson said in a statement published by NPR’s Bobby Allyn. “He said he felt that my voice would be comforting to people.”

Johansson declined to participate. But when OpenAI unveiled voices for ChatGPT in September, users were struck by the uncanny similarity between the voice it called Sky and Johansson’s. And it should not have been a surprise: Altman has said that Her, in which Johansson voices an AI assistant with whom the film’s protagonist falls in love, is his favorite movie. Last week, after OpenAI showcased an updated version of Sky that takes advantage of its latest models, Altman posted the movie’s title suggestively on X.

“When I heard the released demo, I was shocked, angered and in disbelief that Mr. Altman would pursue a voice that sounded so eerily similar to mine that my closest friends and news outlets could not tell the difference,” Johansson said.

But there is even more to the story, Johansson wrote.

“Two days before the ChatGPT 4.0 demo was released, Mr. Altman contacted my agent, asking me to reconsider,” she wrote. “Before we could connect, the system was out there.”

And so even as Altman attempted to renegotiate with Johansson behind the scenes, he intentionally drew attention to Sky’s similarity to Johansson in Her to his nearly 3 million Twitter followers — and released the product before finishing the negotiations.

In the moment, it looked like a grand success. The OpenAI team’s demonstration of a warmer, more emotional chatbot last week was one of the most impressive tech demos I’ve seen. In the end, I found its aggressive flirting unsettling, and seemingly at odds with Altman’s previous statements to me that the company would not pursue AI tools that doubled as romantic companions. But the demo unquestionably succeeded in captivating the tech world — and stealing thunder from Google on the eve of its annual developer conference.

Behind the scenes, though, Johansson was starting to ask questions. She hired attorneys, “who wrote two letters to Mr. Altman and OpenAI, setting out what they had done and asking them to detail the exact process by which they created the ‘Sky’ voice,” she wrote.

That did the trick: on Sunday night, OpenAI announced it would pull the Sky voice from ChatGPT. It also put up a blog post outlining the casting and recording process for the five voices it eventually selected for ChatGPT, Sky included.

OpenAI has taken pains to frame all of this as a big misunderstanding. CTO Mira Murati told reporters last week that Sky is not intended to sound like Johansson. And earlier on Monday, the company’s model behavior lead, Joanne Jang, told The Verge that “We’ve been in conversations with ScarJo’s team because there seems to be some confusion.”

OpenAI did not respond to my requests for comment today. But Johansson’s statement suggests that in reality, there was never any legitimate confusion about what was going on here. Altman asked to license her voice to realize the fantasy of a real-life Her; she declined; and the company proceeded anyway with a voice as similar to hers as it could get.

“I look forward to resolution in the form of transparency and the passage of appropriate legislation to help ensure that individual rights are protected,” she said.

II.

Johansson is one of the world’s most famous actresses, and she speaks for an entire class of creatives who are now wrestling with the fact that automated systems have begun to erode the value of their work. OpenAI’s decision to usurp her voice for its own purposes will now get wide and justified attention.

At the same time, it’s possible that Johansson’s experience wasn’t even the most important thing that happened at OpenAI in the past week.

Last Tuesday, after all, Ilya Sutskever announced he was leaving the company. Sutskever, a co-founder of the company who is renowned both for his technical ability and his concerns about the potential of AI to do harm, was a leader of the faction that briefly deposed Altman in November. Sutskever reversed his position once it became clear that the vast majority of OpenAi employees were prepared to quit if Altman did not return as CEO. But Sutskever hasn’t worked for the company since.

In an anodyne public statement, Sutskever said only that he now plans to work on “a project that is very personally meaningful to” him.

Why not say more about the circumstances of his departure? Vox reported over the weekend that OpenAI’s nondisclosure agreements forbid criticizing the company for the duration of the former employee’s lifetime; forbids them from acknowledging the existence of the NDA; and forces them to give up all vested equity in the company if they refuse to sign it. (After the story drew wide attention, on Saturday Altman apologized and said the company would remove the provision about clawing back equity for violating the NDA.)

As OpenAI’s chief scientist, Sutskever also led its so-called “superalignment team,” which it formed last July to research ways of ensuring that advanced AI systems act safely and in accordance with human intent. Among other things, the company promised to dedicate 20 percent of its scarce and critical computing power to the project. 

But Altman’s ouster led to wide disparagement of AI safety efforts, which were wrongly identified as the reason his co-workers wanted to fire him. (The actual-if-vague reason was that he was not “consistently candid in his communications,” according to the company’s former board, a charge that seems worth revisiting in light of Johansson’s experience with him.) 

And so it perhaps should not be surprising that, upon his return, safety efforts were deprioritized. The superalignment team has been disbanded after less than a year, with its remaining employees absorbed into other teams. Jan Leike, who led the superalignment team under Sutskever, quit on Friday, saying OpenAI’s “safety culture and processes have taken a backseat to shiny products.”

“I joined because I thought OpenAI would be the best place in the world to do this research,” Leike wrote on X. “However, I have been disagreeing with OpenAI leadership about the company’s core priorities for quite some time, until we finally reached a breaking point.”

In response, Altman and OpenAI president Greg Brockman issued a long statement that gestured toward building things responsibly without addressing Laike’s criticism.

III.

What to make of all this? 

On one hand, it’s not unusual for a fast-growing startup to go through some tumult as it scales up. On the other, OpenAI was not founded to be a typical fast-growing startup: it was founded to be a nonprofit research lab focused on the safe development of superhuman intelligence.

Since the release of ChatGPT — and much more so since Altman’s return from exile — OpenAI has looked like anything but. It has consistently pushed the boundaries of what AI can do, both from a technical standpoint and from the vantage of what it believes society can handle.

And last week, determined to show up Google on the eve of I/O, the company went through with its demo of a Johansson voice clone, against Johansson’s explicit wishes, because … Altman is a fan? 

For the moment, the only cost to OpenAI from any of this has been a torrent of aggrieved social media posts. On Monday its APIs were at the center of a range of new PCs unveiled by its largest investor, Microsoft; next month they are expected to be announced as part of a big new partnership with Apple. One analysis last week suggested that the company’s mobile app revenue grew 22 percent the day it showcased the not-Johansson voice demo.

For years now, OpenAI told everyone that these were all secondary concerns — that its deeper ambition was something nobler, and more public-spirited. But since Altman’s return, the company has been telling a different story: a story about winning at all costs. 

And why bother with superalignment, when there’s winning to do? 

Why bother getting actresses’ permission, when the right numbers are all still going up?


Casey Newton writes Platformer, a daily guide to understanding social networks and their relationships with the world. This piece was originally published on Platformer.

More Power

See all Power
power

Delta dips as the Trump administration orders the end of its joint venture with Aeromexico

Shares of Delta Air Lines ticked down on Tuesday morning following the Trump administration’s order that the airline dissolve its approximately 9-year-old joint venture with Aeromexcio by January 1, 2026.

Delta said it was disappointed in the decision, adding that the termination will “cause significant harm to U.S. jobs, communities and consumers traveling between the U.S. and Mexico.” CEO Ed Bastian previously said that the administration’s regulatory stance could be a “breath of fresh air” for the aviation industry.

The Biden administration tentatively decided last year to not renew the antitrust immunity agreement covering the joint venture. At the time, Delta said “$800 million in annual consumer benefits would evaporate” if the partnership were terminated.

Collaboration isn’t over between the two airlines: the Department of Transportation said Delta can maintain its 20% stake in the Mexican airline and the partnership can continue through “arms-length activities such as codesharing, marketing, and frequent flyer cooperation.”

power

The DOJ is suing Uber, alleging the company discriminates against passengers with disabilities

The Department of Justice has filed a lawsuit against Uber on Thursday, alleging that the company routinely and illegally discriminates against passengers with physical disabilities.

The lawsuit, filed in federal court in San Francisco, alleges that Uber’s drivers regularly refuse service to passengers with service animals and stowable wheelchairs. Some passengers are charged cleaning fees for service animals and cancellation fees after being refused a ride, the lawsuit alleges. According to the complaint, others are insulted or denied requests like sitting in the front seat due to mobility issues.

“Ubers discriminatory conduct has caused significant economic, emotional, and physical harm to individuals with disabilities,” the lawsuit reads.

A survey last year by the organization Guide Dogs for the Blind found that more than 83% of people who are blind or visually impaired said they’ve been denied ride-share service.

In a statement to Bloomberg, Uber disagreed with the lawsuit, saying it has a “zero-tolerance policy for confirmed service denials.”

power

Draft Senate bill gives AI companies a two-year pass on federal regulation, Bloomberg reports

Bloomberg reports that a draft bill from Senator Ted Cruz would give AI companies a two-year pass from any federal regulation when they apply to be part of a White House-controlled “regulatory sandbox.” Such a regulatory framework frees participating companies from federal agency oversight while simultaneously handing President Trump broad powers to shape a still nascent and increasingly powerful industry.

The draft bill allows companies approved for the waiver to request renewals for up to eight years, according to the report.

The fast-moving generative-AI boom that took the tech world by storm was kicked off by the release of OpenAI’s ChatGPT less than three years ago. A potential decade free of federal regulations would be a huge win for companies like Meta, Google, OpenAI, and Amazon.

In July, the US Senate voted 99-1 to kill a planned provision from President Trump’s massive tax bill that would have prevented any state from regulating AI for 10 years.

The fast-moving generative-AI boom that took the tech world by storm was kicked off by the release of OpenAI’s ChatGPT less than three years ago. A potential decade free of federal regulations would be a huge win for companies like Meta, Google, OpenAI, and Amazon.

In July, the US Senate voted 99-1 to kill a planned provision from President Trump’s massive tax bill that would have prevented any state from regulating AI for 10 years.

Latest Stories

Sherwood Media, LLC produces fresh and unique perspectives on topical financial news and is a fully owned subsidiary of Robinhood Markets, Inc., and any views expressed here do not necessarily reflect the views of any other Robinhood affiliate, including Robinhood Markets, Inc., Robinhood Financial LLC, Robinhood Securities, LLC, Robinhood Crypto, LLC, or Robinhood Money, LLC.