• Leopold Aschenbrenner spoke about his firing from OpenAI's superalignment team in a podcast. 
  • He said HR warned him after he shared a memo about OpenAI's security with two board members.
  • He said his firing was related to sharing a brainstorming doc with outside researchers. 

A former OpenAI researcher opened up about how he "ruffled some feathers" by writing and sharing some documents related to safety at the company, and was eventually fired.

Leopold Aschenbrenner, who graduated from Columbia University at 19, according to his LinkedIn, worked on OpenAI's superalignment team before he was reportedly "fired for leaking" in April. He spoke out about the experience in a recent interview with podcaster Dwarkesh Patel released Tuesday.

Aschenbrenner said he wrote and shared a memo after a "major security incident" that he didn't specify in the interview, and shared it with a couple of OpenAI board members. In the memo, he wrote that the company's security was "egregiously insufficient" in protecting against the theft of "key algorithmic secrets from foreign actors," Aschenbrenner said. The AI researcher previously shared the memo with others at OpenAI, "who mostly said it was helpful," he added.

HR later gave him a warning about the memo, Aschenbrenner said, telling him that it was "racist" and "unconstructive" to worry about China Communist Party espionage. An OpenAI lawyer later asked him about his views on AI and AGI and whether Aschenbrenner and the superalignment team were "loyal to the company," as the AI researcher put it.

Aschenbrenner claimed the company then went through his OpenAI digital artifacts.

He was fired shortly after, he said, with the company alleging he had leaked confidential information, wasn't forthcoming in its investigation, and referenced his prior warning from HR after sharing the memo with the board members.

Aschenbrenner said the leak in question referred to a "brainstorming document on preparedness, on safety, and security measures" needed for artificial general intelligence, or AGI, that he shared with three external researchers for feedback. He said he had reviewed the document before sharing it for any sensitive information and that it was "totally normal" at the company to share this kind of information for feedback.

Aschenbrenner said OpenAI deemed a line about "planning for AGI by 2027-2028 and not setting timelines for preparedness" as confidential. He said he wrote the document a couple of months after the superalignment team was announced, which referenced a four-year planning horizon.

In its announcement of the superalignment team posted in July 2023, OpenAI said its goal was to "solve the core technical challenges of superintelligence alignment in four years."

"I didn't think that planning horizon was sensitive," Aschenbrenner said in the interview. "You know it's the sort of thing Sam says publicly all the time," he said, referring to CEO Sam Altman.

An OpenAI spokesperson told Business Insider that the concerns Aschenbrenner raised internally and to its Board of Directors "did not lead to his separation."

"While we share his commitment to building safe AGI, we disagree with many of the claims he has since made about our work," the OpenAI spokesperson said.

Aschenbrenner is one of several former employees who have recently spoken out about safety concerns at OpenAI. Most recently, a group of nine current and former OpenAI employees signed a letter calling for more transparency in AI companies and protection for those who express concern about the technology.

Read the original article on Business Insider