What to Know About Elon Musk's New AI Company, xAI

Elon Musk
Elon Musk

Chief Executive Officer of SpaceX and Tesla and owner of Twitter, Elon Musk attends the Viva Technology conference on June 16, 2023 in Paris, France. Credit - Chesnot/Getty Images

Elon Musk wants to “understand the true nature of the universe.” At least that’s what his new AI company, xAI, said on its website as he announced its formation on Wednesday.

Musk incorporated xAI in Nevada in March this year and reportedly purchased “roughly 10,000 graphics processing units”—hardware that is required to develop and run state-of-the-art AI systems. The company has not said how it is financed but the Financial Times reported in April that Musk was discussing getting funding from investors in SpaceX and Tesla, two companies he runs.

The company has not shared much detail about its intentions, but said on its website that its team would be joining a Twitter Spaces call on July 14 to take questions. (Musk is the owner of Twitter.) The AI company will work closely with Twitter, now called X Corp., and Tesla, as well as other companies “to make progress towards our mission,” xAI said on its website.


More from TIME


xAI and Musk’s history with artificial intelligence

The team at xAI, led by Musk, includes former employees of prominent AI companies OpenAI and DeepMind, as well as Microsoft and Tesla. Dan Hendrycks, director of the Center for AI Safety, is listed as an adviser. The Center for AI Safety has been highly vocal about safety concerns. The organization released a statement signed by hundreds of AI scientists and experts, as well as the CEOs of some of the top AI companies in May, which said that “mitigating the risk of extinction from AI should be a global priority alongside other societal-scale risks such as pandemics and nuclear war.” Demis Hassabis, CEO of DeepMind, Sam Altman CEO of OpenAI, and Dario Amodei, CEO of Anthropic were among the signatories.

Musk did not sign the Centre for AI Safety statement, but he did sign an open letter published in March by the Future of Life Institute, which called on AI companies to “immediately pause for at least 6 months the training of AI systems more powerful than GPT-4.

Musk was one of the founding chairs of OpenAI, along with Altman, OpenAI’s CEO. Musk was among a group of investors, including Altman, Greg Brockman, Reid Hoffman, Jessica Livingston, Peter Thiel, Amazon Web Services, Infosys, and YC Research, who committed $1 billion to fund OpenAI in 2015. Musk has said he contributed $100 million of this $1 billion, but has also put the figure at half that.

Read More: A Brief History of Elon Musk Saying One Thing and Doing Another at Twitter

The circumstances of Musk’s departure aren’t entirely clear. According to an OpenAI blog post and later tweets by Musk, he left OpenAI in 2018 to prevent conflicts of interest as Tesla became more focused on AI. Semafor reported in March that Musk had proposed that he assume leadership of OpenAI, and walked away after his proposal was rejected. The Financial Times reported in April that Musk’s departure was also due to clashes with other board members and staff over OpenAI’s approach to AI safety.

Since he left in 2018, Musk has been critical of the direction OpenAI has taken. In an interview with Tucker Carlson on Fox News in April, Musk said of the company “they are now closed source, and they are obviously for-profit, and closely allied with Microsoft.” Microsoft and OpenAI have a multibillion-dollar partnership, which grants OpenAI use of Microsoft’s cloud computing capacity in exchange for OpenAI’s artificial intelligence systems powering Microsoft’s products.

In March, Musk tweeted “I’m still confused as to how a non-profit to which I donated ~$100M somehow became a $30B market cap for-profit. If this is legal, why doesn’t everyone do it?” OpenAI has previously said that it moved from non-profit status to a “hybrid of a for-profit and nonprofit” because the computational demands of training state-of-the-art AI systems mean that OpenAI would need to raise more funding than would be available to a standard non-profit.

In the April interview with Carlson, Musk also said he was worried that AI models are being trained to be “politically correct,” and pledged to create a “Truth GPT,” which he said would be a “maximum truth-seeking AI.”

Read More: Elon Musk Is Bringing the Culture Wars to AI

Musk, who has frequently spoken about the risks posed by advanced AI systems in the past, also said in the interview that he was founding a new AI organization to prevent “AI dystopia.” However, experts and researchers, including xAI adviser Hendrycks, have expressed concerns that the addition of another highly-resourced company to the AI ecosystem could further fuel the race to develop powerful AI systems, at the expense of efforts to make them safe.

Read more: The AI Arms Race Is Changing Everything

In April, in response to reports that Musk would likely start a new AI company, Hendrycks wrote that the “emergence of a new major AI developer will likely increase competitive pressures,” and that the “pressures to race ahead can cause actors to make sacrifices, especially when there are tradeoffs between safety and competitiveness.”