×
The ‘Genesis Mission’: Here’s What’s in Trump’s Most Grandiose AI Executive Order Yet

The ‘Genesis Mission’: Here’s What’s in Trump’s Most Grandiose AI Executive Order Yet

The title of the executive order is on the short side for Trump: “Launching the Genesis Mission.”

It reads in part:

In this pivotal moment, the challenges we face require a historic national effort, comparable in urgency and ambition to the Manhattan Project that was instrumental to our victory in World War II and was a critical basis for the foundation of the Department of Energy (DOE) and its national laboratories.

According to Michael Kratsios, the science advisor to the president, that Manhattan Project comparison is just the beginning. The Genesis Mission is also, we’re being told, “the largest marshaling of federal scientific resources since the Apollo program.”

Then again, the Trump administration says stuff. The president said nuclear weapons tests were going to begin “immediately,” and that was almost a month ago.

But consulting AI.gov, Trump’s special fan page for showing off his love of AI, I find that the president has nine marquee AI executive orders, stretching back to his previous administration, and they have titles like “Promoting the Export of the American AI Technology Stack,” and “Preventing Woke AI in the Federal Government.”

None of them sound nearly as hauntingly mysterious as a “Genesis Mission.” What’s this AI-loving president up to now?

What the “Genesis Mission” is literally supposed to be:

We’re being promised a sort of AI and automation super-platform for the federal government. Based on my read of the program laid out in this order, the Secretary of Energy—fracking mogul Chris Wright—is supposed to unify all Department of Energy datasets with those of all federal agencies, and use those to create “scientific foundation models.” Presumably that means the government’s own LLMs, or other LXMs used for scientific research.

Then our federal government is going to use its new AI models to build programs that “automate research workflows, and accelerate scientific breakthroughs.” We’re getting set-it-and-forget-it federal science, in other words. The AI does the research, and a person can just come along and scoop up the breakthroughs like cream from a milk bucket.

According to Politico, Wright says there will be an “incredible increase in the pace of scientific discovery and innovation.” They’re looking at nuclear fusion, other energy sources, pharmaceuticals, protein folding—all the areas of science and research that pair well with AI hype.

And what does the plan specifically entail?

The executive order does, in all fairness, outline what the next year (and beyond) is supposed to look like for this program to some degree.

By the 60-day mark: A list.

America gets a document identifying 20 core science “challenges” the Genesis Mission can solve.

By the 90-day mark: An inventory.

America is gifted an inventory of computational resources the Genesis Mission can use to build its system.

By the 120-day mark: A plan.

By now, the Mission is supposed to have its data optimized and in place to train the models.

By the 240-day mark: Another inventory.

Wright is supposed to have figured out where robot-driven, automated science experiments can be done. Since it probably sounds like I’m joking, here’s what the order says exactly:

“Within 240 days of the date of this order, the Secretary shall review capabilities across the DOE national laboratories and other participating Federal research facilities for robotic laboratories and production facilities with the ability to engage in AI-directed experimentation and manufacturing, including automated and AI-augmented workflows and the related technical and operational standards needed.”

By the 270 day mark: A demo.

We get some sort of proof of concept for the Genesis Mission platform, focused on one of the 20 aforementioned challenges.

Within one year (and then every year from now on): An evaluation.

Were positive outcomes achieved? Did the Genesis Mission make scientific discoveries? How’s everything going? It’ll all be in the annual report.

And the Genesis Mission had better work, because the other side of this effort is a bunch of federal funding cuts for science. This administration has sought to cancel federal funding for (and subscriptions to) science journals. It has sought to cut $783 million in funding for health research—cuts that, it appears, really will go into effect. It has sought to cut off funding to no less than 100 climate change studies. It has reduced research spending at the National Oceanic and Atmospheric Administration by $100 million, and on, and on.

The cuts may have had many aims, at least one of which was to curb DEI (remember when people used to talk about DEI?). Another one, it seems, is to shift science into the realm of things you can just automate, well before sufficient AI systems exist to justify anyone’s confidence that such a thing is possible.

So buckle up for automated, low-cost scientific breakthroughs, everyone! They’ll be here soon, thanks to the Chris Wright and the Genesis Mission. Otherwise, some of those funding cuts might start to look a little silly in retrospect.

Source link
#Genesis #Mission #Heres #Whats #Trumps #Grandiose #Executive #Order

After two weeks of hearing from assorted witnesses that he was a lying snake, the jury finally heard from the lying snake himself: Sam Altman. At the end of the testimony, his lawyer William Savitt asked him how it felt to be accused of stealing a charity.

“We created, through a ton of hard work, this extremely large charity, and I agree you can’t steal it,” Altman said. “Mr. Musk did try to kill it, I guess. Twice.”

Altman was fully in “nice kid from St. Louis” mode, and did a passable impression of a man who was bewildered at what was happening to him. When he stepped down from the stand holding a stack of evidence binders, he even looked a little like a schoolboy. He seemed nervous at the beginning of his direct testimony, though he warmed up fairly quickly. Overall, he seemed to give credible testimony — and at times, it seemed like the jury liked him.

Throughout this trial I’ve had some difficulty imagining what the jury is making of all this because I am a little too familiar with the figures who are testifying. I have heard some audacious lies under oath, like when Elon Musk told us all he doesn’t lose his temper. (He then proceeded to lose his temper on cross-examination.) Or like when Shivon Zilis, the mother of several of his children, told us that she didn’t know Musk was starting xAI — which seemed to be directly contradicted by her text messages. Or when Greg “What will take me to $1B?” Brockman told us he was all about the mission. I certainly believe Altman isn’t trustworthy — I mean, The New Yorker published more than 17,000 words about how much he lies. But unlike with Musk, there are contemporaneous documents backing Altman’s version of the story. At least, mostly.

“My belief is he wanted to have long-term control”

After OpenAI’s Dota 2 win, discussions for a for-profit arm started in earnest. “Mr. Musk felt very strongly that if we were going to form a for-profit he needed to have total control over it initially,” Altman said. “He only trusted himself to make non-obvious decisions that were going to turn out to be correct.”

Altman testified that he was uncomfortable with Musk’s insistence on control, not just because Musk hadn’t been as involved as everyone else, but because OpenAI existed so no one person would control AGI. And at Y Combinator, the startup incubator where he was president, Altman had seen a lot of control fights; no one wanted to give up power when things were going well. With structures like supervoting shares, founders could retain control forever. Curiously, Altman’s example was not the most famous one (Mark Zuckerberg at Meta); it was Musk and SpaceX. When Altman asked Musk about succession plans for OpenAI, he got a particularly “hair-raising” answer: In the event of Musk’s death, Musk said, “I haven’t thought about it a ton, but maybe control should pass to my children.”

I don’t know about that. But I do know that I saw a 2017 email from Altman to Zilis in which he wrote, “I am worried about control. I don’t think any one person should have control of the world’s first AGI — in fact the whole reason we started OpenAI was so that wouldn’t happen.” He went on to say that he didn’t mind the idea of immediate control and was open to “creative structures” — which I understood to mean that, in order to placate Musk, Altman was willing to give him control up to specific milestones in company development.

“I read a vague, like, a lightweight threat in there”

“My belief is he wanted to have long-term control and that he would’ve had that had we agreed to the structure he wanted,” Altman said on the stand. This sounds basically right. In later video testimony from Sam Teller’s deposition, we heard that Musk no longer invests in anything he doesn’t control. This also fits with Musk’s long-term fixation on making sure he can’t get booted from his own company the way he got booted from PayPal.

Musk also tried to recruit Altman to Tesla. We saw texts between Altman and Teller, in which Teller told Altman that Musk was committed to beefing up Tesla’s AI no matter what, and that he hoped that Altman, Brockman, and Ilya Sutskever would want to join eventually. “I read a vague, like, a lightweight threat in there, that he’s gonna do this inside of Tesla with or without you,” Altman said. But he felt that Tesla was primarily a car company — allowing it to acquire OpenAI would betray OpenAI’s mission.

Later, in Teller’s testimony, we saw texts Teller sent to Zilis at 12:40AM on February 4th, 2018: “I don’t love OpenAI continuing without Elon,” he wrote. “Would rather disable it by recruiting the leaders.”

When Musk stopped his quarterly donations, OpenAI was operating on a “shoestring” with an “extremely short runway of cash.” OpenAI did have other donors, none of whom have sued it or joined Musk’s suit. (One donor in the exhibit that wasn’t called out to the courtroom was Alameda Research, the firm owned by Sam Bankman-Fried, who is now in prison for fraud and money laundering.) Musk’s resignation from the board meant “people wondered if he was gonna try to take, uh, vengeance out on us or something.” On the other hand, Altman said Musk had “demotivated some of our key researchers” and done “huge damage for a long time to the culture of the organization.” So it sure seems like some people were relieved to be rid of him.

I’ve seen some fairly shoddy lawyering from Musk’s side throughout this trial

We saw a lot of evidence that throughout the time Altman was setting up OpenAI’s for-profit arm, he kept Musk apprised of what was going on, either directly or through Zilis or Teller. At no point did Musk object, and whatever he said publicly about the Microsoft investments, there was plenty of evidence that privately he’d been made aware.

On the cross-examination, we were treated to more than 10 minutes of Steven Molo telling Altman that various and assorted people had called him a liar: Sutskever, Mira Murati, Helen Toner, Tasha McCauley, Daniela and Dario Amodei (former OpenAI employees and founders of Anthropic), employees at Altman’s first startup Loopt, that recent New Yorker article, a book called The Optimist, etc. Molo did score some points by asking Altman about testimony in the trial, which Altman said he wasn’t paying close attention to. Molo acted as though this was inconceivable. Surely someone had informed Altman of what was said?

It was a little funny and also a little tiresome. Altman kept his cool, though, seeming hurt and confused by the focus on whether he was a liar. It was also the most successful part of the cross, which declined in focus precipitously afterward. I’ve seen some fairly shoddy lawyering from Musk’s side throughout this trial, and today was pretty bad. At one point, when Molo was trying to capitalize on Altman being both CEO and on the company’s board, Altman said — truthfully — that CEOs are almost always on the boards of the companies they run.

(At this point in my notes, I had written, “Boy, Molo is not very good at this.”)

The point of this trial isn’t to win — it’s to punish Altman, Brockman, and OpenAI

There was also an unconvincing argument about fundraising in nonprofits, specifically that if Stanford could raise $3 billion a year, OpenAI should have remained a nonprofit. Okay, let’s just think about that for a minute. Stanford has a donor network of thousands of graduates. It’s a school, which has very different capital requirements. It is not competing with any reputable for-profit companies. But leave that all aside and assume that some fundraising genius took over at the OpenAI Foundation: $3 billion is the initial two Microsoft investments combined, and not enough to scale OpenAI to where it is now. If compute is the main bottleneck on building AI models, then Molo’s line of argument suggests OpenAI never would have managed to be successful as a nonprofit alone. He’s making the defense’s case for them.

But the thing is, Molo doesn’t actually have to be good at this job, because the point of this trial isn’t to win — though I’m sure Musk wouldn’t mind a win. The point is to punish Altman, Brockman, and OpenAI. Musk has done that pretty thoroughly — reinforcing in the public’s mind that Altman is a liar and a snake. This morning, I read an exclusive in The Wall Street Journal that assorted Republican AGs and the House Oversight committee wanted to look into Sam Altman’s investments. References to the trial are peppered throughout the article.

So yes, Altman was convincing on the stand. He may even win the suit. But it sure seems like Musk’s vengeance has just begun.

Follow topics and authors from this story to see more like this in your personalized homepage feed and to receive email updates.
#Sam #Altman #winning #standAI,OpenAI">Sam Altman was winning on the stand, but it might not be enoughAfter two weeks of hearing from assorted witnesses that he was a lying snake, the jury finally heard from the lying snake himself: Sam Altman. At the end of the testimony, his lawyer William Savitt asked him how it felt to be accused of stealing a charity.“We created, through a ton of hard work, this extremely large charity, and I agree you can’t steal it,” Altman said. “Mr. Musk did try to kill it, I guess. Twice.”Altman was fully in “nice kid from St. Louis” mode, and did a passable impression of a man who was bewildered at what was happening to him. When he stepped down from the stand holding a stack of evidence binders, he even looked a little like a schoolboy. He seemed nervous at the beginning of his direct testimony, though he warmed up fairly quickly. Overall, he seemed to give credible testimony — and at times, it seemed like the jury liked him.Throughout this trial I’ve had some difficulty imagining what the jury is making of all this because I am a little too familiar with the figures who are testifying. I have heard some audacious lies under oath, like when Elon Musk told us all he doesn’t lose his temper. (He then proceeded to lose his temper on cross-examination.) Or like when Shivon Zilis, the mother of several of his children, told us that she didn’t know Musk was starting xAI — which seemed to be directly contradicted by her text messages. Or when Greg “What will take me to B?” Brockman told us he was all about the mission. I certainly believe Altman isn’t trustworthy — I mean, The New Yorker published more than 17,000 words about how much he lies. But unlike with Musk, there are contemporaneous documents backing Altman’s version of the story. At least, mostly.“My belief is he wanted to have long-term control”After OpenAI’s Dota 2 win, discussions for a for-profit arm started in earnest. “Mr. Musk felt very strongly that if we were going to form a for-profit he needed to have total control over it initially,” Altman said. “He only trusted himself to make non-obvious decisions that were going to turn out to be correct.”Altman testified that he was uncomfortable with Musk’s insistence on control, not just because Musk hadn’t been as involved as everyone else, but because OpenAI existed so no one person would control AGI. And at Y Combinator, the startup incubator where he was president, Altman had seen a lot of control fights; no one wanted to give up power when things were going well. With structures like supervoting shares, founders could retain control forever. Curiously, Altman’s example was not the most famous one (Mark Zuckerberg at Meta); it was Musk and SpaceX. When Altman asked Musk about succession plans for OpenAI, he got a particularly “hair-raising” answer: In the event of Musk’s death, Musk said, “I haven’t thought about it a ton, but maybe control should pass to my children.”I don’t know about that. But I do know that I saw a 2017 email from Altman to Zilis in which he wrote, “I am worried about control. I don’t think any one person should have control of the world’s first AGI — in fact the whole reason we started OpenAI was so that wouldn’t happen.” He went on to say that he didn’t mind the idea of immediate control and was open to “creative structures” — which I understood to mean that, in order to placate Musk, Altman was willing to give him control up to specific milestones in company development.“I read a vague, like, a lightweight threat in there”“My belief is he wanted to have long-term control and that he would’ve had that had we agreed to the structure he wanted,” Altman said on the stand. This sounds basically right. In later video testimony from Sam Teller’s deposition, we heard that Musk no longer invests in anything he doesn’t control. This also fits with Musk’s long-term fixation on making sure he can’t get booted from his own company the way he got booted from PayPal.Musk also tried to recruit Altman to Tesla. We saw texts between Altman and Teller, in which Teller told Altman that Musk was committed to beefing up Tesla’s AI no matter what, and that he hoped that Altman, Brockman, and Ilya Sutskever would want to join eventually. “I read a vague, like, a lightweight threat in there, that he’s gonna do this inside of Tesla with or without you,” Altman said. But he felt that Tesla was primarily a car company — allowing it to acquire OpenAI would betray OpenAI’s mission.Later, in Teller’s testimony, we saw texts Teller sent to Zilis at 12:40AM on February 4th, 2018: “I don’t love OpenAI continuing without Elon,” he wrote. “Would rather disable it by recruiting the leaders.”When Musk stopped his quarterly donations, OpenAI was operating on a “shoestring” with an “extremely short runway of cash.” OpenAI did have other donors, none of whom have sued it or joined Musk’s suit. (One donor in the exhibit that wasn’t called out to the courtroom was Alameda Research, the firm owned by Sam Bankman-Fried, who is now in prison for fraud and money laundering.) Musk’s resignation from the board meant “people wondered if he was gonna try to take, uh, vengeance out on us or something.” On the other hand, Altman said Musk had “demotivated some of our key researchers” and done “huge damage for a long time to the culture of the organization.” So it sure seems like some people were relieved to be rid of him.I’ve seen some fairly shoddy lawyering from Musk’s side throughout this trialWe saw a lot of evidence that throughout the time Altman was setting up OpenAI’s for-profit arm, he kept Musk apprised of what was going on, either directly or through Zilis or Teller. At no point did Musk object, and whatever he said publicly about the Microsoft investments, there was plenty of evidence that privately he’d been made aware.On the cross-examination, we were treated to more than 10 minutes of Steven Molo telling Altman that various and assorted people had called him a liar: Sutskever, Mira Murati, Helen Toner, Tasha McCauley, Daniela and Dario Amodei (former OpenAI employees and founders of Anthropic), employees at Altman’s first startup Loopt, that recent New Yorker article, a book called The Optimist, etc. Molo did score some points by asking Altman about testimony in the trial, which Altman said he wasn’t paying close attention to. Molo acted as though this was inconceivable. Surely someone had informed Altman of what was said?It was a little funny and also a little tiresome. Altman kept his cool, though, seeming hurt and confused by the focus on whether he was a liar. It was also the most successful part of the cross, which declined in focus precipitously afterward. I’ve seen some fairly shoddy lawyering from Musk’s side throughout this trial, and today was pretty bad. At one point, when Molo was trying to capitalize on Altman being both CEO and on the company’s board, Altman said — truthfully — that CEOs are almost always on the boards of the companies they run.(At this point in my notes, I had written, “Boy, Molo is not very good at this.”)The point of this trial isn’t to win — it’s to punish Altman, Brockman, and OpenAIThere was also an unconvincing argument about fundraising in nonprofits, specifically that if Stanford could raise  billion a year, OpenAI should have remained a nonprofit. Okay, let’s just think about that for a minute. Stanford has a donor network of thousands of graduates. It’s a school, which has very different capital requirements. It is not competing with any reputable for-profit companies. But leave that all aside and assume that some fundraising genius took over at the OpenAI Foundation:  billion is the initial two Microsoft investments combined, and not enough to scale OpenAI to where it is now. If compute is the main bottleneck on building AI models, then Molo’s line of argument suggests OpenAI never would have managed to be successful as a nonprofit alone. He’s making the defense’s case for them.But the thing is, Molo doesn’t actually have to be good at this job, because the point of this trial isn’t to win — though I’m sure Musk wouldn’t mind a win. The point is to punish Altman, Brockman, and OpenAI. Musk has done that pretty thoroughly — reinforcing in the public’s mind that Altman is a liar and a snake. This morning, I read an exclusive in The Wall Street Journal that assorted Republican AGs and the House Oversight committee wanted to look into Sam Altman’s investments. References to the trial are peppered throughout the article.So yes, Altman was convincing on the stand. He may even win the suit. But it sure seems like Musk’s vengeance has just begun.Follow topics and authors from this story to see more like this in your personalized homepage feed and to receive email updates.Elizabeth LopattoCloseElizabeth LopattoPosts from this author will be added to your daily email digest and your homepage feed.FollowFollowSee All by Elizabeth LopattoAICloseAIPosts from this topic will be added to your daily email digest and your homepage feed.FollowFollowSee All AIOpenAICloseOpenAIPosts from this topic will be added to your daily email digest and your homepage feed.FollowFollowSee All OpenAI#Sam #Altman #winning #standAI,OpenAI

Elon Musk told us all he doesn’t lose his temper. (He then proceeded to lose his temper on cross-examination.) Or like when Shivon Zilis, the mother of several of his children, told us that she didn’t know Musk was starting xAI — which seemed to be directly contradicted by her text messages. Or when Greg “What will take me to $1B?” Brockman told us he was all about the mission. I certainly believe Altman isn’t trustworthy — I mean, The New Yorker published more than 17,000 words about how much he lies. But unlike with Musk, there are contemporaneous documents backing Altman’s version of the story. At least, mostly.

“My belief is he wanted to have long-term control”

After OpenAI’s Dota 2 win, discussions for a for-profit arm started in earnest. “Mr. Musk felt very strongly that if we were going to form a for-profit he needed to have total control over it initially,” Altman said. “He only trusted himself to make non-obvious decisions that were going to turn out to be correct.”

Altman testified that he was uncomfortable with Musk’s insistence on control, not just because Musk hadn’t been as involved as everyone else, but because OpenAI existed so no one person would control AGI. And at Y Combinator, the startup incubator where he was president, Altman had seen a lot of control fights; no one wanted to give up power when things were going well. With structures like supervoting shares, founders could retain control forever. Curiously, Altman’s example was not the most famous one (Mark Zuckerberg at Meta); it was Musk and SpaceX. When Altman asked Musk about succession plans for OpenAI, he got a particularly “hair-raising” answer: In the event of Musk’s death, Musk said, “I haven’t thought about it a ton, but maybe control should pass to my children.”

I don’t know about that. But I do know that I saw a 2017 email from Altman to Zilis in which he wrote, “I am worried about control. I don’t think any one person should have control of the world’s first AGI — in fact the whole reason we started OpenAI was so that wouldn’t happen.” He went on to say that he didn’t mind the idea of immediate control and was open to “creative structures” — which I understood to mean that, in order to placate Musk, Altman was willing to give him control up to specific milestones in company development.

“I read a vague, like, a lightweight threat in there”

“My belief is he wanted to have long-term control and that he would’ve had that had we agreed to the structure he wanted,” Altman said on the stand. This sounds basically right. In later video testimony from Sam Teller’s deposition, we heard that Musk no longer invests in anything he doesn’t control. This also fits with Musk’s long-term fixation on making sure he can’t get booted from his own company the way he got booted from PayPal.

Musk also tried to recruit Altman to Tesla. We saw texts between Altman and Teller, in which Teller told Altman that Musk was committed to beefing up Tesla’s AI no matter what, and that he hoped that Altman, Brockman, and Ilya Sutskever would want to join eventually. “I read a vague, like, a lightweight threat in there, that he’s gonna do this inside of Tesla with or without you,” Altman said. But he felt that Tesla was primarily a car company — allowing it to acquire OpenAI would betray OpenAI’s mission.

Later, in Teller’s testimony, we saw texts Teller sent to Zilis at 12:40AM on February 4th, 2018: “I don’t love OpenAI continuing without Elon,” he wrote. “Would rather disable it by recruiting the leaders.”

When Musk stopped his quarterly donations, OpenAI was operating on a “shoestring” with an “extremely short runway of cash.” OpenAI did have other donors, none of whom have sued it or joined Musk’s suit. (One donor in the exhibit that wasn’t called out to the courtroom was Alameda Research, the firm owned by Sam Bankman-Fried, who is now in prison for fraud and money laundering.) Musk’s resignation from the board meant “people wondered if he was gonna try to take, uh, vengeance out on us or something.” On the other hand, Altman said Musk had “demotivated some of our key researchers” and done “huge damage for a long time to the culture of the organization.” So it sure seems like some people were relieved to be rid of him.

I’ve seen some fairly shoddy lawyering from Musk’s side throughout this trial

We saw a lot of evidence that throughout the time Altman was setting up OpenAI’s for-profit arm, he kept Musk apprised of what was going on, either directly or through Zilis or Teller. At no point did Musk object, and whatever he said publicly about the Microsoft investments, there was plenty of evidence that privately he’d been made aware.

On the cross-examination, we were treated to more than 10 minutes of Steven Molo telling Altman that various and assorted people had called him a liar: Sutskever, Mira Murati, Helen Toner, Tasha McCauley, Daniela and Dario Amodei (former OpenAI employees and founders of Anthropic), employees at Altman’s first startup Loopt, that recent New Yorker article, a book called The Optimist, etc. Molo did score some points by asking Altman about testimony in the trial, which Altman said he wasn’t paying close attention to. Molo acted as though this was inconceivable. Surely someone had informed Altman of what was said?

It was a little funny and also a little tiresome. Altman kept his cool, though, seeming hurt and confused by the focus on whether he was a liar. It was also the most successful part of the cross, which declined in focus precipitously afterward. I’ve seen some fairly shoddy lawyering from Musk’s side throughout this trial, and today was pretty bad. At one point, when Molo was trying to capitalize on Altman being both CEO and on the company’s board, Altman said — truthfully — that CEOs are almost always on the boards of the companies they run.

(At this point in my notes, I had written, “Boy, Molo is not very good at this.”)

The point of this trial isn’t to win — it’s to punish Altman, Brockman, and OpenAI

There was also an unconvincing argument about fundraising in nonprofits, specifically that if Stanford could raise $3 billion a year, OpenAI should have remained a nonprofit. Okay, let’s just think about that for a minute. Stanford has a donor network of thousands of graduates. It’s a school, which has very different capital requirements. It is not competing with any reputable for-profit companies. But leave that all aside and assume that some fundraising genius took over at the OpenAI Foundation: $3 billion is the initial two Microsoft investments combined, and not enough to scale OpenAI to where it is now. If compute is the main bottleneck on building AI models, then Molo’s line of argument suggests OpenAI never would have managed to be successful as a nonprofit alone. He’s making the defense’s case for them.

But the thing is, Molo doesn’t actually have to be good at this job, because the point of this trial isn’t to win — though I’m sure Musk wouldn’t mind a win. The point is to punish Altman, Brockman, and OpenAI. Musk has done that pretty thoroughly — reinforcing in the public’s mind that Altman is a liar and a snake. This morning, I read an exclusive in The Wall Street Journal that assorted Republican AGs and the House Oversight committee wanted to look into Sam Altman’s investments. References to the trial are peppered throughout the article.

So yes, Altman was convincing on the stand. He may even win the suit. But it sure seems like Musk’s vengeance has just begun.

Follow topics and authors from this story to see more like this in your personalized homepage feed and to receive email updates.

#Sam #Altman #winning #standAI,OpenAI">Sam Altman was winning on the stand, but it might not be enough

After two weeks of hearing from assorted witnesses that he was a lying snake, the jury finally heard from the lying snake himself: Sam Altman. At the end of the testimony, his lawyer William Savitt asked him how it felt to be accused of stealing a charity.

“We created, through a ton of hard work, this extremely large charity, and I agree you can’t steal it,” Altman said. “Mr. Musk did try to kill it, I guess. Twice.”

Altman was fully in “nice kid from St. Louis” mode, and did a passable impression of a man who was bewildered at what was happening to him. When he stepped down from the stand holding a stack of evidence binders, he even looked a little like a schoolboy. He seemed nervous at the beginning of his direct testimony, though he warmed up fairly quickly. Overall, he seemed to give credible testimony — and at times, it seemed like the jury liked him.

Throughout this trial I’ve had some difficulty imagining what the jury is making of all this because I am a little too familiar with the figures who are testifying. I have heard some audacious lies under oath, like when Elon Musk told us all he doesn’t lose his temper. (He then proceeded to lose his temper on cross-examination.) Or like when Shivon Zilis, the mother of several of his children, told us that she didn’t know Musk was starting xAI — which seemed to be directly contradicted by her text messages. Or when Greg “What will take me to $1B?” Brockman told us he was all about the mission. I certainly believe Altman isn’t trustworthy — I mean, The New Yorker published more than 17,000 words about how much he lies. But unlike with Musk, there are contemporaneous documents backing Altman’s version of the story. At least, mostly.

“My belief is he wanted to have long-term control”

After OpenAI’s Dota 2 win, discussions for a for-profit arm started in earnest. “Mr. Musk felt very strongly that if we were going to form a for-profit he needed to have total control over it initially,” Altman said. “He only trusted himself to make non-obvious decisions that were going to turn out to be correct.”

Altman testified that he was uncomfortable with Musk’s insistence on control, not just because Musk hadn’t been as involved as everyone else, but because OpenAI existed so no one person would control AGI. And at Y Combinator, the startup incubator where he was president, Altman had seen a lot of control fights; no one wanted to give up power when things were going well. With structures like supervoting shares, founders could retain control forever. Curiously, Altman’s example was not the most famous one (Mark Zuckerberg at Meta); it was Musk and SpaceX. When Altman asked Musk about succession plans for OpenAI, he got a particularly “hair-raising” answer: In the event of Musk’s death, Musk said, “I haven’t thought about it a ton, but maybe control should pass to my children.”

I don’t know about that. But I do know that I saw a 2017 email from Altman to Zilis in which he wrote, “I am worried about control. I don’t think any one person should have control of the world’s first AGI — in fact the whole reason we started OpenAI was so that wouldn’t happen.” He went on to say that he didn’t mind the idea of immediate control and was open to “creative structures” — which I understood to mean that, in order to placate Musk, Altman was willing to give him control up to specific milestones in company development.

“I read a vague, like, a lightweight threat in there”

“My belief is he wanted to have long-term control and that he would’ve had that had we agreed to the structure he wanted,” Altman said on the stand. This sounds basically right. In later video testimony from Sam Teller’s deposition, we heard that Musk no longer invests in anything he doesn’t control. This also fits with Musk’s long-term fixation on making sure he can’t get booted from his own company the way he got booted from PayPal.

Musk also tried to recruit Altman to Tesla. We saw texts between Altman and Teller, in which Teller told Altman that Musk was committed to beefing up Tesla’s AI no matter what, and that he hoped that Altman, Brockman, and Ilya Sutskever would want to join eventually. “I read a vague, like, a lightweight threat in there, that he’s gonna do this inside of Tesla with or without you,” Altman said. But he felt that Tesla was primarily a car company — allowing it to acquire OpenAI would betray OpenAI’s mission.

Later, in Teller’s testimony, we saw texts Teller sent to Zilis at 12:40AM on February 4th, 2018: “I don’t love OpenAI continuing without Elon,” he wrote. “Would rather disable it by recruiting the leaders.”

When Musk stopped his quarterly donations, OpenAI was operating on a “shoestring” with an “extremely short runway of cash.” OpenAI did have other donors, none of whom have sued it or joined Musk’s suit. (One donor in the exhibit that wasn’t called out to the courtroom was Alameda Research, the firm owned by Sam Bankman-Fried, who is now in prison for fraud and money laundering.) Musk’s resignation from the board meant “people wondered if he was gonna try to take, uh, vengeance out on us or something.” On the other hand, Altman said Musk had “demotivated some of our key researchers” and done “huge damage for a long time to the culture of the organization.” So it sure seems like some people were relieved to be rid of him.

I’ve seen some fairly shoddy lawyering from Musk’s side throughout this trial

We saw a lot of evidence that throughout the time Altman was setting up OpenAI’s for-profit arm, he kept Musk apprised of what was going on, either directly or through Zilis or Teller. At no point did Musk object, and whatever he said publicly about the Microsoft investments, there was plenty of evidence that privately he’d been made aware.

On the cross-examination, we were treated to more than 10 minutes of Steven Molo telling Altman that various and assorted people had called him a liar: Sutskever, Mira Murati, Helen Toner, Tasha McCauley, Daniela and Dario Amodei (former OpenAI employees and founders of Anthropic), employees at Altman’s first startup Loopt, that recent New Yorker article, a book called The Optimist, etc. Molo did score some points by asking Altman about testimony in the trial, which Altman said he wasn’t paying close attention to. Molo acted as though this was inconceivable. Surely someone had informed Altman of what was said?

It was a little funny and also a little tiresome. Altman kept his cool, though, seeming hurt and confused by the focus on whether he was a liar. It was also the most successful part of the cross, which declined in focus precipitously afterward. I’ve seen some fairly shoddy lawyering from Musk’s side throughout this trial, and today was pretty bad. At one point, when Molo was trying to capitalize on Altman being both CEO and on the company’s board, Altman said — truthfully — that CEOs are almost always on the boards of the companies they run.

(At this point in my notes, I had written, “Boy, Molo is not very good at this.”)

The point of this trial isn’t to win — it’s to punish Altman, Brockman, and OpenAI

There was also an unconvincing argument about fundraising in nonprofits, specifically that if Stanford could raise $3 billion a year, OpenAI should have remained a nonprofit. Okay, let’s just think about that for a minute. Stanford has a donor network of thousands of graduates. It’s a school, which has very different capital requirements. It is not competing with any reputable for-profit companies. But leave that all aside and assume that some fundraising genius took over at the OpenAI Foundation: $3 billion is the initial two Microsoft investments combined, and not enough to scale OpenAI to where it is now. If compute is the main bottleneck on building AI models, then Molo’s line of argument suggests OpenAI never would have managed to be successful as a nonprofit alone. He’s making the defense’s case for them.

But the thing is, Molo doesn’t actually have to be good at this job, because the point of this trial isn’t to win — though I’m sure Musk wouldn’t mind a win. The point is to punish Altman, Brockman, and OpenAI. Musk has done that pretty thoroughly — reinforcing in the public’s mind that Altman is a liar and a snake. This morning, I read an exclusive in The Wall Street Journal that assorted Republican AGs and the House Oversight committee wanted to look into Sam Altman’s investments. References to the trial are peppered throughout the article.

So yes, Altman was convincing on the stand. He may even win the suit. But it sure seems like Musk’s vengeance has just begun.

Follow topics and authors from this story to see more like this in your personalized homepage feed and to receive email updates.
#Sam #Altman #winning #standAI,OpenAI
Early-stage venture firm A* on Tuesday announced a $450 million Fund III. The firm takes a generalist approach, backing companies across categories including AI applications, fintech, healthcare, and security.
The average check size for this fund will be between $3 million and $5 million, with the aim to back at least 30 startups. The capital will be deployed over the next two to three years, as with the firm’s previous funds. Limited partners include nonprofits, foundations, and endowments; Carnegie Mellon University is among the publicly named backers.

A*, founded in 2020 and run by Kevin Hartz and Bennet Siegel, previously raised a $315 million Fund II in 2024 and a $300 million Fund I in 2021. Hartz is a serial entrepreneur best known for co-founding Xoom, the international money-transfer service PayPal later acquired for $1.1 billion in 2015, and Eventbrite, the event-ticketing platform that went public in 2018. Siegel came up through Boston Consulting Group and Altamont Capital Partners before spending four years as a partner at Coatue Management.

The firm has also drawn attention for backing unusually young founders, even as the practice has become more common since. Hartz told TechCrunch last fall that close to 20% of the firm’s current portfolio involve teenage entrepreneurs. Among others of its investments, it has backed the fintech company Ramp and the AI firm Mercor.

This story was updated to clarify the name of the firm.

#Kevin #Hartzs #closed #fund #million #TechCrunchA* Capital,Kevin Hartz,Startups">Kevin Hartz’s A* just closed its third fund with 0 million | TechCrunch
Early-stage venture firm A* on Tuesday announced a 0 million Fund III. The firm takes a generalist approach, backing companies across categories including AI applications, fintech, healthcare, and security.The average check size for this fund will be between  million and  million, with the aim to back at least 30 startups. The capital will be deployed over the next two to three years, as with the firm’s previous funds. Limited partners include nonprofits, foundations, and endowments; Carnegie Mellon University is among the publicly named backers.

A*, founded in 2020 and run by Kevin Hartz and Bennet Siegel, previously raised a 5 million Fund II in 2024 and a 0 million Fund I in 2021. Hartz is a serial entrepreneur best known for co-founding Xoom, the international money-transfer service PayPal later acquired for .1 billion in 2015, and Eventbrite, the event-ticketing platform that went public in 2018. Siegel came up through Boston Consulting Group and Altamont Capital Partners before spending four years as a partner at Coatue Management.







The firm has also drawn attention for backing unusually young founders, even as the practice has become more common since. Hartz told TechCrunch last fall that close to 20% of the firm’s current portfolio involve teenage entrepreneurs. Among others of its investments, it has backed the fintech company Ramp and the AI firm Mercor.

This story was updated to clarify the name of the firm. 
#Kevin #Hartzs #closed #fund #million #TechCrunchA* Capital,Kevin Hartz,Startups

$450 million Fund III. The firm takes a generalist approach, backing companies across categories including AI applications, fintech, healthcare, and security.
The average check size for this fund will be between $3 million and $5 million, with the aim to back at least 30 startups. The capital will be deployed over the next two to three years, as with the firm’s previous funds. Limited partners include nonprofits, foundations, and endowments; Carnegie Mellon University is among the publicly named backers.

A*, founded in 2020 and run by Kevin Hartz and Bennet Siegel, previously raised a $315 million Fund II in 2024 and a $300 million Fund I in 2021. Hartz is a serial entrepreneur best known for co-founding Xoom, the international money-transfer service PayPal later acquired for $1.1 billion in 2015, and Eventbrite, the event-ticketing platform that went public in 2018. Siegel came up through Boston Consulting Group and Altamont Capital Partners before spending four years as a partner at Coatue Management.

The firm has also drawn attention for backing unusually young founders, even as the practice has become more common since. Hartz told TechCrunch last fall that close to 20% of the firm’s current portfolio involve teenage entrepreneurs. Among others of its investments, it has backed the fintech company Ramp and the AI firm Mercor.

This story was updated to clarify the name of the firm.

#Kevin #Hartzs #closed #fund #million #TechCrunchA* Capital,Kevin Hartz,Startups">Kevin Hartz’s A* just closed its third fund with $450 million | TechCrunch

Early-stage venture firm A* on Tuesday announced a $450 million Fund III. The firm takes a generalist approach, backing companies across categories including AI applications, fintech, healthcare, and security.
The average check size for this fund will be between $3 million and $5 million, with the aim to back at least 30 startups. The capital will be deployed over the next two to three years, as with the firm’s previous funds. Limited partners include nonprofits, foundations, and endowments; Carnegie Mellon University is among the publicly named backers.

A*, founded in 2020 and run by Kevin Hartz and Bennet Siegel, previously raised a $315 million Fund II in 2024 and a $300 million Fund I in 2021. Hartz is a serial entrepreneur best known for co-founding Xoom, the international money-transfer service PayPal later acquired for $1.1 billion in 2015, and Eventbrite, the event-ticketing platform that went public in 2018. Siegel came up through Boston Consulting Group and Altamont Capital Partners before spending four years as a partner at Coatue Management.

The firm has also drawn attention for backing unusually young founders, even as the practice has become more common since. Hartz told TechCrunch last fall that close to 20% of the firm’s current portfolio involve teenage entrepreneurs. Among others of its investments, it has backed the fintech company Ramp and the AI firm Mercor.

This story was updated to clarify the name of the firm.

#Kevin #Hartzs #closed #fund #million #TechCrunchA* Capital,Kevin Hartz,Startups

Post Comment