Beta Technologies shares surged more than 9% after air taxi maker Eve Air Mobility announced an up to $1 billion deal to buy motors from the Vermont-based company.
Eve, which was started by Brazilian airplane maker Embraer and is now under Eve Holding, said the manufacturing deal could equal as much as $1 billion over 10 years. The Florida-based company said it has a backlog of 2,800 vehicles.
Shares of Eve Holding gained 14%.
Eve CEO Johann Bordais called the deal a “pivotal milestone” in the advancement of the company’s electric vertical takeoff and landing, or eVTOL, technology.
“Their electric motor technology will play a critical role in powering our aircraft during cruise, supporting the maturity of our propulsion architecture as we progress toward entry into service,” he said in a release.
Amazon’s cloud unit on Tuesday announced AI-enabled software designed to help clients better understand and recover from outages.
DevOps Agent, as the artificial intelligence tool from Amazon Web Services is called, predicts the cause of technical hiccups using input from third-party tools such as Datadog and Dynatrace. AWS said customers can sign up to use the tool Tuesday in a preview, before Amazon starts charging for the service.
The AI outage tool from AWS is intended to help companies more quickly figure out what caused an outage and implement fixes, Swami Sivasubramanian, vice president of agentic AI at AWS, told CNBC. It’s what site reliability engineers, or SREs, do at many companies that provide online services.
SREs try to prevent downtime and jump into action during live incidents. Startups such as Resolve and Traversal have started marketing AI assistants for these experts. Microsoft’s Azure cloud group introduced an SRE Agent in May.
Rather than waiting for on-call staff members to figure out what happened, the AWS DevOps Agent automatically assigns work to agents that look into different hypotheses, Sivasubramanian said.
“By the time the on-call ops team member dials in, they have an incident report with preliminary investigation of what could be the likely outcome, and then suggest what could be the remediation as well,” Sivasubramanian told CNBC ahead of AWS’ Reinvent conference in Las Vegas this week.
Commonwealth Bank of Australia has tested the AWS DevOps Agent. In under 15 minutes, the software found the root cause of an issue that would have taken a veteran engineer hours, AWS said in a statement.
The tool relies on Amazon’s in-house AI models and those from other providers, a spokesperson said.
AWS has been selling software in addition to raw infrastructure for many years. Amazon was early to start renting out server space and storage to developers since the mid-2000s, and technology companies such as Google, Microsoft and Oracle have followed.
Since the launch of ChatGPT in 2022, these cloud infrastructure providers have been trying to demonstrate how generative AI models, which are often training in large cloud computing data centers, can speed up work for software developers.
Over the summer, Amazon announced Kiro, a so-called vibe coding tool that produces and modifies source code based on user text prompts. In November, Google debuted similar software for individual software developers called Antigravity, and Microsoft sells subscriptions to GitHub Copilot.
Attendees pass an Amazon Web Services logo during AWS re:Invent 2024, a conference hosted by Amazon Web Services, at The Venetian hotel in Las Vegas on Dec. 3, 2024.
Noah Berger | Getty Images
Amazon has found a way to let cloud clients extensively customize generative AI models. The catch is that the system costs $100,000 per year.
The Nova Forge offering from Amazon Web Services gives organizations access to Amazon’s AI models in various stages of training so they can incorporate their own data earlier in the process.
Already, companies can fine-tune large language models after they’ve been trained. The results with Nova Forge will lean more heavily on the data that customers supply. Nova Forge customers will also have the option to refine open-weight models, but training data and computing infrastructure are not included.
Organizations that assemble their own models might end up spending hundreds of millions or billions of dollars, which means using Nova Forge is more affordable, Amazon said.
AWS released its own models under the Nova brand in 2024, but they aren’t the first choice for most software developers. A July survey from Menlo Ventures said that by the middle of this year, Amazon-backed Anthropic controlled 32% of the market for enterprise LLMs, followed by OpenAI with 25%, Google with 20% and Meta with 9% — Amazon Nova had a less than 5% share, a Menlo spokesperson said.
The Nova models are available through AWS’ Bedrock service for running models on Amazon cloud infrastructure, as are Anthropic’s Claude 4.5 models.
“We are a frontier lab that has focused on customers,” Rohit Prasad, Amazon head scientist for artificial general intelligence, told CNBC in an interview. “Our customers wanted it. We have invented on their behalf to make this happen.”
Nova Forge is also in use by internal Amazon customers, including teams that work on the company’s stores and the Alexa AI assistant, Prasad said.
Reddit needed an AI model for moderating content that would be sophisticated about the many subjects people discuss on the social network. Engineers found that a Nova model enhanced with Reddit data through Forge performed better than commercially available large-scale models, Prasad said. Booking.com, Nimbus Therapeutics, the Nomura Research Institute and Sony are also building models with Forge, Amazon said.
Organizations can request that Amazon engineers help them build their Forge models, but that assistance is not included in the new service’s $100,000 annual fee.
AWS is also introducing new models for developers at its Reinvent conference in Las Vegas this week.
Nova 2 Pro is a reasoning model whose tests show it performs at least as well as Anthropic’s Claude Sonnet 4.5, OpenAI’s GPT-5 and GPT-5.1, and Google’s Gemini 3.0 Pro Preview, Amazon said. Reasoning involves running a series of computations that might take extra time in response to requests to produce better answers. Nova 2 Pro will be available in early access to AWS customers with Forge subscriptions, Prasad said. That means Forge customers and Amazon engineers will be able to try Nova 2 Pro at the same time.
Nova 2 Omni is another reasoning model that can process incoming images, speech, text and videos, and it generates images and text. It’s the first reasoning model with that range of capability, Amazon said. Amazon hopes that, by delivering a multifaceted model, it can lower the cost and complexity of incorporating AI models into applications.
Tens of thousands of organizations are using Nova models each week, Prasad said. AWS has said it has millions of customers. Nova is the second-most popular family of models in Bedrock, Prasad said. The top group of models are from Anthropic.
Elon Musk, chief executive officer of Tesla Inc., during the US-Saudi Investment Forum at the Kennedy Center in Washington, DC, US, on Wednesday, Nov. 19, 2025.
Bloomberg | Bloomberg | Getty Images
Elon Musk has again sounded the alarm on the dangers of AI and listed what he considers as the three most important ingredients to ensure a positive future with the technology.
The billionaire CEO of Tesla, SpaceX, xAI, X and The Boring Company, appeared on a podcast with Indian billionaire Nikhil Kamath on Sunday.
“It’s not that we’re guaranteed to have a positive future with AI,” Musk said on the podcast. “There’s some danger when you create a powerful technology, that a powerful technology can be potentially destructive.”
Musk was a co-founder of OpenAI alongside Sam Altman, but left its board in 2018 and publicly criticized the company for ditching its founding mission as a non-profit to develop AI safely after it launched ChatGPT in 2022. Musk’s xAI developed its own chatbot, Grok, in 2023.
Musk has previously warned that “one of the biggest risks to the future of civilization is AI,” and stressed that rapid advancements are leading AI to become a bigger risk to society than cars or planes or medicines.
On the podcast, the tech billionaire emphasized the importance of ensuring AI technologies pursue truth instead of repeating inaccuracies. “That can be very dangerous,” Musk told Kamath, who is also the co-founder of retail stockbroker Zerodha.
“Truth and beauty and curiosity. I think those are the three most important things for AI,” he said.
He said that, without strictly adhering to truths, AI will learn information from online sources where it “will absorb a lot of lies and then have trouble reasoning because these lies are incompatible with reality.”
He added: “You can make an AI go insane if you force it to believe things that aren’t true because it will lead to conclusions that are also bad.”
“Hallucination” — responses that are incorrect or misleading — is a major challenge facing AI. Earlier this year, an AI feature launched by Apple on its iPhones generated fake news alerts.
These included a false summary from the BBC News app notifications on a story about the PDC World Darts Championship semi-final, where it wrongly claimed that the British darts player Luke Littler had won the championship. Littler did not win the tournament’s final until the next day.
Apple told the BBC at the time that it was working on an update to resolve the problem which clarifies when Apple Intelligence is responsible for the text shown in the notifications.
Musk added that “some appreciation of beauty is important” and that “you know it when you see it.”
Musk said AI should want to know more about the nature of reality because humanity is more interesting than machines.
“It’s more interesting to see the continuation if not the prosperity of humanity than to exterminate humanity,” he said.
Geoffrey Hinton, a computer scientist and ex-vice president at Google known as a “Godfather of AI,” said earlier this year that there’s a “10% to 20% chance” that AI will “wipe us out,” in an episode of the Diary of a CEO podcast. Some of the shorter term risks he cited included hallucinations and the automation of entry level jobs.
“The hope is that if enough smart people do enough research with enough resources, we’ll figure out a way to build them so they’ll never want to harm us,” Hinton added.