But these image-generating programs — which look like toys today — could be the start of a big wave in technology. Technologists call them generative models, or generative AI.
“In the last three months, the words ‘generative AI’ went from, ‘no one even discussed this’ to the buzzword du jour,” said David Beisel, a venture capitalist at NextView Ventures.
In the past year, generative AI has gotten so much better that it’s inspired people to leave their jobs, start new companies and dream about a future where artificial intelligence could power a new generation of tech giants.
The field of artificial intelligence has been having a boom phase for the past half-decade or so, but most of those advancements have been related to making sense of existing data. AI models have quickly grown efficient enough to recognize whether there’s a cat in a photo you just took on your phone and reliable enough to power results from a Google search engine billions of times per day.
But generative AI models can produce something entirely new that wasn’t there before — in other words, they’re creating, not just analyzing.
“The impressive part, even for me, is that it’s able to compose new stuff,” said Boris Dayma, creator of the Craiyon generative AI. “It’s not just creating old images, it’s new things that can be completely different to what it’s seen before.”
Sequoia Capital — historically the most successful venture capital firm in the history of the industry, with early bets on companies like Apple and Google — says in a blog post on its website that “Generative AI has the potential to generate trillions of dollars of economic value.” The VC firm predicts that generative AI could change every industry that requires humans to create original work, from gaming to advertising to law.
In a twist, Sequoia also notes in the post that the message was partially written by GPT-3, a generative AI that produces text.
How generative AI works
Kif Leswing/Craiyon
Image generation uses techniques from a subset of machine learning called deep learning, which has driven most of the advancements in the field of artificial intelligence since a landmark 2012 paper about image classification ignited renewed interest in the technology.
Deep learning uses models trained on large sets of data until the program understands relationships in that data. Then the model can be used for applications, like identifying if a picture has a dog in it, or translating text.
Image generators work by turning this process on its head. Instead of translating from English to French, for example, they translate an English phrase into an image. They usually have two main parts, one that processes the initial phrase, and the second that turns that data into an image.
The first wave of generative AIs was based on an approach called GAN, which stands for generative adversarial networks. GANs were famously used in a tool that generates photos of people who don’t exist. Essentially, they work by having two AI models compete against each other to better create an image that fits with a goal.
Newer approaches generally use transformers, which were first described in a 2017 Google paper. It’s an emerging technique that can take advantage of bigger datasets that can cost millions of dollars to train.
The first image generator to gain a lot of attention was DALL-E, a program announced in 2021 by OpenAI, a well-funded startup in Silicon Valley. OpenAI released a more powerful version this year.
“With DALL-E 2, that’s really the moment when when sort of we crossed the uncanny valley,” said Christian Cantrell, a developer focusing on generative AI.
Another commonly used AI-based image generator is Craiyon, formerly known as Dall-E Mini, which is available on the web. Users can type in a phrase and see it illustrated in minutes in their browser.
Since launching in July 2021, it’s now generating about 10 million images a day, adding up to 1 billion images that have never existed before, according to Dayma. He’s made Craiyon his full-time job after usage skyrocketed earlier this year. He says he’s focused on using advertising to keep the website free to users because the site’s server costs are high.
For example, Stable Diffusion was integrated into Adobe Photoshop through a plug-in, allowing users to generate backgrounds and other parts of images that they can then directly manipulate inside the application using layers and other Photoshop tools, turning generative AI from something that produces finished images into a tool that can be used by professionals.
“I wanted to meet creative professionals where they were and I wanted to empower them to bring AI into their workflows, not blow up their workflows,” said Cantrell, developer of the plug-in.
Cantrell, who was a 20-year Adobe veteran before leaving his job this year to focus on generative AI, says the plug-in has been downloaded tens of thousands of times. Artists tell him they use it in myriad ways that he couldn’t have anticipated, such as animating Godzilla or creating pictures of Spider-Man in any pose the artist could imagine.
“Usually, you start from inspiration, right? You’re looking at mood boards, those kinds of things,” Cantrell said. “So my initial plan with the first version, let’s get past the blank canvas problem, you type in what you’re thinking, just describe what you’re thinking and then I’ll show you some stuff, right?”
An emerging art to working with generative AIs is how to frame the “prompt,” or string of words that lead to the image. A search engine called Lexica catalogs Stable Diffusion images and the exact string of words that can be used to generate them.
Startups, cloud providers, and chip makers could thrive
Image generated by DALL-E with prompt: A cat on sitting on the moon, in the style of Pablo Picasso, detailed, stars
Screenshot/OpenAI
Some investors are looking at generative AI as a potentially transformative platform shift, like the smartphone or the early days of the web. These kinds of shifts greatly expand the total addressable market of people who might be able to use the technology, moving from a few dedicated nerds to business professionals — and eventually everyone else.
“It’s not as though AI hadn’t been around before this — and it wasn’t like we hadn’t had mobile before 2007,” said Beisel, the seed investor. “But it’s like this moment where it just kind of all comes together. That real people, like end-user consumers, can experiment and see something that’s different than it was before.”
Cantrell sees generative machine learning as akin to an even more foundational technology: the database. Originally pioneered by companies like Oracle in the 1970s as a way to store and organize discrete bits of information in clearly delineated rows and columns — think of an enormous Excel spreadsheet, databases have been re-envisioned to store every type of data for every conceivable type of computing application from the web to mobile.
“Machine learning is kind of like databases, where databases were a huge unlock for web apps. Almost every app you or I have ever used in our lives is on top of a database,” Cantrell said. “Nobody cares how the database works, they just know how to use it.”
Michael Dempsey, managing partner at Compound VC, says moments where technologies previously limited to labs break into the mainstream are “very rare” and attract a lot of attention from venture investors, who like to make bets on fields that could be huge. Still, he warns that this moment in generative AI might end up being a “curiosity phase” closer to the peak of a hype cycle. And companies founded during this era could fail because they don’t focus on specific uses that businesses or consumers would pay for.
Others in the field believe that startups pioneering these technologies today could eventually challenge the software giants that currently dominate the artificial intelligence space, including Google, Facebook parent Meta and Microsoft, paving the way for the next generation of tech giants.
“There’s going to be a bunch of trillion-dollar companies — a whole generation of startups who are going to build on this new way of doing technologies,” said Clement Delangue, the CEO of Hugging Face, a developer platform like GitHub that hosts pre-trained models, including those for Craiyon and Stable Diffusion. Its goal is to make AI technology easier for programmers to build on.
Some of these firms are already sporting significant investment.
Hugging Face was valued at $2 billion after raising money earlier this year from investors including Lux Capital and Sequoia; and OpenAI, the most prominent startup in the field, has received over $1 billion in funding from Microsoft and Khosla Ventures.
Meanwhile, Stability AI, the maker of Stable Diffusion, is in talks to raise venture funding at a valuation of as much as $1 billion, according to Forbes. A representative for Stability AI declined to comment.
Cloud providers like Amazon, Microsoft and Google could also benefit because generative AI can be very computationally intensive.
Meta and Google have hired some of the most prominent talent in the field in hopes that advances might be able to be integrated into company products. In September, Meta announced an AI program called “Make-A-Video” that takes the technology one step farther by generating videos, not just images.
“This is pretty amazing progress,” Meta CEO Mark Zuckerberg said in a post on his Facebook page. “It’s much harder to generate video than photos because beyond correctly generating each pixel, the system also has to predict how they’ll change over time.”
On Wednesday, Google matched Meta and announced and released code for a program called Phenaki that also does text to video, and can generate minutes of footage.
The boom could also bolster chipmakers like Nvidia, AMD and Intel, which make the kind of advanced graphics processors that are ideal for training and deploying AI models.
At a conference last week, Nvidia CEO Jensen Huang highlighted generative AI as a key use for the company’s newest chips, saying these kind of programs could soon “revolutionize communications.”
Profitable end uses for Generative AI are currently rare. A lot of today’s excitement revolves around free or low-cost experimentation. For example, some writers have been experimented with using image generators to make images for articles.
Prompt: “A cat sitting on the moon, in the style of picasso, detailed”
Screenshot/Craiyon
Ultimately, everyone developing generative AI will have to grapple with some of the ethical issues that come up from image generators.
First, there’s the jobs question. Even though many programs require a powerful graphics processor, computer-generated content is still going to be far less expensive than the work of a professional illustrator, which can cost hundreds of dollars per hour.
That could spell trouble for artists, video producers and other people whose job it is to generate creative work. For example, a person whose job is choosing images for a pitch deck or creating marketing materials could be replaced by a computer program very shortly.
“It turns out, machine-learning models are probably going to start being orders of magnitude better and faster and cheaper than that person,” said Compound VC’s Dempsey.
There are also complicated questions around originality and ownership.
Generative AIs are trained on huge amounts of images, and it’s still being debated in the field and in courts whether the creators of the original images have any copyright claims on images generated to be in the original creator’s style.
One artist won an art competition in Colorado using an image largely created by a generative AI called MidJourney, although he said in interviews after he won that he processed the image after choosing it from one of hundreds he generated and then tweaking it in Photoshop.
Some images generated by Stable Diffusion seem to have watermarks, suggesting that a part of the original datasets were copyrighted. Some prompt guides recommend using specific living artists’ names in prompts in order to get better results that mimic the style of that artist.
Image generators can also be used to create new images of trademarked characters or objects, such as the Minions, Marvel characters or the throne from Game of Thrones.
As image-generating software gets better, it also has the potential to be able to fool users into believing false information or to display images or videos of events that never happened.
Developers also have to grapple with the possibility that models trained on large amounts of data may have biases related to gender, race or culture included in the data, which can lead to the model displaying that bias in its output. For its part, Hugging Face, the model-sharing website, publishes materials such as an ethics newsletter and holds talks about responsible development in the AI field.
“What we’re seeing with these models is one of the short-term and existing challenges is that because they’re probabilistic models, trained on large datasets, they tend to encode a lot of biases,” Delangue said, offering an example of a generative AI drawing a picture of a “software engineer” as a white man.
Participants at the presentation of new iPhone models from Apple try out the new thinner iPhone Air.
Andrej Sokolow | Picture Alliance | Getty Images
Apple has postponed the launch of its new iPhone Air model in China due to regulatory issues surrounding its eSIM design, the company said.
Wireless carriers in China need a special license from the government before they can sell a new device with an eSIM, and the carriers haven’t secured that approval yet, Apple said. The company added that it’s working to make the device available in China as soon as possible.
Apple announced the iPhone Air at its annual event on Tuesday. The device, which is 5.6 millimeters thick, marks the first major new iPhone design since the iPhone X was introduced in 2017. The iPhone Air doesn’t support a physical SIM card, and instead features an eSIM built into the device.
CEO Tim Cook told CNBC’s Jim Cramer on Friday that the eSIM is what allows the device to still have “great” battery life.
“It’s eSIM only, and so we were able to take the battery and extend the battery to areas that previously had the physical cell,” Cook said.
Read more CNBC tech news
The company previously said the iPhone Air would become available for pre-order in the region on Friday at 2 a.m. EST before it goes on sale September 19.
As of Friday morning, the iPhone Air product page on Apple’s China website stated, “Release information will be updated later.”
The website notes that China Mobile, China Telecom and China Unicom will offer eSIM support for the iPhone Air, “with specific timing subject to regulatory approval.”
Jensen Huang, co-founder and chief executive officer of Nvidia, at the London Tech Week exposition in London, UK, on Monday, June 9, 2025.
Bloomberg | Bloomberg | Getty Images
Nvidia and OpenAI are in discussions about backing a major investment in Britain focused on boosting artificial intelligence infrastructure in the country.
The two tech firms are discussing a sizable deal to support data center development in the country which could ultimately be worth billions of dollars, a person familiar with the matter told CNBC, confirming earlier reporting from the Financial Times.
The companies are still working through various processes at the moment with Nvidia and cloud computing firm Nscale, said the person, who did not want to be named due to the sensitivity of the issue.
They added that an investment agreement has not yet been finalized. It is expected to be unveiled next week during U.S. President Donald Trump’s state visit to the U.K.
Nvidia and Nscale did did not immediately respond to CNBC’s request for comment. OpenAI declined to comment on the discussions.
Countries around the world have been courting major U.S. AI players in a bid to boost their own national infrastructure and technological ambitions.
The topic of so-called “sovereign” AI — the idea of onshoring the data processing infrastructure behind advanced artificial intelligence systems — has been top of mind for officials as governments look to reduce their dependency on foreign countries for critical technologies.
The U.K. government declined to comment when asked by CNBC about the investment discussions with OpenAI, Nvidia and Nscale. Nvidia CEO Jensen Huang is set to join Trump on his state visit to Britain next week.
Earlier this year, the Nvidia boss called the U.K. an “incredible place to invest” and said his multitrillion-dollar chipmaker would boost investment in the country. “The U.K. is in a Goldilocks circumstance,” Huang said at the time in a panel discussion with British Prime Minister Keir Starmer.
Apple AirPods Pro 3 models are displayed during Apple’s “Awe-Dropping” event at the Steve Jobs Theater on the Apple Park campus in Cupertino, California, on Sept. 9, 2025.
Nic Coury | AFP | Getty Images
For decades, shows like “Star Trek” and novels like “The Hitchhikers Guide to the Galaxy” have showcased fictional universal translators, capable of seamlessly converting any language into English and vice versa.
Now, those gadgets once limited to works of science fiction are inching close to reality.
During its iPhone unveiling event on Tuesday, Apple included a video of many travelers’ dream scenario. It showed an English-speaking tourist buying flowers in an unnamed Spanish-speaking country. The florist addressed the tourist in Spanish, but what the tourist heard was in clear, coherent English.
“Today all the red carnations are 50% off,” the tourist heard in English in her headphones, at essentially the same time that the clerk was speaking.
The video was marketing material for Apple’s latest AirPods Pro 3, but the feature is one of many of its kind coming from tech companies that also include Google parent Alphabet and Meta, which makes Facebook and Instagram.
Apple introduces live translation to airpods.
Courtesy: Apple
Technological advancements spurred by the arrival of OpenAI’s ChatGPT in late 2022 have ushered in an era of generative artificial intelligence. Almost three years later, those advancements are resulting in real-time language translators.
For Apple, Live Translation is a key selling point for the AirPods Pro 3, which the company unveiled on Tuesday. The new $250 earbuds go on sale next week, and with Live Translation, users will be able to immediately hear French, German, Portuguese and Spanish translated to English. Live Translation will also arrive as an update to AirPods 4 and AirPods Pro 2 on Monday.
And when two people are speaking to each other wearing AirPods, the conversation can be translated both ways simultaneously inside each user’s headphones. In Apple’s video demo, it looked like two people talking to each other in different languages.
Analysts are excited that the feature could mark a step forward for Apple’s AI strategy. The translation feature needs to be paired with a new-enough iPhone to run Apple Intelligence, Apple’s AI software suite.
“If we can actually use the AirPods for live translations, that’s a feature that would actually get people to upgrade,” DA Davidson analyst Gil Luria told CNBC on Wednesday.
Translation is emerging as a key battleground in the technology industry as AI gets good enough to translate languages as quickly as people speak.
But Apple is not alone.
Host Jimmy Fallon holds Pixel 10 Pro Fold mobile phone during the ‘Made by Google’ event, organised to introduce the latest additions to Google’s Pixel portfolio of devices, in Brooklyn, New York, U.S., August 20, 2025.
Brendan McDermid | Reuters
A crowded market
In the past year, Google and Meta have also released hardware products featuring real-time translation capabilities.
Google’s Pixel 10 phone has a capability that can translate what a speaker is saying to the listener’s language during phone calls. That feature, called Voice Translate is designed to also preserve the speaker’s voice inflections. Voice Translate will start showing up on people’s phones through a software update on Monday.
In Google’s live demo in August, Voice Translate was able to translate a sentence from entertainer Jimmy Fallon into Spanish, and it actually sounded like the comedian. Apple’s feature does not try to imitate the user’s voice.
Meanwhile, Meta in May announced that its Ray-Ban Meta glasses would be able to translate what a person is saying in another language using the device’s speakers, and the other party in the conversations would be able to see translated responses transcribed on the user’s phone.
Meta will hold its own product keynote on Wednesday, where the company is expected to announce the next generation of its smart glasses, which will feature a small display in one of the lenses, CNBC reported in August. It’s unclear if Meta will announce more translation features.
Meta employee Sara Nicholson poses with the Ray-Ban sunglasses at the Meta Connect annual event at the company’s headquarters in Menlo Park, California, U.S., September 24, 2024.
Manuel Orbegozo | Reuters
And OpenAI in June showcased an intelligent voice assistant mode for ChatGPT that has fluid translation built-in as one of many features. ChatGPT is integrated with Apple’s Siri, but not in voice mode. OpenAI is planning to release new hardware products with Apple’s former design guru Jony Ive in the coming years.
The rise of live translation could also reshape entire industries. Translators and interpreters are the number one type of job threatened by AI, and 98% of translators’ work activities overlap with what AI can do, a Microsoft Research study published in August found.
Purpose built translators
In the past several years, a number of purpose-built translation gadgets have entered the market, taking advantage of global high-speed cellular service and improving online translation services to produce puck-like devices or headphones with translation built-in for a couple hundred dollars.
“What I love about what Apple is doing is it really just illuminates the fact that how pressing of an issue this is,” said Joe Miller, U.S. general manager of Japan-based Pocketalk, which makes a $299 translation device that goes between two people conversing in different languages and translates their conversation in audio and text.
Given Apple’s massive scale and the fact that the Apple shipped about 18 million sets of wireless headphones in the first quarter alone, according to Canalys, the company’s entry into the market will expose a wider subset of customers to improvements translation tech has made in recent years.
Despite Apple’s entry into the market, makers of purpose-built devices say their focus on accuracy and knowledge of linguistics will provide better translations than what’s available for free with a new phone.
“We actually hired linguists,” said Aleksander Alski, head of U.S. and Canada for Poland-based Vasco Electronics, which is releasing translation headphones that can imitate the user’s voice, like Google’s feature. “We combined the AI with with human input, and thanks to that, we were able to secure much higher accuracy throughout all the languages we offer.”
There’s also home-field advantage. Vasco Electronics’ largest market is Europe, and Apple’s Live Translation isn’t available for EU users, Apple said on its website.
Some of the products being introduced by tech companies are less than universal, and are limited to a small number of languages for now. Apple’s feature is only available in 5 languages, versus Pocketalk’s 95.
Pocketalk’s Miller believes that the potential of the technology goes far beyond a tourist ordering a glass of wine in France. He says that it’s most powerful when its used in workplaces like schools and hospitals, which require privacy and security features that go beyond what Apple and Google provide.
“This isn’t about luxury tourism and travel,” Miller said. “This is about the intersection of language and friction, when a discussion needs to be had.”