Big players, including Microsoft, with Copilot, Google, with Gemini, and OpenAI, with GPT-4o, are making AI chatbot technology previously restricted to test labs more accessible to the general public.
How do these large language model (LLM) programs work? OpenAI’s GPT-3 told us that AI uses “a series of autocomplete-like programs to learn language” and that these programs analyze “the statistical properties of the language” to “make educated guesses based on the words you’ve typed previously.”
Or, in the words of James Vincent, a human person: “These AI tools are vast autocomplete systems, trained to predict which word follows the next in any given sentence. As such, they have no hard-coded database of ‘facts’ to draw on — just the ability to write plausible-sounding statements. This means they have a tendency to present false information as truth since whether a given sentence sounds plausible does not guarantee its factuality.”
But there are so many more pieces to the AI landscape that are coming into play (and so many name changes — remember when we were talking about Bing and Bard before those tools were rebranded?), but you can be sure to see it all unfold here on The Verge.
Highlights
- AI is confusing — here’s your cheat sheet
- Figma explains how its AI tool ripped off Apple’s design
- Early Apple tech bloggers are shocked to find their name and work have been AI-zombified
- Google and OpenAI are racing to rewire the internet
- We have to stop ignoring AI’s hallucination problem
- OpenAI releases GPT-4o, a faster model that’s free for all ChatGPT users
- The true promise of AI: Siri that doesn’t suck.
Dec 8
X’s new ‘Aurora’ image generator is gone.The “Grok 2 + Aurora” option has vanished from Grok’s model selector menu only a day after it appeared, Engadget reports, replaced by “Grok 2 + FLUX beta” instead. The model still makes photorealistic images, but it was less willing to reproduce celebrities when I asked.
X owner Elon Musk wrote yesterday that the photorealistic and largely unrestricted model is a beta “internal image generator.”
Update: Added testing detail.
Dec 7
X gives Grok a new photorealistic AI image generator
Image: The VergeX has given Grok a new AI image generator model called “Aurora” that seems to create far more photorealistic imagery than Grok’s other image generator, with similarly few apparent restrictions on what it will produce, TechCrunch reports. Like Grok, anyone can use Aurora. It lives in a new “Grok 2 + Aurora beta” option in the Grok model selector, though you’ll only get a few queries before you hit the X Premium subscription paywall and have to wait.
Read Article >TechCrunch found that the model, which X employee Chris Park posted is available this morning, was willing to create copyrighted characters and public figures, including Mickey Mouse and “a bloodied Donald Trump,” but that it “stopped short of nudes.” Its lack of restrictions isn’t surprising, given our experience with Grok’s other model.
Dec 5
Sundar Pichai says Google Search will ‘change profoundly’ in 2025
Image: Laura Normand / The VergeGoogle CEO Sundar Pichai says the company’s search engine will “change profoundly” in 2025. “I think we are going to be able to tackle more complex questions than ever before,” Pichai said during the NYT’s DealBook Summit on Wednesday.
Read Article >“I think you’ll be surprised, even early in ‘25, the kind of newer things Search can do compared to where it is today.”
Dec 4
ChatGPT now has over 300 million weekly users
Image: The VergeChatGPT now has over 300 million people using the AI chatbot each week. OpenAI CEO Sam Altman revealed the milestone during The New York Times’ DealBook Summit on Wednesday, which comes just months after ChatGPT hit 200 million weekly users in August.
Read Article >“Our product has scaled ... now we have more than 300 million weekly active users,” Altman said. “We have users sending more than 1 billion messages per day to ChatGPT.”
Dec 4
OpenAI’s 12 days of ‘shipmas’ include Sora and new reasoning model
Image: Alex Parkin / The VergeHappy holidays from OpenAI. The AI startup plans to kick off a “shipmas” period of new features, products, and demos for 12 days, starting on December 5th. The announcements will include OpenAI’s long-awaited text-to-video AI tool Sora and a new reasoning model, sources familiar with OpenAI’s plans tell The Verge.
Read Article >OpenAI CEO Sam Altman confirmed the 12 days of announcements onstage at The New York Times’ DealBook conference on Wednesday morning, though he didn’t say exactly what was coming. OpenAI plans to launch or demo something every day for 12 days straight.
Dec 3
ChatGPT’s search results for news are ‘unpredictable’ and frequently inaccurate
Illustration: The VergeBased on testing done by Columbia’s Tow Center for Digital Journalism researchers, OpenAI’s ChatGPT search tool has some issues when it comes to responding with the truth.
Read Article >OpenAI launched the tool for subscribers in October, saying it could give “fast, timely answers with links to relevant web sources.” Instead, Futurism points out that the researchers said ChatGPT search struggled to correctly identify quotes from articles, even when they came from publishers with arrangements to share data with OpenAI.
Nov 27
Former Android leaders are building an ‘operating system for AI agents’
Image: The VergeA new startup created by former Android leaders aims to build an operating system for AI agents. Among them is Hugo Barra, Google’s former VP of Android product management, who says the new company — named “/dev/agents” — will revisit the leaders’ “Android roots.”
Read Article >AI companies are pushing AI agents as the next big leap in AI tools, promising digital assistants that can carry out tasks and make decisions autonomously and with little human input. Companies like Microsoft, Google, and OpenAI each have plans to launch some version of this concept as a product in the next few months.
Nov 27
Bluesky won’t use your posts for AI training, but can it stop anyone else?Today, a Hugging Face employee published data from 1 million Bluesky posts scraped from its API to the AI repository. He’s removed it and apologized, but 404 Media notes the set was “trending” all day.
Bluesky says it’s looking into ways to “specify consent (or not) for AI training.” but acknowledges that “It will be up to outside developers to respect these settings.”
Nov 26
Artists say they leaked OpenAI’s Sora video model in protest
A still from a video generated by OpenAI’s Sora. OpenAIOpenAI first teased its text-to-video AI model, Sora, back in February and hasn’t provided any meaningful updates on when it will be released since then. Now, it looks like some artists leaked access to the model in protest of being used by the company for what they claim is “unpaid R&D and PR.”
Read Article >On Tuesday, a group of Sora beta testers claimed to have leaked early access to Sora with a working interface for generating videos. In a post on Hugging Face, a public repository of AI models, they say that people were able to create lots of AI videos — all of which resemble OpenAI’s own Sora demos — before the company intervened to shut down access. (TechCrunch first reported on the alleged leak.)
Nov 26
Anthropic says Claude AI can match your unique writing style
Anthropic is adding custom styles to Claude that quickly adjust the tone and length of responses. Image: AnthropicAnthropic is adding a new feature to its Claude AI assistant that will give users more control over how the chatbot responds to different writing tasks. The new custom styles are available to all Claude AI users, enabling anyone to train it to match their own communication style or select from preset options to quickly adjust the tone and level of detail it provides.
Read Article >This update aims to personalize the chatbot’s replies and make them feel more natural or appropriate for specific applications, such as writing detailed technical documents or professional emails.
Nov 22
Is AI hitting a wall?
Image: Cath Virginia / The Verge, Adobe StockAI’s rate of progress is hitting a wall. It may not matter.
Read Article >That was the prevailing theme from this week’s Cerebral Valley AI Summit in San Francisco — a gathering of about 350 CEOs, engineers, and investors in the AI industry that I attended on Wednesday.
Nov 20
Messenger video calls will look and sound better — and add AI backgrounds
Image: MetaMeta has announced that it’s adding HD video calling and noise suppression to its Messenger app, along with several other features like AI backgrounds that help bring it up to par with other popular smartphone video calling apps.
Read Article >HD video calling is enabled by default on Wi-Fi but can be turned on for calls using your cell data plan in the “Call settings” menu. You’ll also find options to adjust noise cancellation or switch your call’s audio output. Meta says it’s giving users the ability to leave voice or video messages as well.
Nov 20
Niantic is building a ‘geospatial’ AI model based on Pokémon Go player data
Photo by Sam Byford / The VergeNiantic has announced that it’s building a new “Large Geospatial Model” (LGM) that combines millions of scans taken from the smartphones of players of Pokémon Go and other Niantic products. This AI model could allow computers and robots to understand and interact with the world in new ways, the company said in a blog post spotted by Garbage Day.
Read Article >The LGM’s “spatial intelligence” is built on the neural networks developed as part of Niantic’s Visual Positioning System. The blog post explains that “Over the past five years, Niantic has focused on building our Visual Positioning System (VPS), which uses a single image from a phone to determine its position and orientation using a 3D map built from people scanning interesting locations in our games and Scaniverse,” and “This data is unique because it is taken from a pedestrian perspective and includes places inaccessible to cars.”
Nov 19
Google Gemini can remember things now
Illustration: The VergeGoogle Gemini can now “remember” certain things about you, such as your interests and personal preferences. The change is rolling out to subscribers with access to Gemini Advanced, and it will allow the AI chatbot to tailor its responses based on your needs, according to Gemini’s most recent release notes.
Read Article >As noted by Google, you can share your interests, as well as details about your work, hobbies, or aspirations, while chatting with Gemini. You can also enter relevant details about yourself within the bot’s “Saved Info” page, such as telling it to only suggest vegetarian recipes, or that you work as an accountant.
Nov 16
“Oregon -14 Wisconsin Texas.”I’m glad Apple Intelligence summaries exist. Because this is a beautiful mess.
And life-changing, just like Tim Cook said.
Nov 15
ESPN is testing a generative AI avatar called ‘FACTS’
Image: ESPNESPN is testing an AI-generated avatar with the Saturday college football show SEC Nation. Dubbed FACTS, it’s going to be “...promoting education and fun around sports analytics” with information drawn from ESPN Analytics, which includes data like the Football Power Index (FPI), player and team statistics, and game schedules. We haven’t seen the avatar in action, but it sounds like a bot-ified version of stats encyclopedia Howie Schwab, who was ESPN’s first statistician and eventually the star of a mid-2000s game show, Stump the Schwab.
Read Article >ESPN has already brought generative AI to its website with AI-written game recaps. FACTS is still in development, and there’s no word on when it could make its first appearance on the network.
Nov 14
Google’s Gemini AI now has its own iPhone app
In the AI chatbot world, ubiquity is everything. Companies have raced to build desktop and mobile apps for their bots, in order to both give them new capabilities but also to make sure they’re right in front of your face as often as possible.
Read Article >There’s no better example of that than Google’s new Gemini app for iPhone, which quietly hit the App Store around the world this week. The free app is simple and straightforward: it’s just a chat window and a list of your previous chats. You can query the bot with text, voice, or your camera, and it’ll give you answers. It’s effectively identical to the Gemini section of the Google app, or what you’d get by opening a browser and going to the Gemini website.
Nov 13
Perplexity is starting its ads experiment this week.The AI-powered search engine will begin with showing labeled ads formatted as “sponsored follow-up questions” that will appear next to answers in the US.
TechCrunch reports:
“Perplexity’s embrace of ads stands in contrast to OpenAI’s decision not to launch its AI-powered search tool, ChatGPT Search, with ads. But rival Google has similarly piloted ads in its AI search experience, AI Overviews — recently bringing ads to mobile in the U.S. for certain queries”
Nov 12
Line go up.Anthropic co-founder Darius Amodei said on Lex Fridman’s podcast yesterday that, “if you believe the straight-line extrapolation,” we’ll have artificial general intelligence “in 2026 or 2027.” He also indicates that’s a big if, listing reasons why (which don’t include that LLMs might not be the way to AGI).
Makes sense — after all, according to growth trends, I should have been the size of a dinosaur by age 10.
Nov 11
Google’s AI ‘learning companion’ takes chatbot answers a step further
Google’s Learn About website. Image: GoogleGoogle has launched an experimental new AI tool called Learn About, which is different from the chatbots we’re used to, like Gemini and ChatGPT. It’s built on the LearnLM AI model that Google introduced this spring, saying it’s “grounded in educational research and tailored to how people learn.” The answers it provides have more visual and interactive elements with educational formatting.
Read Article >We tested Learn About and Google Gemini with a simple prompt: “How big is the universe?” Both answered that “the observable universe” is “about 93 billion light-years in diameter.”
Nov 9
Sold.A piece of artwork called “A.I. God. Portrait of Alan Turing,” created in part by AI-equipped robot Ai-Da, went for almost $1.1 million at auction, writes The New York Times.
There was a lot of human involvement, starting with combining the bot’s paintings of parts of Turing’s face:
The works were then photographed and uploaded to a computer that used Ai-Da’s language model to decide on the assembly of a single painting, which was then completed using a 3-D textured printer; studio assistants helped to create a more realistic finished product on the canvas. Ai-Da then added marks and textures to the portrait to complete it.
A.I.-Powered Painting of Alan Turing Sells for $1.1 Million[The New York Times]
Oct 30
Google says Sundar didn’t actually entertain the idea of splitting off AI from search.On Google’s Q3 earnings call, Wells Fargo analyst Ken Gawrelski asked:
Why doesn’t it make sense to have two completely different search experiences? One, an agent-like answers engine and then, two, a links-based more traditional search engine? You could innovate on both and let the consumer decide.
Sundar replied: “I do think having two surfaces for us allows us to experiment more.”
But Google rep Chris Pappas tells us he’s referring to two different AI surfaces — AI Overviews and the Gemini App — not good ol’ link-based search.
Oct 28
Universal Music partners with AI company building an ‘ethical’ music generator
UMG and AI company Klay are forming their own foundational Music-generating AI model. Cath Virginia / The Verge | Photo from Getty ImageUniversal Music Group (UMG) announced a new deal centered on creating an “ethical” foundational model for AI music generation. It’s partnered with a company called Klay Vision that’s creating a “Large Music Model” named KLayMM and plans to launch out of stealth mode with a product within months. Ary Attie, its founder and CEO, said the company believes “the next Beatles will play with KLAY.”
Read Article >The two say the model will work “in collaboration with the music industry and its creators,” without many details about how, while Klay plans to make music AI “more than a short-lived gimmick.”
Oct 27
Hospitals use a transcription tool powered by an error-prone OpenAI model
Image: The VergeA few months ago, my doctor showed off an AI transcription tool he used to record and summarize patient meetings. In my case, the summary was fine, but researchers cited in this report by The Associated Press have found that’s not always the case for transcriptions created by OpenAI’s Whisper, which powers a tool many hospitals use — sometimes it just makes things up entirely.
Read Article >Whisper is used by a company called Nabla for a tool that it estimates has transcribed 7 million medical conversations, according to AP. More than 30,000 clinicians and 40 health systems use it, the outlet writes. The report says that Nabla officials “are aware that Whisper can hallucinate and are addressing the problem.” In a blog post published Monday, execs wrote that their model includes improvements to account for the “well-documented limitations of Whisper.”