❌

Reading view

There are new articles available, click to refresh the page.

12 days of OpenAI: The Ars Technica recap

Over the past 12 business days, OpenAI has announced a new product or demoed an AI feature every weekday, calling the PR event "12 days of OpenAI." We've covered some of the major announcements, but we thought a look at each announcement might be useful for people seeking a comprehensive look at each day's developments.

The timing and rapid pace of these announcementsβ€”particularly in light of Google's competing releasesβ€”illustrates the intensifying competition in AI development. What might normally have been spread across months was compressed into just 12 business days, giving users and developers a lot to process as they head into 2025.

Humorously, we asked ChatGPT what it thought about the whole series of announcements, and it was skeptical that the event even took place. "The rapid-fire announcements over 12 days seem plausible," wrote ChatGPT-4o, "But might strain credibility without a clearer explanation of how OpenAI managed such an intense release schedule, especially given the complexity of the features."

Read full article

Comments

Β© J Studios via Getty Images

OpenAI announces o3 and o3-mini, its next simulated reasoning models

On Friday, during Day 12 of its "12 days of OpenAI," OpenAI CEO Sam Altman announced its latest AI "reasoning" models, o3 and o3-mini, which build upon the o1 models launched earlier this year. The company is not releasing them yet but will make these models available for public safety testing and research access today.

The models use what OpenAI calls "private chain of thought," where the model pauses to examine its internal dialog and plan ahead before responding, which you might call "simulated reasoning" (SR)β€”a form of AI that goes beyond basic large language models (LLMs).

The company named the model family "o3" instead of "o2" to avoid potential trademark conflicts with British telecom provider O2, according to The Information. During Friday's livestream, Altman acknowledged his company's naming foibles, saying, "In the grand tradition of OpenAI being really, truly bad at names, it'll be called o3."

Read full article

Comments

Β© Benj Edwards / Andriy Onufriyenko via Getty Images

Call ChatGPT from any phone with OpenAI’s new 1-800 voice service

On Wednesday, OpenAI launched a 1-800-CHATGPT (1-800-242-8478) telephone number that anyone in the US can call to talk to ChatGPT via voice chat for up to 15 minutes for free. The company also says that people outside the US can send text messages to the same number for free using WhatsApp.

Upon calling, users hear a voice say, "Hello again, it's ChatGPT, an AI assistant. Our conversation may be reviewed for safety. How can I help you?" Callers can ask ChatGPT anything they would normally ask the AI assistant and have a live, interactive conversation.

During a livestream demo of "Calling with ChatGPT" during Day 10 of "12 Days of OpenAI," OpenAI employees demonstrated several examples of the telephone-based voice chat in action, asking ChatGPT to identify a distinctive house in California and for help in translating a message into Spanish for a friend. For fun, they showed calls from an iPhone, a flip phone, and a vintage rotary phone.

Read full article

Comments

Β© Charles Taylor via Getty Images

Google goes β€œagentic” with Gemini 2.0’s ambitious AI agent features

On Wednesday, Google unveiled Gemini 2.0, the next generation of its AI-model family, starting with an experimental release called Gemini 2.0 Flash. The model family can generate text, images, and speech while processing multiple types of input including text, images, audio, and video. It's similar to multimodal AI models like GPT-4o, which powers OpenAI's ChatGPT.

"Gemini 2.0 Flash builds on the success of 1.5 Flash, our most popular model yet for developers, with enhanced performance at similarly fast response times," said Google in a statement. "Notably, 2.0 Flash even outperforms 1.5 Pro on key benchmarks, at twice the speed."

Gemini 2.0 Flashβ€”which is the smallest model of the 2.0 family in terms of parameter countβ€”launches today through Google's developer platforms like Gemini API, AI Studio, and Vertex AI. However, its image generation and text-to-speech features remain limited to early access partners until January 2025. Google plans to integrate the tech into products like Android Studio, Chrome DevTools, and Firebase.

Read full article

Comments

Β© Google

Reddit debuts AI-powered discussion searchβ€”but will users like it?

On Monday, Reddit announced it would test an AI-powered search feature called "Reddit Answers" that uses an AI model to create summaries from existing Reddit posts to respond to user questions, reports Reuters.

The feature generates responses by searching through Reddit's vast collection of community discussions and comments. When users ask questions, Reddit Answers provides summaries of relevant conversations and includes links to related communities and posts.

The move potentially puts Reddit in competition with traditional search engines like Google and newer AI search tools like those from OpenAI and Perplexity. But while other companies pull information from across the Internet, Reddit Answers focuses only on content within Reddit's platform.

Read full article

Comments

Β© Reddit

Ten months after first tease, OpenAI launches Sora video generation publicly

On Monday, OpenAI released Sora Turbo, a new version of its text-to-video generation model, making it available to ChatGPT Plus and Pro subscribers through a dedicated website. The model generates videos up to 20 seconds long at resolutions reaching 1080 p from a text or image prompt.

Open AI announced that Sora would be available today for ChatGPT Plus and Pro subscribers in the US and many parts of the world but is not yet available in Europe. As of early Monday afternoon, though, even existing Plus subscribers trying to use the tool are being presented with a message that "sign ups are temporarily unavailable" thanks to "heavy traffic."

Out of an abundance of caution, OpenAI is limiting Sora's ability to generate videos of people for the time being. At launch, uploads involving human subjects face restrictions while OpenAI refines its deepfake prevention systems. The platform also blocks content involving CSAM and sexual deepfakes. OpenAI says it maintains an active monitoring system and conducted testing to identify potential misuse scenarios before release.

Read full article

Comments

Β© OpenAI

OpenAI announces full β€œo1” reasoning model, $200 ChatGPT Pro tier

On Thursday during a live demo as part of its "12 days of OpenAI" event, OpenAI announced a new tier of ChatGPT with higher usage limits for $200 a month and the full version of "o1," the full version of a so-called reasoning model the company debuted in September.

Unlike o1-preview, o1 can now process images as well as text (similar to GPT-4o), and it is reportedly much faster than o1-preview. In a demo question about a Roman emperor, o1 took 14 seconds for an answer, and 1 preview took 33 seconds. According to OpenAI, o1 makes major mistakes 34 percent less often than o1-preview, while "thinking" 50 percent faster. The model will also reportedly become even faster once deployment is finished transitioning the GPUs to the new model.

Whether the new ChatGPT Pro subscription will be worth the $200 a month fee isn't yet fully clear, but the company specified that users will have access to an even more capable version of o1 called "o1 Pro Mode" that will do even deeper reasoning searches and provide "more thinking power for more difficult problems" before answering.

Read full article

Comments

Β© OpenAI / Benj Edwards

Soon, the tech behind ChatGPT may help drone operators decide which enemies to kill

As the AI industry grows in size and influence, the companies involved have begun making stark choices about where they land on issues of life and death. For example, can their AI models be used to guide weapons or make targeting decisions? Different companies have answered this question in different ways, but for ChatGPT maker OpenAI, what started as a hard line against weapons development and military applications has slipped away over time.

On Wednesday, defense-tech company Anduril Industriesβ€”started by Oculus founder Palmer Luckey in 2017β€”announced a partnership with OpenAI to develop AI models (similar to the GPT-4o and o1 models that power ChatGPT) to help US and allied forces identify and defend against aerial attacks.

The companies say their AI models will process data to reduce the workload on humans. "As part of the new initiative, Anduril and OpenAI will explore how leading-edge AI models can be leveraged to rapidly synthesize time-sensitive data, reduce the burden on human operators, and improve situational awareness," Anduril said in a statement.

Read full article

Comments

Β© OpenAI / Getty Images / Benj Edwards

OpenAI teases 12 days of mystery product launches starting tomorrow

On Wednesday, OpenAI CEO Sam Altman announced a "12 days of OpenAI" period starting December 5, which will unveil new AI features and products for 12 consecutive weekdays.

Altman did not specify the exact features or products OpenAI plans to unveil, but a report from The Verge about this "12 days of shipmas" event suggests the products may include a public release of the company's text-to-video model Sora and a new "reasoning" AI model similar to o1-preview. Perhaps we may even see DALL-E 4 or a new image generator based on GPT-4o's multimodal capabilities.

Altman's full tweet included hints at releases both big and small:

Read full article

Comments

Β© ThomasVogel via Getty Images

Certain names make ChatGPT grind to a halt, and we know why

OpenAI's ChatGPT is more than just an AI language model with a fancy interface. It's a system consisting of a stack of AI models and content filters that make sure its outputs don't embarrass OpenAI or get the company into legal trouble when its bot occasionally makes up potentially harmful facts about people.

Recently, that reality made the news when people discovered that the name "David Mayer" breaks ChatGPT. 404 Media also discovered that the names "Jonathan Zittrain" and "Jonathan Turley" caused ChatGPT to cut conversations short. And we know another name, likely the first, that started the practice last year: Brian Hood. More on that below.

The chat-breaking behavior occurs consistently when users mention these names in any context, and it results from a hard-coded filter that puts the brakes on the AI model's output before returning it to the user.

Read full article

Comments

Β© stocksnapper via Getty Images

Amazon pours another $4B into Anthropic, OpenAI’s biggest rival

On Friday, Anthropic announced that Amazon has increased its investment in the AI startup by $4 billion, bringing its total stake to $8 billion while maintaining its minority investor position. Anthropic makes Claude, an AI assistant rival to OpenAI's ChatGPT.

One reason behind the deal involves chips. The computing demands of training large AI models have made access to specialized processors a requirement for AI companies. While Nvidia currently dominates the AI chip market with customers that include most major tech companies, some cloud providers like Amazon have begun developing their own AI-specific processors.

Under the agreement, Anthropic will train and deploy its foundation models using Amazon's custom-built Trainium (for training AI models) and its Inferentia chips (for AI inference, the term for running trained models). The company will also work with Amazon's Annapurna Labs division to advance processor development for AI applications.

Read full article

Comments

Β© Bloomberg via Getty Images

Niantic uses PokΓ©mon Go player data to build AI navigation system

Last week, Niantic announced plans to create an AI model for navigating the physical world using scans collected from players of its mobile games, such as PokΓ©mon Go, and from users of its Scaniverse app, reports 404 Media.

All AI models require training data. So far, companies have collected data from websites, YouTube videos, books, audio sources, and more, but this is perhaps the first we've heard of AI training data collected through a mobile gaming app.

"Over the past five years, Niantic has focused on building our Visual Positioning System (VPS), which uses a single image from a phone to determine its position and orientation using a 3D map built from people scanning interesting locations in our games and Scaniverse," Niantic wrote in a company blog post.

Read full article

Comments

Β© https://www.gettyimages.com/detail/news-photo/man-plays-pokemon-go-game-on-a-smartphone-on-july-22-2016-news-photo/578680184

❌