❌

Normal view

There are new articles available, click to refresh the page.
Today β€” 4 March 2025Main stream

Amazon is working on a new 'reasoning' AI model that competes with OpenAI and Anthropic

4 March 2025 at 02:00
AWS VP Ruba Borno (left) and CEO Matt Garman
AWS VP Ruba Borno (left) and CEO Matt Garman

Amazon

  • Amazon plans to launch a new AI model with advanced reasoning capabilities.
  • The model aims to offer hybrid reasoning, a mix of quick answers and more complex thinking.
  • Amazon is prioritizing cost efficiency and external benchmark performance.

Amazon is building its own AI model that incorporates advanced "reasoning" capabilities, Business Insider has learned.

The offering is tentatively scheduled to launch by June under the Nova brand, a group of generative AI models Amazon unveiled late last year, according to a person directly involved in the project. This person asked not to be identified because they were not authorized to speak with the media.

Amazon wants the new model to take a "hybrid reasoning" approach that provides quick answers and more complex extended thinking within a single system, this person added. An Amazon spokesperson didn't respond to a request for comment.

Reasoning models have recently become the next frontier in AI. They often work more slowly but can also tackle tougher problems by trying multiple solutions and backtracking via chain-of-thought techniques. Companies including Google, OpenAI, and Anthropic have released their own reasoning models recently, while DeepSeek drew a lot of attention for building a similar offering more efficiently.

One of Amazon's priorities is to make its Nova reasoning model more price-efficient than competitors, which include OpenAI's o1, Anthropic's Claude 3.7 Sonnet, and Google's Gemini 2.0 Flash Thinking, according to the person involved in the project.

Amazon previously said that its existing in-house Nova models are at least 75% cheaper than third-party models available via its Bedrock AI development platform.

Another goal is to get it Amazon's upcoming reasoning model ranked in the top 5 for performance, based on external benchmarks that evaluate software development and math skills, such as the SWE, Berkeley Function Calling Leaderboard, and AIME, among others, this person added.

The move reflects Amazon's commitment to invest in its own family of AI models, even as it preaches the need to offer a variety of model choices through Bedrock. Amazon's AGI team, run by head scientist Rohit Prasad, has been working on this new model.

It also puts Amazon in more direct competition with Anthropic, the AI startup that just launched its newest model. Claude 3.7 Sonnet uses a similar hybrid approach, combining quick answers and longer chain-of-thought outputs.

Amazon has invested $8 billion in Anthropic so far, and the two companies have been close partners, collaborating in areas including AI chips and cloud computing.

Have a tip? Contact this reporter via email at [email protected] or Signal, Telegram, or WhatsApp at 650-942-3061. Use a personal email address and a nonwork device; here's our guide to sharing information securely.

Read the original article on Business Insider

Before yesterdayMain stream

Ten cool science stories we almost missed

There is rarely time to write about every cool science paper that comes our way; many worthy candidates sadly fall through the cracks over the course of the year. But as 2024 comes to a close, we've gathered ten of our favorite such papers at the intersection of science and culture as a special treat, covering a broad range of topics: from reenacting Bronze Age spear combat and applying network theory to the music of Johann Sebastian Bach, to Spider-Man inspired web-slinging tech and a mathematical connection between a turbulent phase transition and your morning cup of coffee. Enjoy!

Reenacting Bronze Age spear combat

Experiment with experienced fighters who spar freely using different styles. An experiment with experienced fighters who spar freely using different styles. Credit: Valerio Gentile/CC BY

The European Bronze Age saw the rise of institutionalized warfare, evidenced by the many spearheads and similar weaponry archaeologists have unearthed. But how might these artifacts be used in actual combat? Dutch researchers decided to find out by constructing replicas of Bronze Age shields and spears and using them in realistic combat scenarios. They described their findings in an October paper published in the Journal of Archaeological Science.

There have been a couple of prior experimental studies on bronze spears, but per Valerio Gentile (now at the University of Gottingen) and coauthors, practical research to date has been quite narrow in scope, focusing on throwing weapons against static shields. Coauthors C.J. van Dijk of the National Military Museum in the Netherlands and independent researcher O. Ter Mors each had more than a decade of experience teaching traditional martial arts, specializing in medieval polearms and one-handed weapons. So they were ideal candidates for testing the replica spears and shields.

Read full article

Comments

Β© APS/Carin Cain

Chip startups are making these New Year's resolutions to take on Nvidia in 2025

27 December 2024 at 02:00
Jensen Huang speaking on stage
Nvidia CEO Jensen Huang.

Chip Somodevilla/Getty Images

  • The AI computing market may shift in 2025, opening opportunities for smaller companies.
  • Nvidia dominates AI computing. Evolving workloads could benefit competitors.
  • Companies like Groq, Positron, and SambaNova focus on inference to challenge Nvidia's market hold.

In 2025, the tides may turn for companies hoping to compete with the $3 trillion gorilla in AI computing.

Nvidia holds an estimated 90% of the market share for AI computing. Still, as the use of AI grows,Β workloads are expected to change, and this evolution may give companies with competitive hardware an opening.

In 2024, the majority of AI compute spend shifted to inference, Thomas Sohmers, CEO of chip startup Positron AI, told BI. This will "continue to grow on what looks like an exponential curve," he added.

In AI, inference is the computation needed to produce the response to a user's query or request. The computing required to teach the model the knowledge needed to answer is called "training." Creating OpenAI's image generation platform Sora, for example, represents training. Each user who instructs it to create an image represents an inference workload.

OpenAI's other models have Sohmers and others excited about the growth in computing needs in 2025.

OpenAI's o1 and o3, Google's Gemini 2.0 Flash Thinking, and a handful of other AI models useΒ more compute-intensive strategies to improve results after training. These strategies are often called inference-time computing, chain-of-thought, chain-of-reasoning, or reasoning models.

Simply put, if the models think more before they answer, the responses are better. That thinking comes at a cost of time and money.

The startups vying for some of Nvidia's market share are attempting to optimize one or both.

Nvidia already benefits from these innovations, CEO Jensen Huang said on the company's November earnings call. Huang's wannabe competitors are betting that in 2025, new post-training strategies for AI will benefit all purveyors of inference chips.

Business Insider spoke to three challengers about their hopes and expectations for 2025. Here are their New Year's resolutions.

What's one thing within your control that could make 2025 a big year for alternative chips?

A tattooed man in a black shirt and jeans stands on a stage with a pink and black background that read Groq: what's next?
Mark Heaps is the chief technology evangelist for Nvidia challenger Groq.

Groq

Mark Heaps, chief technology evangelist, Groq:

"Execution, execution, execution. Right now, everybody at Groq has decided not to take a holiday break this year. Everyone is executing and building the systems. We are all making sure that we deliver to the opportunity that we've got because that is in our control.

I tell everyone our funnel right now is carbonated and bubbling over. It's unbelievable, the amount of customer interest. We have to build more systems, and we have to stand up those systems so we can serve the demand that we've got. We want to serve all those customers. We want to increase rate limits for everybody."

Rodrigo Liang, CEO, SambaNova Systems:

"For SambaNova, the most critical factor is executing on the shift from training to inference. The industry is moving rapidly toward real-time applications, and inference workloads are becoming the lion's share of AI demand. Our focus is on ensuring our technology enables enterprises to scale efficiently and sustainably."

Thomas Sohmers, CEO, Positron:

"My belief is if we can actually deploy enough compute β€” which thankfully I think we can from a supply chain perspective β€” by deploying significantly more inference-specific compute, we're going to be able to grow the adoption rate of 'chain of thoughts' and other inference-additional compute."

What's one thing you're hoping for that's not in your control for 2025?

Rodrigo Liang SambaNova Systems
Rodrigo Liang, CEO and cofounder of SambaNova Systems.

SambaNova Systems

Heaps:

"It's about customers recognizing that there are novel advancements against incumbent technologies. There's a lot of folks that have told us, 'We like what you have, but to use the old adage and rephrase it: No one ever got fired for buying from β€” insert incumbent.'

But we know that it's starting to boil up. People are realizing it's hard for them to get chips from the incumbent, and it's also not as performant as Groq is. So my wish would be that people are willing to take that chance and actually look to some of these new technologies."

Liang:

"If I had a magic wand, I'd address the power challenges around deploying AI. Today, most of the market is stuck using power-hungry hardware that wasn't designed for inference at scale. The result is an unsustainable approach β€” economically and environmentally.

At SambaNova, we've proven there's a better way. Our architecture consumes 10 times less power, making it possible for enterprises to deploy AI systems that meet their goals without blowing past their power budgets or carbon targets. I'd like to see the market move faster toward adopting technologies that prioritize efficiency and sustainability β€” because that's how we ensure AI can scale globally without overwhelming the infrastructure that supports it."

Sohmers:

"I would like people to actually adopt these chain of thought capabilities at the fastest rate possible. I think that is a huge shift β€” from a capabilities perspective. You have 8 billion parameter models surpassing 70 billion parameter models. So I'm trying to do everything I can to make that happen."

What's your New Year's resolution?

Positron AI executives stand near the startup's products
Positron AI executives. From left to right: Edward Kmett, Thomas Sohmers, Adam Huson, and Greg Davis.

Positron AI

Heaps:

"In the last six months, I've gone to a number of hackathons, and I've met developers. It's deeply inspiring. So my New Year's resolution is to try to amplify the signal of the good that people are doing with AI."

Liang:

"Making time for music. Playing guitar is something I've always loved, and I would love to get back into it. Music has this incredible way of clearing the mind and sparking creativity, which I find invaluable as we work to bring SambaNova's AI to new corners of the globe."

Sohmers:

I want to do as much to encourage the usage of these new tools to help, you know, my mom. Part of the reason I got into technology was because I wanted to see these tools lift up people to be able to do more with their time β€” to learn everything that they want beyond whatever job they're in. I think that bringing the cost down of these things will enable that proliferation.

I also personally want to see and try to use more of these things outside of my just work context because I've been obsessively using the o1 Pro model for the past few weeks, and it's been amazing for my personal work. But when I gave access to my mom what she would do with it was pretty interesting β€” those sort of normal, everyday person tasks for these things where it truly is being an assistant."

Read the original article on Business Insider

Amazon announces Nova, a new family of multimodal AI models

3 December 2024 at 09:57

At its re:Invent conference on Tuesday, Amazon Web Services (AWS), Amazon’s cloud computing division, announced a new family of multimodal generative AI models it calls Nova. There are four text-generating models in total: Micro, Lite, Pro, and Premier. Micro, Lite, and Pro are available Tuesday to AWS customers, while Premier will arrive in early 2025, […]

Β© 2024 TechCrunch. All rights reserved. For personal use only.

❌
❌