❌

Normal view

There are new articles available, click to refresh the page.
Before yesterdayMain stream

Why AI language models choke on too much text

Large language models represent text using tokens, each of which is a few characters. Short words are represented by a single token (like "the" or "it"), whereas larger words may be represented by several tokens (GPT-4o represents "indivisible" with "ind," "iv," and "isible").

When OpenAI released ChatGPT two years ago, it had a memoryβ€”known as a context windowβ€”of just 8,192 tokens. That works out to roughly 6,000 words of text. This meant that if you fed it more than about 15 pages of text, it would β€œforget” information from the beginning of its context. This limited the size and complexity of tasks ChatGPT could handle.

Today’s LLMs are far more capable:

Read full article

Comments

Β© Aurich Lawson | Getty Images

Not to be outdone by OpenAI, Google releases its own β€œreasoning” AI model

19 December 2024 at 13:49

It's been a really busy month for Google as it apparently endeavors to outshine OpenAI with a blitz of AI releases. On Thursday, Google dropped its latest party trick: Gemini 2.0 Flash Thinking Experimental, which is a new AI model that uses runtime "reasoning" techniques similar to OpenAI's o1 to achieve "deeper thinking" on problems fed into it.

The experimental model builds on Google's newly released Gemini 2.0 Flash and runs on its AI Studio platform, but early tests conducted by TechCrunch reporter Kyle Wiggers reveal accuracy issues with some basic tasks, such as incorrectly counting that the word "strawberry" contains two R's.

These so-called reasoning models differ from standard AI models by incorporating feedback loops of self-checking mechanisms, similar to techniques we first saw in early 2023 with hobbyist projects like "Baby AGI." The process requires more computing time, often adding extra seconds or minutes to response times. Companies have turned to reasoning models as traditional scaling methods at training time have been showing diminishing returns.

Read full article

Comments

Β© Alan Schein via Getty Images

Google releases its own β€˜reasoning’ AI model

19 December 2024 at 09:22

Google has released what it’s calling a new β€œreasoning” AI model β€” but it’s in the experimental stages, and from our brief testing, there’s certainly room for improvement. The new model, called Gemini 2.0 Flash Thinking Experimental (a mouthful, to be sure), is available in AI Studio, Google’s AI prototyping platform. A model card describes […]

Β© 2024 TechCrunch. All rights reserved. For personal use only.

Google DeepMind unveils a new video model to rival Sora

16 December 2024 at 09:00

Google DeepMind, Google’s flagship AI research lab, wants to beat OpenAI at the video-generation game β€” and it might just, at least for a little while. On Monday, DeepMind announced Veo 2, a next-gen video-generating AI and the successor to Veo, which powers a growing number of products across Google’s portfolio. Veo 2 can create […]

Β© 2024 TechCrunch. All rights reserved. For personal use only.

Watch Google DeepMind’s Genie 2 generate playable 3D worlds

11 December 2024 at 11:00

Last week, Google DeepMind announced Genie 2, a new video model that generates plausible, consistent, playable 3D environments based on a prompt image. DeepMind claims that Genie 2 has a slew of new and exciting emerging capabilities that improve the lighting, reflections, and can even generate videos from real-world images. Within these generated worlds, the […]

Β© 2024 TechCrunch. All rights reserved. For personal use only.

❌
❌