❌

Normal view

There are new articles available, click to refresh the page.
Before yesterdayMain stream

AI video just took a startling leap in realism. Are we doomed?

Last week, Google introduced Veo 3, its newest video generation model that can create 8-second clips with synchronized sound effects and audio dialogβ€”a first for the company's AI tools. The model, which generates videos at 720p resolution (based on text descriptions called "prompts" or still image inputs), represents what may be the most capable consumer video generator to date, bringing video synthesis close to a point where it is becoming very difficult to distinguish between "authentic" and AI-generated media.

Google also launched Flow, an online AI filmmaking tool that combines Veo 3 with the company's Imagen 4 image generator and Gemini language model, allowing creators to describe scenes in natural language and manage characters, locations, and visual styles in a web interface.

An AI-generated video from Veo 3: "ASMR scene of a woman whispering "Moonshark" into a microphone while shaking a tambourine"

Both tools are now available to US subscribers of Google AI Ultra, a plan that costs $250 a month and comes with 12,500 credits. Veo 3 videos cost 150 credits per generation, allowing 83 videos on that plan before you run out. Extra credits are available for the price of 1 cent per credit in blocks of $25, $50, or $200. That comes out to about $1.50 per video generation. But is the price worth it? We ran some tests with various prompts to see what this technology is truly capable of.

Read full article

Comments

Β© Google

Google’s Will Smith double is better at eating AI spaghetti … but it’s crunchy?

On Tuesday, Google launched Veo 3, a new AI video synthesis model that can do something no major AI video generator has been able to do before: create a synchronized audio track. While from 2022 to 2024, we saw early steps in AI video generation, each video was silent and usually very short in duration. Now you can hear voices, dialog, and sound effects in eight-second high-definition video clips.

Shortly after the new launch, people began asking the most obvious benchmarking question: How good is Veo 3 at faking Oscar-winning actor Will Smith at eating spaghetti?

First, a brief recap. The spaghetti benchmark in AI video traces its origins back to March 2023, when we first covered an early example of horrific AI-generated video using an open source video synthesis model called ModelScope. The spaghetti example later became well-known enough that Smith parodied it almost a year later in February 2024.

Read full article

Comments

Google adds Veo 2 video generation to Gemini app

Google has announced that yet another AI model is coming to Gemini, but this time, it's more than a chatbot. The company's Veo 2 video generator is rolling out to the Gemini app and website, giving paying customers a chance to create short video clips with Google's allegedly state-of-the-art video model.

Veo 2 works like other video generators, including OpenAI's Soraβ€”you input text describing the video you want, and a Google data center churns through tokens until it has an animation. Google claims that Veo 2 was designed to have a solid grasp of real-world physics, particularly the way humans move. Google's examples do look good, but presumably that's why they were chosen.

Prompt: Aerial shot of a grassy cliff onto a sandy beach where waves crash against the shore, a prominent sea stack rises from the ocean near the beach, bathed in the warm, golden light of either sunrise or sunset, capturing the serene beauty of the Pacific coastline.

Read full article

Comments

A new, uncensored AI video model may spark a new AI hobbyist movement

19 December 2024 at 07:50

The AI-generated video scene has been hopping this year (or twirling wildly, as the case may be). This past week alone we've seen releases or announcements of OpenAI's Sora, Pika AI's Pika 2, Google's Veo 2, and Minimax's video-01-live. It's frankly hard to keep up, and even tougher to test them all. But recently, we put a new open-weights AI video synthesis model, Tencent's HunyuanVideo, to the testβ€”and it's surprisingly capable for being a "free" model.

Unlike the aforementioned models, HunyuanVideo's neural network weights are openly distributed, which means they can be run locally under the right circumstances (people have already demonstrated it on a consumer 24 GB VRAM GPU) and it can be fine-tuned or used with LoRAs to teach it new concepts.

Notably, a few Chinese companies have been at the forefront of AI video for most of this year, and some experts speculate that the reason is less reticence to train on copyrighted materials, use images and names of famous celebrities, and incorporate some uncensored video sources. As we saw with Stable Diffusion 3's mangled release, including nudity or pornography in training data may allow these models achieve better results by providing more information about human bodies. HunyuanVideo notably allows uncensored outputs, so unlike the commercial video models out there, it can generate videos of anatomically realistic, nude humans.

Read full article

Comments

Β© Tencent

Twirling body horror in gymnastics video exposes AI’s flaws

13 December 2024 at 06:12

On Wednesday, a video from OpenAI's newly launched Sora AI video generator went viral on social media, featuring a gymnast who sprouts extra limbs and briefly loses her head during what appears to be an Olympic-style floor routine.

As it turns out, the nonsensical synthesis errors in the videoβ€”what we like to call "jabberwockies"β€”hint at technical details about how AI video generators work and how they might get better in the future.

But before we dig into the details, let's take a look at the video.

Read full article

Comments

Β© OpenAI / Deedy

Ten months after first tease, OpenAI launches Sora video generation publicly

On Monday, OpenAI released Sora Turbo, a new version of its text-to-video generation model, making it available to ChatGPT Plus and Pro subscribers through a dedicated website. The model generates videos up to 20 seconds long at resolutions reaching 1080 p from a text or image prompt.

Open AI announced that Sora would be available today for ChatGPT Plus and Pro subscribers in the US and many parts of the world but is not yet available in Europe. As of early Monday afternoon, though, even existing Plus subscribers trying to use the tool are being presented with a message that "sign ups are temporarily unavailable" thanks to "heavy traffic."

Out of an abundance of caution, OpenAI is limiting Sora's ability to generate videos of people for the time being. At launch, uploads involving human subjects face restrictions while OpenAI refines its deepfake prevention systems. The platform also blocks content involving CSAM and sexual deepfakes. OpenAI says it maintains an active monitoring system and conducted testing to identify potential misuse scenarios before release.

Read full article

Comments

Β© OpenAI

❌
❌