❌

Normal view

There are new articles available, click to refresh the page.
Before yesterdayMain stream

A stealth AI model beat DALL-E and Midjourney on a popular benchmark β€” its creator just landed $30M

5 May 2025 at 14:12
Recraft, the startup behind a mysterious image model that beat OpenAI’s DALL-E and Midjourney on a respected industry benchmark last year, has raised a $30 million Series B round led by Accel, it exclusively told TechCrunch.Β  Other investors in the round include Khosla Ventures and Madrona. Based in San Francisco, Recraft previously raised a $12 […]

Midjourney introduces first new image generation model in over a year

AI image generator Midjourney released its first new model in quite some time today; dubbed V7, it's a ground-up rework that is available in alpha to users now.

There are two areas of improvement in V7: the first is better images, and the second is new tools and workflows.

Starting with the image improvements, V7 promises much higher coherence and consistency for hands, fingers, body parts, and "objects of all kinds." It also offers much more detailed and realistic textures and materials, like skin wrinkles or the subtleties of a ceramic pot.

Read full article

Comments

Β© Xeophon

Midjourney releases V7, its first new AI image model in nearly a year

3 April 2025 at 21:59
Midjourney, one of the earliest AI image-generating services on the web, has released its first new AI image model in nearly a year. Dubbed V7, the model began rolling out in alpha around midnight EST on Thursday, comes a week after OpenAI debuted a new image generator in ChatGPT that quickly went viral for its […]

New AI text diffusion models break speed barriers by pulling words from noise

27 February 2025 at 13:14

On Thursday, Inception Labs released Mercury Coder, a new AI language model that uses diffusion techniques to generate text faster than conventional models. Unlike traditional models that create text word by wordβ€”such as the kind that powers ChatGPTβ€”diffusion-based models like Mercury produce entire responses simultaneously, refining them from an initially masked state into coherent text.

Traditional large language models build text from left to right, one token at a time. They use a technique called "autoregression." Each word must wait for all previous words before appearing. Inspired by techniques from image-generation models like Stable Diffusion, DALL-E, and Midjourney, text diffusion language models like LLaDA (developed by researchers from Renmin University and Ant Group) and Mercury use a masking-based approach. These models begin with fully obscured content and gradually "denoise" the output, revealing all parts of the response at once.

While image diffusion models add continuous noise to pixel values, text diffusion models can't apply continuous noise to discrete tokens (chunks of text data). Instead, they replace tokens with special mask tokens as the text equivalent of noise. In LLaDA, the masking probability controls the noise level, with high masking representing high noise and low masking representing low noise. The diffusion process moves from high noise to low noise. Though LLaDA describes this using masking terminology and Mercury uses noise terminology, both apply a similar concept to text generation rooted in diffusion.

Read full article

Comments

Β© akinbostanci via Getty Images

❌
❌