Close Menu
  • Home
  • AI
  • Education
  • Entertainment
  • Food Health
  • Health
  • Sports
  • Tech
  • Well Being

Subscribe to Updates

Subscribe to our newsletter and never miss our latest news

Subscribe my Newsletter for New Posts & tips Let's stay updated!

What's Hot

NYC nursing walkout ends as last striking nurses approve new contract

February 22, 2026

Altman Says Musk’s Idea to Put Data Centers in Space Is ‘Ridiculous’

February 21, 2026

Sam Altman would like remind you that humans use a lot of energy, too

February 21, 2026
Facebook X (Twitter) Instagram
  • Home
  • About Us
  • Advertise With Us
  • Contact us
  • DMCA
  • Privacy Policy
  • Terms & Conditions
Facebook X (Twitter) Instagram
IQ Times Media – Smart News for a Smarter YouIQ Times Media – Smart News for a Smarter You
  • Home
  • AI
  • Education
  • Entertainment
  • Food Health
  • Health
  • Sports
  • Tech
  • Well Being
IQ Times Media – Smart News for a Smarter YouIQ Times Media – Smart News for a Smarter You
Home » Inception raises $50 million to build diffusion models for code and text
AI

Inception raises $50 million to build diffusion models for code and text

IQ TIMES MEDIABy IQ TIMES MEDIANovember 6, 2025No Comments3 Mins Read
Facebook Twitter Pinterest LinkedIn Tumblr Email
Share
Facebook Twitter LinkedIn Pinterest Email


With so much money flooding into AI startups, it’s a good time to be an AI researcher with an idea to test out. And if the idea is novel enough, it might be easier to get the resources you need as an independent company instead of inside one of the big labs.

That’s the story of Inception, a startup developing diffusion-based AI models that just raised $50 million in seed funding led by Menlo Ventures. Andrew Ng and Andrej Karpathy provided additional angel funding.

The leader of the project is Stanford professor Stefano Ermon, whose research focuses on diffusion models — which generate outputs through iterative refinement rather than word-by-word. These models power image-based AI systems like Stable Diffusion, Midjourney and Sora. Having worked on those systems since before the AI boom made them exciting, Ermon is using Inception to apply the same models to a broader range of tasks.

Together with the funding, the company released a new version of its Mercury model, designed for software development. Mercury has already been integrated into a number of development tools, including ProxyAI, Buildglare, and Kilo Code. Most importantly, Ermon says the diffusion approach will help Inception’s models conserve on two of the most important metrics: latency (response time) and compute cost.

“These diffusion-based LLMs are much faster and much more efficient than what everybody else is building today,” Ermon says. “It’s just a completely different approach where there is a lot of innovation that can still be brought to the table.”

Understanding the technical difference requires a bit of background. Diffusion models are structurally different from auto-regression models, which dominate text-based AI services. Auto-regression models like GPT-5 and Gemini work sequentially, predicting each next word or word fragment based on the previously processed material. Diffusion models, trained for image generation, take a more holistic approach, modifying the overall structure of a response incrementally until it matches the desired result.

The conventional wisdom is to use auto-regression models for text applications, and that approach has been hugely successful for recent generations of AI models. But a growing body of research suggests diffusion models may perform better when a model is processing large quantities of text or managing data constraints. As Ermon tells it, those qualities become a real advantage when performing operations over large codebases.

Techcrunch event

San Francisco
|
October 13-15, 2026

Diffusion models also have more flexibility in how they utilize hardware, a particularly important advantage as the infrastructure demands of AI become clear. Where auto-regression models have to execute operations one after another, diffusion models can process many operations simultaneously, allowing for significantly lower latency in complex tasks.

“We’ve been benchmarked at over 1,000 tokens per second, which is way higher than anything that’s possible using the existing autoregressive technologies,” Ermon says, “because our thing is built to be parallel. It’s built to be really, really fast.”



Source link

Share. Facebook Twitter Pinterest LinkedIn Tumblr Email
IQ TIMES MEDIA
  • Website

Related Posts

Sam Altman would like remind you that humans use a lot of energy, too

February 21, 2026

Microsoft’s new gaming CEO vows not to flood the ecosystem with ‘endless AI slop’

February 21, 2026

Google VP warns that two types of AI startups may not survive

February 21, 2026
Add A Comment
Leave A Reply Cancel Reply

Editors Picks

Court clears way for Louisiana law requiring Ten Commandments in classrooms to take effect

February 20, 2026

Endangered Floreana tortoises reintroduced to native island habitat

February 20, 2026

Kentucky Supreme Court rules charter school funding unconstitutional

February 19, 2026

Students in Savannah grieve teacher killed in crash with driver pursued by ICE

February 19, 2026
Education

Court clears way for Louisiana law requiring Ten Commandments in classrooms to take effect

By IQ TIMES MEDIAFebruary 20, 20260

The 5th Circuit Court of Appeals cleared the way Tuesday for a Louisiana law requiring…

Endangered Floreana tortoises reintroduced to native island habitat

February 20, 2026

Kentucky Supreme Court rules charter school funding unconstitutional

February 19, 2026

Students in Savannah grieve teacher killed in crash with driver pursued by ICE

February 19, 2026
IQ Times Media – Smart News for a Smarter You
Facebook X (Twitter) Instagram Pinterest Vimeo YouTube
  • Home
  • About Us
  • Advertise With Us
  • Contact us
  • DMCA
  • Privacy Policy
  • Terms & Conditions
© 2026 iqtimes. Designed by iqtimes.

Type above and press Enter to search. Press Esc to cancel.