Inception Unveils Mercury: A New Era for AI Models
Inception, a Palo Alto-based company founded by Stanford professor Stefano Ermon, has introduced a new type of AI model called Mercury, which is part of their diffusion-based large language models (dLLMs). Announced on their website, Mercury is designed to be significantly faster and more cost-effective than existing large language models (LLMs).
The Mercury family of models leverages diffusion technology, traditionally used in image and audio generation, to enhance text generation capabilities. Unlike traditional LLMs that generate text sequentially, Mercury's diffusion models generate and refine text in parallel, allowing for faster processing speeds. This approach enables Mercury to achieve speeds of over 1000 tokens per second on standard hardware, a feat previously only possible with specialized chips.
Inception offers Mercury through an API and supports on-premises deployments, making it accessible for various enterprise applications. The company claims that its models can run up to 10 times faster and at a fraction of the cost of traditional models, providing a significant advantage in latency-sensitive applications. Mercury Coder, a model optimized for code generation, is already available for testing and has shown impressive performance on standard coding benchmarks.
With the introduction of Mercury, Inception aims to set a new standard for AI models, offering enhanced speed and efficiency that could transform how language models are utilized across industries.
We hope you enjoyed this article.
Consider subscribing to one of several newsletters we publish. For example, in the Daily AI Brief you can read the most up to date AI news round-up 6 days per week.
Also, consider following our LinkedIn page AI Brief.
Subscribe to Daily AI Brief
Daily report covering major AI developments and industry news, with both top stories and complete market updates