News
ControlNets for Stable Diffusion 3.5 Large
Today we are adding new capabilities to Stable Diffusion 3.5 Large by releasing three ControlNets: Blur, Canny, and Depth.
Introducing Stable Diffusion 3.5
Today we are introducing Stable Diffusion 3.5. This open release includes multiple model variants, including Stable Diffusion 3.5 Large and Stable Diffusion 3.5 Large Turbo, and as of October 29th, Stable Diffusion 3.5 Medium.
Introducing Stable Fast 3D: Rapid 3D Asset Generation From Single Images
We are excited to introduce Stable Fast 3D, Stability AI’s latest breakthrough in 3D asset generation technology. This innovative model transforms a single input image into a detailed 3D asset, setting a new standard for speed and quality in the field of 3D reconstruction.
Introducing Stable Video 4D, Our Latest AI Model for Dynamic Multi-Angle Video Generation
We are pleased to announce the availability of Stable Video 4D, an innovative model that allows users to upload a single video and receive dynamic novel-view videos of eight new angles/views, delivering a new level of versatility and creativity.
Stability AI Releases Stable Assistant Features
Today we’re announcing new features to Stable Assistant, our user-friendly chatbot that leverages Stable Image Ultra, our most advanced image generation technology based on Stable Diffusion 3.
Announcing the Open Release of Stable Diffusion 3 Medium, Our Most Sophisticated Image Generation Model to Date
We are excited to announce the launch of Stable Diffusion 3 Medium, the latest and most advanced text-to-image AI model in our Stable Diffusion 3 series.
Introducing Stable Audio Open - An Open Source Model for Audio Samples and Sound Design
We’re excited to announce Stable Audio Open, an open source model optimised for generating short audio samples, sound effects and production elements using text prompts.
Stable Diffusion 3 API Now Available
We are pleased to announce the availability of Stable Diffusion 3 and Stable Diffusion 3 Turbo on the Stability AI Developer Platform API.
Introducing Stable LM 2 12B
Introducing the latest additions to our Stable LM 2 language model series: a 12 billion parameter base model and an instruction-tuned variant, trained on 2 trillion tokens in seven languages: English, Spanish, German, Italian, French, Portuguese, and Dutch. This medium-sized model balances strong performance, efficiency, memory requirements, and speed, following our established Stable LM 2 1.6B framework as detailed in our previously released technical report.
Introducing Stable Audio 2.0
Today, we are pleased to introduce Stable Audio 2.0. This model enables high-quality, full tracks with coherent musical structure up to three minutes long at 44.1 kHz stereo from a single natural language prompt.
Introducing Stable Code Instruct 3B
Stable Code Instruct 3B is our latest instruction-tuned large language model, built on top of Stable Code 3B. This model enhances code completion and supports natural language interactions, aiming to improve the efficiency and intuitiveness of programming and software development related tasks. Our analysis suggests that Stable Code Instruct 3B outperforms comparable models such as Codellama 7B Instruct, and DeepSeek-Coder Instruct 1.3B in various coding-related tasks.
Image Services on Stability AI Developer Platform
Today we are introducing a powerful suite of image services on the Stability AI Developer Platform API to enhance media creation. Utilizing our cutting-edge image models, these services offer developers and builders comprehensive tools for image generation, enhancement, outpainting, and editing, streamlining the development of user-focused solutions and simplifying high-quality image production without complex prompt engineering.
Introducing Stable Video 3D: Quality Novel View Synthesis and 3D Generation from Single Images
When we released Stable Video Diffusion, we highlighted the versatility of our video model across various applications. Building upon this foundation, we are excited to release Stable Video 3D. This new model advances the field of 3D technology, delivering greatly improved quality and multi-view when compared to the previously released Stable Zero123, as well as outperforming other open source alternatives such as Zero123-XL.
Introducing TripoSR: Fast 3D Object Generation from Single Images
We have partnered with Tripo AI to develop TripoSR, a fast 3D object reconstruction model inspired by the recent work of LRM: Large Reconstruction Model For Single Image to 3D. This new image-to-3D model is designed to cater to the growing demands of entertainment, gaming, industrial design, and architecture professionals, with responsive outputs for visualizing detailed 3D objects.
Stable Diffusion 3
Announcing Stable Diffusion 3 in early preview, our most capable text-to-image model with greatly improved performance in multi-subject prompts, image quality, and spelling abilities.
Introducing Stable Cascade
Today marks the launch of Stable Cascade in its research preview. This innovative text to image model introduces an interesting three-stage approach, setting new benchmarks for quality, flexibility, fine-tuning, and efficiency with a focus on further eliminating hardware barriers.
Introducing Stable LM 2 1.6B
Today, we are introducing our first language model from the new Stable LM 2 series: the 1.6 billion parameter base model and an instruction-tuned version. The base model is trained on…
Stable Code 3B: Coding on the Edge
Stable Code, an upgrade from Stable Code Alpha 3B, specializes in code completion and outperforms predecessors in efficiency and multi-language support. It is compatible with standard laptops, including non-GPU models, and features capabilities like FIM and expanded context size.
Stable Video Diffusion Now Available on Stability AI Developer Platform API
Stable Diffusion Video API transforms images into 2-second, high-quality videos. Ideal for businesses and researchers, customization, and a competitive edge in various industries like advertising, TV, and gaming.
Introducing Stable Zero123: Quality 3D Object Generation from Single Images
Stable Zero123 is an AI-powered model for generating novel views of 3D objects with improved quality. Released for non-commercial and research purposes, it uses an improved dataset and elevation conditioning for higher-quality predictions.