Stability AI · General LLM
StableLM
Stability AI's open-source language model series designed for transparent, accessible AI with strong performance at smaller parameter counts.
Overview
StableLM is Stability AI's family of open-source language models, continuing the company's mission to make AI technology accessible. The StableLM 2 series offers 1.6B and 12B parameter models with training transparency and permissive licensing. StableLM Zephyr, the instruction-tuned variant, demonstrates competitive chat performance for its size. The models are designed to provide a fully open alternative for developers who need transparent, auditable language model foundations.
Parameters
1.6B / 12B (StableLM 2)
Context Window
4K tokens
Training Data
2T tokens (diverse multilingual)
Architecture
Decoder-only transformer
License
Stability AI Community License / Apache 2.0
Capabilities
General-purpose text generation
Conversational AI through instruction-tuned variants
Code understanding and generation
Lightweight deployment on consumer hardware
Use Cases
Building transparent AI applications with fully open model weights
Deploying lightweight chat assistants on modest hardware
Research into language model training and behavior
Creating open-source AI products without licensing restrictions
Pros
- +Fully open-source with transparent training methodology
- +Compact models suitable for resource-constrained deployments
- +Active community and regular model updates
- +No restrictions on commercial use for most variants
Cons
- -Smaller models lag behind frontier models on complex tasks
- -Stability AI's business challenges create uncertainty about support
- -Shorter context window limits document processing capabilities
- -Less competitive on benchmarks compared to peers like Phi-3 or Gemma
Pricing
Free and open-source. Small enough to run on consumer laptops. No API costs when self-hosted.