Llama 3.1 8B
Llama 3.1 CommunityMeta · 8B · Llama 3.1 Community
Meta's versatile 8B — great quality/speed ratio
Find out which AI models your machine can actually run.
Estimates based on browser APIs. Actual specs may vary.
Meta · 8B · Llama 3.1 Community
Meta's versatile 8B — great quality/speed ratio
Alibaba · 9B · Apache 2.0
Multimodal Qwen 3.5 mid-size
Microsoft · 14B · MIT
Microsoft's reasoning-focused model
OpenAI · 21B · Apache 2.0
OpenAI's open-weight MoE with configurable reasoning
Mistral AI · 24B · Apache 2.0
Multimodal Mistral with vision support
Google · 27B · Gemma
Google's flagship Gemma 3 model
Alibaba · 32B · Apache 2.0
Best open-source coding model at release
Alibaba · 32B · Apache 2.0
Qwen 3 flagship dense model
DeepSeek · 32B · MIT
R1 reasoning distilled into Qwen 32B — sweet spot
Meta · 70B · Llama 3.3 Community
Best open model at 70B class
Meta · 109B · Llama 4 Community
MoE with 16 experts, 17B active params
OpenAI · 117B · Apache 2.0
OpenAI's flagship open-weight MoE — 52.6% SWE-bench
Mistral AI · 123B · MRL
Dense 123B coding model — 72.2% SWE-bench Verified
DeepSeek · 671B · MIT
Massive MoE reasoning model — 37B active
DeepSeek · 685B · MIT
State-of-the-art MoE — 37B active params
Moonshot AI · 1T · Kimi
1T-param MoE with 384 experts — 32B active, strong agentic coding
Alibaba · 0.8B · Apache 2.0
Ultra-tiny model for embedded and edge
Meta · 1B · Llama 3.2 Community
Meta's smallest Llama for edge devices
Google · 1B · Gemma
Google's tiny Gemma for on-device
Community · 1.1B · Apache 2.0
Ultralight model for constrained devices
Alibaba · 1.5B · Apache 2.0
Ultra-lightweight coding model
DeepSeek · 1.5B · MIT
Tiny reasoning model distilled from R1
Alibaba · 1.7B · Apache 2.0
Compact multilingual Qwen 3
Alibaba · 2B · Apache 2.0
Small multimodal Qwen 3.5
Google · 2B · Gemma
Google's compact open model
Meta · 3B · Llama 3.2 Community
Lightweight Llama for mobile and edge
HuggingFace · 3B · Apache 2.0
Lightweight multilingual reasoning
Microsoft · 3.8B · MIT
Microsoft's efficient small model with long context
Microsoft · 3.8B · MIT
Lightweight reasoning model
Alibaba · 4B · Apache 2.0
Compact Qwen 3 for general tasks
Google · 4B · Gemma
Multimodal Gemma with 128K context
Alibaba · 4B · Apache 2.0
Small multimodal Qwen 3.5
Mistral AI · 7B · Apache 2.0
High-quality 7B with sliding window attention
Alibaba · 7B · Apache 2.0
Strong multilingual and coding capabilities
Alibaba · 7B · Apache 2.0
Dedicated coding model
DeepSeek · 7B · MIT
R1 reasoning distilled into Qwen 7B
Alibaba · 8B · Apache 2.0
Qwen 3 with thinking mode support
Mistral AI · 8B · MRL
Mistral's efficient 8B model
Google · 9B · Gemma
Google's best mid-size open model
Zhipu AI · 9B · GLM-4
Multilingual model supporting 26 languages with 128K context
NVIDIA · 9B · NVIDIA Open
Hybrid Mamba2 architecture for reasoning
Meta · 11B · Llama 3.2 Community
Multimodal vision and text model
Google · 12B · Gemma
Multimodal Gemma with 128K context
Mistral AI · 12B · Apache 2.0
Multilingual 12B with 128K context
Alibaba · 14B · Apache 2.0
Excellent quality for its size class
Alibaba · 14B · Apache 2.0
Strong all-rounder with thinking mode
DeepSeek · 14B · MIT
R1 reasoning distilled into Qwen 14B
Liquid AI · 24B · Liquid AI
Hybrid MoE with convolution+attention layers — 2.3B active
Mistral AI · 24B · Apache 2.0
Coding-focused model with 256K context — 68% SWE-bench
Google · 27B · Gemma
Google's largest Gemma 2 model
Alibaba · 27.8B · Apache 2.0
Flagship native multimodal Qwen 3.5
Alibaba · 30B · Apache 2.0
MoE with only 3.3B active — extremely efficient
NVIDIA · 30B · NVIDIA Open
MoE with 1M context and 3B active
Alibaba · 32B · Apache 2.0
High-quality reasoning and multilingual
LG AI · 32B · EXAONE AI
Hybrid reasoning, multilingual
Allen AI · 32B · Apache 2.0
Fully open research model by Allen AI
Cohere · 35B · CC BY-NC 4.0
Optimized for retrieval-augmented generation
Alibaba · 35B · Apache 2.0
Efficient multimodal MoE with 3B active
Mistral AI · 47B · Apache 2.0
MoE with 12.9B active params
Alibaba · 72B · Qwen
Alibaba's flagship open model
Alibaba · 122B · Apache 2.0
Large multimodal MoE with 10B active
Mistral AI · 141B · Apache 2.0
Large MoE with 39B active params
Alibaba · 235B · Apache 2.0
Massive MoE with 22B active — frontier quality
Alibaba · 397B · Apache 2.0
Largest multimodal Qwen 3.5 MoE
Meta · 400B · Llama 4 Community
Multimodal MoE with 128 experts — 17B active, 1M context
Meta · 405B · Llama 3.1 Community
Largest open-weight dense model by Meta
Alibaba · 480B · Apache 2.0
Largest open coding MoE — 35B active
DeepSeek · 671B · MIT
Improved V3 with hybrid thinking and tool use
No models found
Try adjusting your search or filters