LMmy choosing

Model Catalog

New & noteworthy local models you can run on your own machine.

Ministral 3
3B
3B
8B
8B
14B
14B
Ministral 3 series, available in three model sizes: 3B, 8B, and 14B parameters. Provides best of class cost-to-performance ratio.
55.1K
22
6
Updated 2 days ago
Qwen3 Next
80B
Hybrid attention architecture, high-sparsity Mixture-of-Experts 80B model (active 3B).
16.1K
14
Updated 3 days ago
Olmo 3
7B
7B
32B
Olmo 3 is a family of Open language models designed to enable the science of language models.
20.6K
16
3
Updated 14 days ago
olmOCR 2
7B
The olmOCR 2 model is a Vision Language Model (VLM) from Allen AI.
23.6K
7
Updated 15 days ago
minimax-m2
230B
MiniMax M2 is a 230B MoE (10B active) model built for coding and agentic workflows
21.6K
16
Updated 29 days ago
gpt-oss-safeguard
20B
120B
gpt-oss-safeguard-20b and gpt-oss-safeguard-120b are open safety models from OpenAI, building on gpt-oss. Trained to help classify text content based on customizable policies.
6.7K
19
2
Updated 1 month ago
Qwen3-VL
2B
4B
8B
30B
32B
Qwen's latest vision-language model. Includes comprehensive upgrades to visual perception, spatial reasoning, and image understanding.
364K
51
5
Updated 1 month ago
Granite 4.0
3B
3B
7B
32B
Granite 4.0 language models are lightweight, state-of-the-art open models that natively support multilingual capabilities, coding tasks, RAG, tool use, and JSON output.
48.5K
34
4
Updated 1 month ago
seed-oss
36B
Advanced reasoning model from ByteDance with flexible "thinking budget" control and ability to reflect on the length of its own reasoning
39K
19
Updated 1 month ago
Qwen3
4B
4B
30B
30B
235B
235B
The latest version of the Qwen3 model family, featuring 4B, 30B, and 235B dense and MoE models, both thinking and non-thinking variants.
317.4K
109
6
Updated 1 month ago
gpt-oss
20B
120B
OpenAI's first open source LLM. Comes in 2 sizes: 20B and 120B. Supports configurable reasoning effort (low, medium, high). Trained for tool use. Apache 2.0 licensed.
1M
226
2
Updated 1 month ago
Qwen3-Coder
30B
480B
State-of-the-art, Mixture-of-Experts local coding model with native support for 256K context length. Available in 30B (3B active) and 480B (35B active) sizes.
186.5K
81
2
Updated 1 month ago
Ernie-4.5
21B
Medium-size Mixture-of-Experts model from Baidu's new Ernie 4.5 line of foundation models.
13.3K
8
Updated 1 month ago
LFM2
350M
700M
1.2B
LFM2 is a new generation of hybrid models developed by Liquid AI, specifically designed for edge AI and on-device deployment. It sets a new standard in terms of quality, speed, and memory efficiency.
51.2K
36
3
Updated 1 month ago
devstral
23.6B
24B
Devstral is a coding model from Mistral AI. It excels at using tools to explore codebases, editing multiple files and power software engineering agents.
71.5K
30
2
Updated 1 month ago
gemma-3n
4.5B
6.9B
Gemma 3n is a generative AI model optimized for use in everyday devices, such as phones, laptops, and tablets.
150.1K
65
2
Updated 1 month ago
Mistral Small
24B
Mistrall Small is a 'knowledge-dense' 24B multi-modal (image input) local model that supports up to 128 token context length.
63.2K
17
Updated 1 month ago
Magistral
23.6B
24B
MistralAI's open-weight reasoning model. 24B dense transformer model supporting up to 128K token context window. The model is capable of long chains of reasoning traces before providing answers.
130.8K
44
2
Updated 1 month ago
mistral-nemo
12B
General purpose dense transformer designed for multilingual use cases. Built in collaboration between MistralAI and NVIDIA.
22.7K
2
Updated 1 month ago
qwen2.5-vl
3B
7B
32B
72B
Qwen2.5-VL is a performant vision-language model, capable of recognizing common objects and text. Supports context length of 128k tokens in a variety of human languages.
64.4K
17
4
Updated 1 month ago
gemma-3
270M
1B
4B
12B
27B
State-of-the-art image + text input models from Google, built from the same research and tech used to create the Gemini models
680.4K
93
5
Updated 1 month ago
phi-4-reasoning
3.8B
14.7B
14.7B
Phi-4-mini-reasoning is a lightweight open model built upon synthetic data with a focus on high-quality, reasoning dense data.
106K
26
3
Updated 1 month ago
phi-4
3B
14B
phi-4 is a state-of-the-art open model built upon a blend of synthetic datasets, data from filtered public domain websites, and acquired academic books and Q&A datasets.
20.1K
7
2
Updated 1 month ago
Codestral
22B
Mistral AI's latest coding model, Codestral can handle both instructions and code completions with ease in over 80 programming languages.
30.2K
17
Updated 1 month ago
Mistral
7B
One of the most popular open-source LLMs, Mistral's 7B Instruct model's balance of speed, size, and performance makes it a great general-purpose daily driver.
70.7K
31
Updated 1 month ago
Qwen3 (1st Generation)
4B
8B
14B
30B
32B
235B
The first batch of Qwen3 models (Qwen3-2504), a collection of dense and MoE models ranging from 4B to 235B. These are general purpose models that score highly on benchmarks.
333K
37
6
Updated 1 month ago
deepseek-r1
7B
8B
8B
14B
32B
70B
Distilled version of the DeepSeek-R1-0528 model, created by continuing the post-training process on the Qwen3 8B Base model using Chain-of-Thought (CoT) from DeepSeek-R1-0528.
432.2K
110
6
Updated 1 month ago