AI & ML interests
None defined yet.
Recent Activity
	View all activity
	
- 
	
	
	  mlx-community/Qwen3-VL-4B-Instruct-3bitImage-Text-to-Text • 0.9B • Updated • 271
- 
	
	
	  mlx-community/Qwen3-VL-4B-Instruct-5bitImage-Text-to-Text • 1B • Updated • 124
- 
	
	
	  mlx-community/Qwen3-VL-4B-Instruct-6bitImage-Text-to-Text • 1B • Updated • 101
- 
	
	
	  mlx-community/Qwen3-VL-4B-Instruct-8bitImage-Text-to-Text • 2B • Updated • 567 • 3
Best in Class MoE, better than Qwen3. Optimised for Smaller devices sub 16 GB (M1/2/3/4) Apple Silicon.
			
	
	💻 Significant Performance: among open models on Agentic Coding, Agentic Browser-Use, and other foundational coding tasks, achieving ~Claude Sonnet.
			
	
	- 
	
	
	  mlx-community/Qwen3-Coder-30B-A3B-Instruct-4bitText Generation • 31B • Updated • 630 • 10
- 
	
	
	  mlx-community/Qwen3-Coder-480B-A35B-Instruct-4bitText Generation • 480B • Updated • 444 • 18
- 
	
	
	  mlx-community/Qwen3-Coder-30B-A3B-Instruct-8bitText Generation • 31B • Updated • 427 • 2
- 
	
	
	  mlx-community/Qwen3-Coder-30B-A3B-Instruct-8bit-DWQ-lr9e8Text Generation • 31B • Updated • 131 • 1
Alibaba's first hybrid model, designed to cut resources and speed things up.
			
	
	- 
	
	
	  mlx-community/Qwen3-Next-80B-A3B-Thinking-8bitText Generation • 80B • Updated • 164 • 2
- 
	
	
	  mlx-community/Qwen3-Next-80B-A3B-Thinking-6bitText Generation • 80B • Updated • 45
- 
	
	
	  mlx-community/Qwen3-Next-80B-A3B-Thinking-5bitText Generation • 80B • Updated • 38
- 
	
	
	  mlx-community/Qwen3-Next-80B-A3B-Thinking-4bitText Generation • 80B • Updated • 210 • 2
Very Small smart model created for the mobile
			
	
	- 
	
	
	  mlx-community/lille-130m-instruct-bf16Text Generation • 0.1B • Updated • 45
- 
	
	
	  mlx-community/lille-130m-instruct-fp16Text Generation • 0.1B • Updated • 43 • 1
- 
	
	
	  mlx-community/lille-130m-instruct-8bitText Generation • 35.8M • Updated • 7
- 
	
	
	  mlx-community/lille-130m-instruct-6bitText Generation • 27.8M • Updated • 5
SwissAI's Apertus models that support 1k languages
			
	
	- 
	
	
	  mlx-community/Apertus-8B-Instruct-2509-bf16Text Generation • 8B • Updated • 345 • 4
- 
	
	
	  mlx-community/Apertus-8B-Instruct-2509-8bitText Generation • 8B • Updated • 117
- 
	
	
	  mlx-community/Apertus-8B-Instruct-2509-6bitText Generation • 8B • Updated • 52
- 
	
	
	  mlx-community/Apertus-8B-Instruct-2509-4bitText Generation • 1B • Updated • 203 • 1
Image Quality Assessment
			
	
	- 
	
	
	  mlx-community/VisualQuality-R1-7B-bf16Reinforcement Learning • 8B • Updated • 19
- 
	
	
	  mlx-community/VisualQuality-R1-7B-8bitReinforcement Learning • Updated • 9
- 
	
	
	  mlx-community/VisualQuality-R1-7B-6bitReinforcement Learning • Updated • 8
- 
	
	
	  mlx-community/VisualQuality-R1-7B-4bitReinforcement Learning • Updated • 10
Apple's text based diffusion model
			
	
	Google's Gemma 3n converted to MLX using mlx-lm
			
	
	- 
	
	
	  mlx-community/gemma-3n-E4B-it-lm-bf16Text Generation • 7B • Updated • 93 • 4
- 
	
	
	  mlx-community/gemma-3n-E2B-it-lm-bf16Text Generation • 4B • Updated • 100
- 
	
	
	  mlx-community/gemma-3n-E4B-it-lm-4bitText Generation • 1B • Updated • 5.42k • 4
- 
	
	
	  mlx-community/gemma-3n-E2B-it-lm-4bitText Generation • 0.7B • Updated • 5.3k • 1
This collection houses Nanonets-OCR-s
			
	
	- 
	
	
	  mlx-community/DeepSeek-R1-0528-4bitText Generation • 105B • Updated • 223 • 17
- 
	
	
	  mlx-community/DeepSeek-R1-0528-Qwen3-8B-4bitText Generation • 1B • Updated • 688 • 4
- 
	
	
	  mlx-community/DeepSeek-R1-0528-Qwen3-8B-4bit-DWQText Generation • 1B • Updated • 151 • 8
- 
	
	
	  mlx-community/DeepSeek-R1-0528-Qwen3-8B-8bitText Generation • 2B • Updated • 44 • 1
- 
	
	
	  mlx-community/Devstral-Small-2505-3bitText Generation • 3B • Updated • 21 • 1
- 
	
	
	  mlx-community/Devstral-Small-2505-4bitText Generation • 4B • Updated • 23 • 2
- 
	
	
	  mlx-community/Devstral-Small-2505-6bitText Generation • Updated • 12 • 1
- 
	
	
	  mlx-community/Devstral-Small-2505-8bitText Generation • Updated • 22 • 1
- 
	
	
	  mlx-community/Llama-OuteTTS-1.0-1B-fp16Text-to-Speech • 1B • Updated • 28 • 3
- 
	
	
	  mlx-community/Llama-OuteTTS-1.0-1B-4bitText-to-Speech • 0.2B • Updated • 116 • 1
- 
	
	
	  mlx-community/Llama-OuteTTS-1.0-1B-8bitText-to-Speech • 0.4B • Updated • 15 • 1
- 
	
	
	  mlx-community/Llama-OuteTTS-1.0-1B-6bitText-to-Speech • 0.3B • Updated • 6
Gemma 3 distilled weight quantized (DWQ) models
			
	
	- 
	
	
	  mlx-community/gemma-3-4b-it-4bit-DWQText Generation • 0.7B • Updated • 119 • 1
- 
	
	
	  mlx-community/gemma-3-12b-it-4bit-DWQText Generation • 2B • Updated • 110 • 2
- 
	
	
	  mlx-community/gemma-3-1b-it-4bit-DWQText Generation • 0.2B • Updated • 65
- 
	
	
	  mlx-community/gemma-3-27b-it-4bit-DWQText Generation • 4B • Updated • 116 • 3
Abliterated, and further fine-tuned to be the most uncensored models available. Now in MLX
			
	
	- 
	
	
	  mlx-community/Josiefied-Qwen3-30B-A3B-abliterated-v2-bf16Text Generation • 31B • Updated • 28
- 
	
	
	  mlx-community/Josiefied-Qwen3-30B-A3B-abliterated-v2-8bitText Generation • 31B • Updated • 115
- 
	
	
	  mlx-community/Josiefied-Qwen3-30B-A3B-abliterated-v2-6bitText Generation • 31B • Updated • 34
- 
	
	
	  mlx-community/Josiefied-Qwen3-30B-A3B-abliterated-v2-4bitText Generation • 31B • Updated • 160 • 2
Quantization Aware Trained (QAT) Gemma 3 checkpoints. The model preserves similar quality as half precision while using 3x less memory.
			
	
	- 
	
	
	  mlx-community/gemma-3-27b-it-qat-bf16Image-Text-to-Text • Updated • 123 • 5
- 
	
	
	  mlx-community/gemma-3-27b-it-qat-8bitImage-Text-to-Text • Updated • 171 • 9
- 
	
	
	  mlx-community/gemma-3-27b-it-qat-6bitImage-Text-to-Text • Updated • 25
- 
	
	
	  mlx-community/gemma-3-27b-it-qat-4bitImage-Text-to-Text • Updated • 83.6k • 20
- 
	
	
	  mlx-community/answerdotai-ModernBERT-base-8bitFill-Mask • 53M • Updated • 3
- 
	
	
	  mlx-community/answerdotai-ModernBERT-base-4bitFill-Mask • 29.5M • Updated • 10
- 
	
	
	  mlx-community/answerdotai-ModernBERT-base-bf16Fill-Mask • 0.2B • Updated • 18 • 1
- 
	
	
	  mlx-community/answerdotai-ModernBERT-Large-Instruct-4bitFill-Mask • 70M • Updated • 4
Kokoro is an open-weight TTS model with 82 million parameters. Despite its lightweight architecture, it delivers amazing quality.
			
	
	- 
	
	
	  mlx-community/Qwen2.5-7B-Instruct-1M-4bitText Generation • 1B • Updated • 131 • 10
- 
	
	
	  mlx-community/Qwen2.5-7B-Instruct-1M-6bitText Generation • 2B • Updated • 1 • 2
- 
	
	
	  mlx-community/Qwen2.5-7B-Instruct-1M-3bitText Generation • 1.0B • Updated • 2
- 
	
	
	  mlx-community/Qwen2.5-7B-Instruct-1M-8bitText Generation • 2B • Updated • 13 • 3
Convert HTML content to LLM-friendly Markdown/JSON content
			
	
	- 
	
	
	  mlx-community/QVQ-72B-Preview-4bitImage-Text-to-Text • 11B • Updated • 8 • 7
- 
	
	
	  mlx-community/QVQ-72B-Preview-6bitImage-Text-to-Text • 16B • Updated • 2 • 2
- 
	
	
	  mlx-community/QVQ-72B-Preview-3bitImage-Text-to-Text • 9B • Updated • 2 • 5
- 
	
	
	  mlx-community/QVQ-72B-Preview-8bitImage-Text-to-Text • 21B • Updated • 1 • 3
The best uncensored models
			
	
	- 
	
	
	  mlx-community/Josiefied-Qwen2.5-Coder-7B-Instruct-abliterated-v1Text Generation • 8B • Updated • 12
- 
	
	
	  mlx-community/Josiefied-Qwen2.5-Coder-7B-Instruct-abliterated-v1-8bitText Generation • 2B • Updated • 11
- 
	
	
	  mlx-community/Josiefied-Qwen2.5-Coder-7B-Instruct-abliterated-v1-6bitText Generation • 2B • Updated • 8
- 
	
	
	  mlx-community/Josiefied-Qwen2.5-Coder-7B-Instruct-abliterated-v1-4bitText Generation • 1B • Updated • 31 • 1
EXAONE 3.5, a collection of instruction-tuned bilingual generative models ranging from 2.4B to 32B parameters, developed by LG AI.
			
	
	- 
	
	
	  mlx-community/paligemma2-3b-ft-docci-448-8bitImage-Text-to-Text • 0.9B • Updated
- 
	
	
	  mlx-community/paligemma2-3b-ft-docci-448-6bitImage-Text-to-Text • 0.7B • Updated
- 
	
	
	  mlx-community/paligemma2-3b-ft-docci-448-bf16Image-Text-to-Text • 3B • Updated • 8 • 1
- 
	
	
	  mlx-community/paligemma2-10b-ft-docci-448-bf16Image-Text-to-Text • 10B • Updated • 25 • 3
- 
	
	
	  mlx-community/SmolVLM-Instruct-4bitImage-Text-to-Text • 0.5B • Updated • 228 • 5
- 
	
	
	  mlx-community/SmolVLM-Instruct-6bitImage-Text-to-Text • 0.6B • Updated • 8
- 
	
	
	  mlx-community/SmolVLM-Instruct-8bitImage-Text-to-Text • 0.7B • Updated • 23 • 9
- 
	
	
	  mlx-community/SmolVLM-Instruct-bf16Image-Text-to-Text • 2B • Updated • 13 • 5
- 
	
	
	  mlx-community/Florence-2-base-ft-4bitImage-Text-to-Text • 48.8M • Updated • 64 • 1
- 
	
	
	  mlx-community/Florence-2-large-ft-bf16Image-Text-to-Text • 0.8B • Updated • 52 • 1
- 
	
	
	  mlx-community/Florence-2-base-ft-bf16Image-Text-to-Text • 0.3B • Updated • 10 • 1
- 
	
	
	  mlx-community/Florence-2-base-ft-8bitImage-Text-to-Text • 81.7M • Updated • 29 • 1
Code-specific model series based on Qwen2.5
			
	
	- 
	
	
	  mlx-community/Qwen2.5-Coder-32B-Instruct-8bitText Generation • 9B • Updated • 90 • 11
- 
	
	
	  mlx-community/Qwen2.5-Coder-14B-Instruct-4bitText Generation • 2B • Updated • 165 • 4
- 
	
	
	  mlx-community/Qwen2.5-Coder-14B-Instruct-bf16Text Generation • 15B • Updated • 15 • 2
- 
	
	
	  mlx-community/Qwen2.5-Coder-3B-Instruct-8bitText Generation • 0.9B • Updated • 8
A collection of Neversleep's RP focused Lumimaid LLMs.
			
	
	Qwen1.5 is the improved version of Qwen, the large language model series developed by Alibaba Cloud.
			
	
	- 
	
	
	  mlx-community/Qwen1.5-1.8B-Chat-4bitText Generation • 0.5B • Updated • 9 • 2
- 
	
	
	  mlx-community/Qwen1.5-0.5B-Chat-4bitText Generation • 72.6M • Updated • 3.47k • 4
- 
	
	
	  mlx-community/Qwen1.5-14B-Chat-4bitText Generation • 3B • Updated • 7 • 1
- 
	
	
	  mlx-community/Qwen1.5-7B-Chat-4bitText Generation • 2B • Updated • 7 • 2
- 
	
	
	  mlx-community/Meta-Llama-3-8B-Instruct-4bitText Generation • 2B • Updated • 1.68k • 79
- 
	
	
	  mlx-community/Meta-Llama-3-8B-4bitText Generation • 2B • Updated • 42 • 8
- 
	
	
	  mlx-community/Meta-Llama-Guard-2-8B-4bitText Generation • 2B • Updated • 5
- 
	
	
	  mlx-community/Meta-Llama-3-70B-4bitText Generation • 11B • Updated • 58 • 9
- 
	
	
	  mlx-community/Phi-3-mini-4k-instruct-4bitText Generation • 0.6B • Updated • 669 • 12
- 
	
	
	  mlx-community/Phi-3-mini-128k-instruct-4bitText Generation • 0.6B • Updated • 179 • 12
- 
	
	
	  mlx-community/Phi-3-mini-128k-instruct-8bitText Generation • 1B • Updated • 38 • 10
- 
	
	
	  mlx-community/Phi-3-mini-4k-instruct-8bitText Generation • 1B • Updated • 19 • 2
A family of Open-source Efficient Language Models from Apple.
			
	
	Mamba is a new LLM architecture that integrates the Structured State Space sequence model to manage lengthy data sequences.
			
	
	EnCodec models in MLX
			
	
	This collection houses Nanonets-OCR2 models
			
	
	- 
	
	
	  mlx-community/Nanonets-OCR2-3B-bf16Image-Text-to-Text • 4B • Updated • 302
- 
	
	
	  mlx-community/Nanonets-OCR2-3B-8bitImage-Text-to-Text • 2B • Updated • 271
- 
	
	
	  mlx-community/Nanonets-OCR2-3B-6bitImage-Text-to-Text • 1B • Updated • 115
- 
	
	
	  mlx-community/Nanonets-OCR2-3B-4bitImage-Text-to-Text • 1B • Updated • 328
Apriel-1.5-15b-Thinker is a multimodal reasoning model in ServiceNow’s Apriel SLM series which achieves competitive performance against models 10 time
			
	
	- 
	
	
	  mlx-community/Apriel-1.5-15b-Thinker-4bitText Generation • Updated • 687 • 2
- 
	
	
	  mlx-community/Apriel-1.5-15b-Thinker-5bitText Generation • Updated • 119
- 
	
	
	  mlx-community/Apriel-1.5-15b-Thinker-6bit-MLXImage-Text-to-Text • Updated • 194 • 1
- 
	
	
	  mlx-community/Apriel-1.5-15b-Thinker-3bit-MLXImage-Text-to-Text • Updated • 80
- 
	
	
	  mlx-community/Granite-4.0-H-Tiny-4bit-DWQText Generation • 1B • Updated • 902 • 2
- 
	
	
	  mlx-community/granite-4.0-h-micro-8bitText Generation • 0.9B • Updated • 498 • 1
- 
	
	
	  mlx-community/granite-4.0-h-small-4bitText Generation • 32B • Updated • 325
- 
	
	
	  mlx-community/granite-4.0-tiny-preview-4bitText Generation • 1B • Updated • 61
SEA-LION mlx models by AI Singapore.
			
	
	- 
	
	
	  mlx-community/Gemma-SEA-LION-v4-27B-IT-mlx-4bitText Generation • 27B • Updated • 5 • 1
- 
	
	
	  mlx-community/Llama-SEA-LION-v3.5-8B-R-mlx-4bitText Generation • 2B • Updated • 2
- 
	
	
	  mlx-community/Gemma-SEA-LION-v3-9B-IT-mlx-4bitText Generation • 9B • Updated • 4
- 
	
	
	  mlx-community/Llama-SEA-LION-v3-8B-IT-mlx-4bitText Generation • 2B • Updated • 1
- 
	
	
	  mlx-community/embeddinggemma-300m-4bitSentence Similarity • 48.1M • Updated • 166 • 2
- 
	
	
	  mlx-community/embeddinggemma-300m-5bitSentence Similarity • 57.7M • Updated • 43
- 
	
	
	  mlx-community/embeddinggemma-300m-6bitSentence Similarity • 67.4M • Updated • 56
- 
	
	
	  mlx-community/embeddinggemma-300m-8bitSentence Similarity • 86.6M • Updated • 509 • 2
A fine-tuned Gemma 3 1B instruction model specialized for English-to-Swahili translation and Swahili conversational AI. The model accepts input in bot
			
	
	- 
	
	
	  mlx-community/gemma-3-270m-it-4bitText Generation • 41.9M • Updated • 268 • 8
- 
	
	
	  mlx-community/gemma-3-270m-it-5bitText Generation • 81.8M • Updated • 15
- 
	
	
	  mlx-community/gemma-3-270m-it-6bitText Generation • 95.4M • Updated • 13
- 
	
	
	  mlx-community/gemma-3-270m-it-8bitText Generation • 0.1B • Updated • 1.07k • 2
- 
	
	
	  mlx-community/ERNIE-4.5-300B-A47B-PT-4bitText Generation • 299B • Updated • 20 • 2
- 
	
	
	  mlx-community/ERNIE-4.5-21B-A3B-PT-bf16Text Generation • 22B • Updated • 20 • 1
- 
	
	
	  mlx-community/ERNIE-4.5-21B-A3B-PT-8bitText Generation • 22B • Updated • 20 • 2
- 
	
	
	  mlx-community/ERNIE-4.5-21B-A3B-PT-6bitText Generation • 22B • Updated • 5
Series of code models by JetBrains
			
	
	This collection houses  BitNet-1.58, Falcon3-1.58 and Falcon-E quants.
			
	
	- 
	
	
	  mlx-community/bitnet-b1.58-2B-4TText Generation • 0.8B • Updated • 51 • 1
- 
	
	
	  mlx-community/bitnet-b1.58-2B-4T-4bitText Generation • 0.6B • Updated • 119
- 
	
	
	  mlx-community/bitnet-b1.58-2B-4T-8bitText Generation • 0.6B • Updated • 115
- 
	
	
	  mlx-community/bitnet-b1.58-2B-4T-6bitText Generation • 0.6B • Updated • 10
High-quality 4-bit quants of the Qwen3 model family.
			
	
	- 
	
	
	  mlx-community/Qwen3-14B-4bit-DWQ-053125Text Generation • 2B • Updated • 142 • 4
- 
	
	
	  mlx-community/Qwen3-8B-4bit-DWQ-053125Text Generation • 1B • Updated • 156 • 1
- 
	
	
	  mlx-community/Qwen3-4B-4bit-DWQ-053125Text Generation • 0.6B • Updated • 112 • 2
- 
	
	
	  mlx-community/Qwen3-1.7B-4bit-DWQ-053125Text Generation • 0.3B • Updated • 126 • 2
- 
	
	
	  mlx-community/AceReason-Nemotron-7B-4bitText Generation • 1B • Updated • 61
- 
	
	
	  mlx-community/AceReason-Nemotron-7B-8bitText Generation • 2B • Updated
- 
	
	
	  mlx-community/AceReason-Nemotron-7B-bf16Text Generation • 8B • Updated • 3
- 
	
	
	  mlx-community/AceReason-Nemotron-14B-4bitText Generation • 2B • Updated • 6
Collection of Gemma 3 variants for performance on medical text and image comprehension to accelerate building healthcare-based AI applications.
			
	
	- 
	
	
	  mlx-community/medgemma-4b-it-4bitImage-Text-to-Text • 0.9B • Updated • 55 • 2
- 
	
	
	  mlx-community/medgemma-4b-it-6bitImage-Text-to-Text • 1B • Updated • 24 • 1
- 
	
	
	  mlx-community/medgemma-4b-it-8bitImage-Text-to-Text • 1B • Updated • 62 • 1
- 
	
	
	  mlx-community/medgemma-4b-it-bf16Image-Text-to-Text • 5B • Updated • 47 • 1
Nvidia's ASR models, now in MLX!
			
	
	- 
	
	
	  mlx-community/parakeet-ctc-0.6bAutomatic Speech Recognition • 0.6B • Updated • 457 • 2
- 
	
	
	  mlx-community/parakeet-rnnt-0.6bAutomatic Speech Recognition • 0.6B • Updated • 1.09k
- 
	
	
	  mlx-community/parakeet-ctc-1.1bAutomatic Speech Recognition • 1B • Updated • 15 • 1
- 
	
	
	  mlx-community/parakeet-rnnt-1.1bAutomatic Speech Recognition • 1B • Updated • 25 • 1
The GLM-4 and Z1 series are powerful open-source language models excelling in reasoning, code, and complex tasks.
			
	
	- 
	
	
	  mlx-community/GLM-Z1-32B-0414-4bitText Generation • 5B • Updated • 142 • 2
- 
	
	
	  mlx-community/GLM-4-32B-0414-4bitText Generation • 5B • Updated • 255 • 5
- 
	
	
	  mlx-community/GLM-4-32B-Base-0414-8bitText Generation • 9B • Updated • 22
- 
	
	
	  mlx-community/GLM-4-32B-Base-0414-6bitText Generation • 7B • Updated • 26
- 
	
	
	  mlx-community/Llama-4-Scout-17B-16E-Instruct-4bitImage-Text-to-Text • Updated • 369 • 9
- 
	
	
	  mlx-community/Llama-4-Scout-17B-16E-Instruct-6bitImage-Text-to-Text • Updated • 210 • 5
- 
	
	
	  mlx-community/Llama-4-Scout-17B-16E-Instruct-8bitImage-Text-to-Text • Updated • 222 • 3
- 
	
	
	  mlx-community/Llama-4-Maverick-17B-16E-Instruct-4bitText Generation • 63B • Updated • 283 • 7
A collection of lightweight, state-of-the-art open models built from the same research and technology that powers the Gemini 2.0 models
			
	
	- 
	
	
	  mlx-community/gemma-3-4b-it-8bitImage-Text-to-Text • 2B • Updated • 625 • 5
- 
	
	
	  mlx-community/gemma-3-4b-pt-4bitImage-Text-to-Text • 1B • Updated • 45 • 3
- 
	
	
	  mlx-community/gemma-3-4b-it-bf16Image-Text-to-Text • 5B • Updated • 157 • 1
- 
	
	
	  mlx-community/gemma-3-4b-pt-6bitImage-Text-to-Text • 1B • Updated • 9
- 
	
	
	  mlx-community/OLMoE-1B-7B-0125-InstructText Generation • 7B • Updated • 4
- 
	
	
	  mlx-community/OLMoE-1B-7B-0125-Instruct-8bitText Generation • 2B • Updated • 5
- 
	
	
	  mlx-community/OLMoE-1B-7B-0125-Instruct-6bitText Generation • 2B • Updated
- 
	
	
	  mlx-community/OLMoE-1B-7B-0125-Instruct-4bitText Generation • 1B • Updated • 28 • 2
FuseAI is attempting to merge CoT models to achieve newer models that are more than the sum of their parts.
			
	
	- 
	
	
	  mlx-community/Qwen2.5-VL-72B-Instruct-8bitImage-Text-to-Text • 21B • Updated • 59 • 2
- 
	
	
	  mlx-community/Qwen2.5-VL-72B-Instruct-6bitImage-Text-to-Text • 16B • Updated • 21 • 1
- 
	
	
	  mlx-community/Qwen2.5-VL-72B-Instruct-4bitImage-Text-to-Text • 12B • Updated • 228 • 7
- 
	
	
	  mlx-community/Qwen2.5-VL-72B-Instruct-3bitImage-Text-to-Text • 10B • Updated • 38 • 5
Kyutai's Helium-1 2B Model, outperforming other state of the art small models.
			
	
	- 
	
	
	  mlx-community/helium-1-preview-2b-float32Text Generation • 2B • Updated • 2
- 
	
	
	  mlx-community/helium-1-preview-2bText Generation • 2B • Updated
- 
	
	
	  mlx-community/helium-1-preview-2b-8bitText Generation • 0.6B • Updated • 8 • 1
- 
	
	
	  mlx-community/helium-1-preview-2b-4bitText Generation • 0.3B • Updated • 2 • 1
- 
	
	
	  mlx-community/deepseek-vl2-6bitImage-Text-to-Text • 6B • Updated • 45 • 1
- 
	
	
	  mlx-community/deepseek-vl2-small-4bitImage-Text-to-Text • 3B • Updated • 41
- 
	
	
	  mlx-community/deepseek-vl2-4bitImage-Text-to-Text • 4B • Updated • 69 • 1
- 
	
	
	  mlx-community/deepseek-vl2-small-6bitImage-Text-to-Text • 4B • Updated • 20
- 
	
	
	  mlx-community/Llama-3.3-70B-Instruct-8bitText Generation • 20B • Updated • 373 • 14
- 
	
	
	  mlx-community/Llama-3.3-70B-Instruct-6bitText Generation • 15B • Updated • 69 • 5
- 
	
	
	  mlx-community/Llama-3.3-70B-Instruct-3bitText Generation • 9B • Updated • 106 • 7
- 
	
	
	  mlx-community/Llama-3.3-70B-Instruct-4bitText Generation • 11B • Updated • 1.07k • 30
Falcon Mamba models compatible with MLX
			
	
	Google’s Code-Gemma
			
	
	The Qwen 2.5 models are a series of AI models trained on 18 trillion tokens, supporting 29 languages and offering advanced features such as instructio
			
	
	- 
	
	
	  mlx-community/Qwen2.5-72B-Instruct-bf16Text Generation • 73B • Updated • 8
- 
	
	
	  mlx-community/Qwen2.5-72B-Instruct-8bitText Generation • 20B • Updated • 21 • 3
- 
	
	
	  mlx-community/Qwen2.5-72B-Instruct-4bitText Generation • 11B • Updated • 101 • 5
- 
	
	
	  mlx-community/Qwen2.5-32B-Instruct-bf16Text Generation • 33B • Updated • 12
OpenAI Whisper speech recognition models in MLX format
			
	
	A series of smol LLMs: 135M, 360M and 1.7B.
			
	
	- 
	
	
	  mlx-community/Meta-Llama-3.1-70B-bf16Text Generation • 71B • Updated • 18 • 4
- 
	
	
	  mlx-community/Meta-Llama-3.1-70B-Instruct-bf16Text Generation • 71B • Updated • 9 • 2
- 
	
	
	  mlx-community/Meta-Llama-3.1-8B-Instruct-bf16Text Generation • 8B • Updated • 235 • 3
- 
	
	
	  mlx-community/Meta-Llama-3.1-8B-Instruct-8bitText Generation • 2B • Updated • 318 • 10
Meta goes small with Llama3.2, both text only 1B and 3B, and the 11B Vision models.
			
	
	- 
	
	
	  mlx-community/Llama-3.2-11B-Vision-Instruct-abliteratedImage-Text-to-Text • 11B • Updated • 667 • 7
- 
	
	
	  mlx-community/Llama-3.2-11B-Vision-Instruct-abliterated-8-bitImage-Text-to-Text • 3B • Updated • 74
- 
	
	
	  mlx-community/Llama-3.2-11B-Vision-Instruct-abliterated-4-bitImage-Text-to-Text • 2B • Updated • 91 • 1
- 
	
	
	  mlx-community/Llama-3.2-11B-Vision-Instruct-8bitImage-to-Text • 3B • Updated • 581 • 10
- 
	
	
	  mlx-community/Qwen3-VL-4B-Instruct-3bitImage-Text-to-Text • 0.9B • Updated • 271
- 
	
	
	  mlx-community/Qwen3-VL-4B-Instruct-5bitImage-Text-to-Text • 1B • Updated • 124
- 
	
	
	  mlx-community/Qwen3-VL-4B-Instruct-6bitImage-Text-to-Text • 1B • Updated • 101
- 
	
	
	  mlx-community/Qwen3-VL-4B-Instruct-8bitImage-Text-to-Text • 2B • Updated • 567 • 3
This collection houses Nanonets-OCR2 models
			
	
	- 
	
	
	  mlx-community/Nanonets-OCR2-3B-bf16Image-Text-to-Text • 4B • Updated • 302
- 
	
	
	  mlx-community/Nanonets-OCR2-3B-8bitImage-Text-to-Text • 2B • Updated • 271
- 
	
	
	  mlx-community/Nanonets-OCR2-3B-6bitImage-Text-to-Text • 1B • Updated • 115
- 
	
	
	  mlx-community/Nanonets-OCR2-3B-4bitImage-Text-to-Text • 1B • Updated • 328
Best in Class MoE, better than Qwen3. Optimised for Smaller devices sub 16 GB (M1/2/3/4) Apple Silicon.
			
	
	Apriel-1.5-15b-Thinker is a multimodal reasoning model in ServiceNow’s Apriel SLM series which achieves competitive performance against models 10 time
			
	
	- 
	
	
	  mlx-community/Apriel-1.5-15b-Thinker-4bitText Generation • Updated • 687 • 2
- 
	
	
	  mlx-community/Apriel-1.5-15b-Thinker-5bitText Generation • Updated • 119
- 
	
	
	  mlx-community/Apriel-1.5-15b-Thinker-6bit-MLXImage-Text-to-Text • Updated • 194 • 1
- 
	
	
	  mlx-community/Apriel-1.5-15b-Thinker-3bit-MLXImage-Text-to-Text • Updated • 80
💻 Significant Performance: among open models on Agentic Coding, Agentic Browser-Use, and other foundational coding tasks, achieving ~Claude Sonnet.
			
	
	- 
	
	
	  mlx-community/Qwen3-Coder-30B-A3B-Instruct-4bitText Generation • 31B • Updated • 630 • 10
- 
	
	
	  mlx-community/Qwen3-Coder-480B-A35B-Instruct-4bitText Generation • 480B • Updated • 444 • 18
- 
	
	
	  mlx-community/Qwen3-Coder-30B-A3B-Instruct-8bitText Generation • 31B • Updated • 427 • 2
- 
	
	
	  mlx-community/Qwen3-Coder-30B-A3B-Instruct-8bit-DWQ-lr9e8Text Generation • 31B • Updated • 131 • 1
- 
	
	
	  mlx-community/Granite-4.0-H-Tiny-4bit-DWQText Generation • 1B • Updated • 902 • 2
- 
	
	
	  mlx-community/granite-4.0-h-micro-8bitText Generation • 0.9B • Updated • 498 • 1
- 
	
	
	  mlx-community/granite-4.0-h-small-4bitText Generation • 32B • Updated • 325
- 
	
	
	  mlx-community/granite-4.0-tiny-preview-4bitText Generation • 1B • Updated • 61
Alibaba's first hybrid model, designed to cut resources and speed things up.
			
	
	- 
	
	
	  mlx-community/Qwen3-Next-80B-A3B-Thinking-8bitText Generation • 80B • Updated • 164 • 2
- 
	
	
	  mlx-community/Qwen3-Next-80B-A3B-Thinking-6bitText Generation • 80B • Updated • 45
- 
	
	
	  mlx-community/Qwen3-Next-80B-A3B-Thinking-5bitText Generation • 80B • Updated • 38
- 
	
	
	  mlx-community/Qwen3-Next-80B-A3B-Thinking-4bitText Generation • 80B • Updated • 210 • 2
SEA-LION mlx models by AI Singapore.
			
	
	- 
	
	
	  mlx-community/Gemma-SEA-LION-v4-27B-IT-mlx-4bitText Generation • 27B • Updated • 5 • 1
- 
	
	
	  mlx-community/Llama-SEA-LION-v3.5-8B-R-mlx-4bitText Generation • 2B • Updated • 2
- 
	
	
	  mlx-community/Gemma-SEA-LION-v3-9B-IT-mlx-4bitText Generation • 9B • Updated • 4
- 
	
	
	  mlx-community/Llama-SEA-LION-v3-8B-IT-mlx-4bitText Generation • 2B • Updated • 1
Very Small smart model created for the mobile
			
	
	- 
	
	
	  mlx-community/lille-130m-instruct-bf16Text Generation • 0.1B • Updated • 45
- 
	
	
	  mlx-community/lille-130m-instruct-fp16Text Generation • 0.1B • Updated • 43 • 1
- 
	
	
	  mlx-community/lille-130m-instruct-8bitText Generation • 35.8M • Updated • 7
- 
	
	
	  mlx-community/lille-130m-instruct-6bitText Generation • 27.8M • Updated • 5
- 
	
	
	  mlx-community/embeddinggemma-300m-4bitSentence Similarity • 48.1M • Updated • 166 • 2
- 
	
	
	  mlx-community/embeddinggemma-300m-5bitSentence Similarity • 57.7M • Updated • 43
- 
	
	
	  mlx-community/embeddinggemma-300m-6bitSentence Similarity • 67.4M • Updated • 56
- 
	
	
	  mlx-community/embeddinggemma-300m-8bitSentence Similarity • 86.6M • Updated • 509 • 2
SwissAI's Apertus models that support 1k languages
			
	
	- 
	
	
	  mlx-community/Apertus-8B-Instruct-2509-bf16Text Generation • 8B • Updated • 345 • 4
- 
	
	
	  mlx-community/Apertus-8B-Instruct-2509-8bitText Generation • 8B • Updated • 117
- 
	
	
	  mlx-community/Apertus-8B-Instruct-2509-6bitText Generation • 8B • Updated • 52
- 
	
	
	  mlx-community/Apertus-8B-Instruct-2509-4bitText Generation • 1B • Updated • 203 • 1
A fine-tuned Gemma 3 1B instruction model specialized for English-to-Swahili translation and Swahili conversational AI. The model accepts input in bot
			
	
	- 
	
	
	  mlx-community/gemma-3-270m-it-4bitText Generation • 41.9M • Updated • 268 • 8
- 
	
	
	  mlx-community/gemma-3-270m-it-5bitText Generation • 81.8M • Updated • 15
- 
	
	
	  mlx-community/gemma-3-270m-it-6bitText Generation • 95.4M • Updated • 13
- 
	
	
	  mlx-community/gemma-3-270m-it-8bitText Generation • 0.1B • Updated • 1.07k • 2
Image Quality Assessment
			
	
	- 
	
	
	  mlx-community/VisualQuality-R1-7B-bf16Reinforcement Learning • 8B • Updated • 19
- 
	
	
	  mlx-community/VisualQuality-R1-7B-8bitReinforcement Learning • Updated • 9
- 
	
	
	  mlx-community/VisualQuality-R1-7B-6bitReinforcement Learning • Updated • 8
- 
	
	
	  mlx-community/VisualQuality-R1-7B-4bitReinforcement Learning • Updated • 10
- 
	
	
	  mlx-community/ERNIE-4.5-300B-A47B-PT-4bitText Generation • 299B • Updated • 20 • 2
- 
	
	
	  mlx-community/ERNIE-4.5-21B-A3B-PT-bf16Text Generation • 22B • Updated • 20 • 1
- 
	
	
	  mlx-community/ERNIE-4.5-21B-A3B-PT-8bitText Generation • 22B • Updated • 20 • 2
- 
	
	
	  mlx-community/ERNIE-4.5-21B-A3B-PT-6bitText Generation • 22B • Updated • 5
Apple's text based diffusion model
			
	
	Series of code models by JetBrains
			
	
	Google's Gemma 3n converted to MLX using mlx-lm
			
	
	- 
	
	
	  mlx-community/gemma-3n-E4B-it-lm-bf16Text Generation • 7B • Updated • 93 • 4
- 
	
	
	  mlx-community/gemma-3n-E2B-it-lm-bf16Text Generation • 4B • Updated • 100
- 
	
	
	  mlx-community/gemma-3n-E4B-it-lm-4bitText Generation • 1B • Updated • 5.42k • 4
- 
	
	
	  mlx-community/gemma-3n-E2B-it-lm-4bitText Generation • 0.7B • Updated • 5.3k • 1
This collection houses Nanonets-OCR-s
			
	
	This collection houses  BitNet-1.58, Falcon3-1.58 and Falcon-E quants.
			
	
	- 
	
	
	  mlx-community/bitnet-b1.58-2B-4TText Generation • 0.8B • Updated • 51 • 1
- 
	
	
	  mlx-community/bitnet-b1.58-2B-4T-4bitText Generation • 0.6B • Updated • 119
- 
	
	
	  mlx-community/bitnet-b1.58-2B-4T-8bitText Generation • 0.6B • Updated • 115
- 
	
	
	  mlx-community/bitnet-b1.58-2B-4T-6bitText Generation • 0.6B • Updated • 10
High-quality 4-bit quants of the Qwen3 model family.
			
	
	- 
	
	
	  mlx-community/Qwen3-14B-4bit-DWQ-053125Text Generation • 2B • Updated • 142 • 4
- 
	
	
	  mlx-community/Qwen3-8B-4bit-DWQ-053125Text Generation • 1B • Updated • 156 • 1
- 
	
	
	  mlx-community/Qwen3-4B-4bit-DWQ-053125Text Generation • 0.6B • Updated • 112 • 2
- 
	
	
	  mlx-community/Qwen3-1.7B-4bit-DWQ-053125Text Generation • 0.3B • Updated • 126 • 2
- 
	
	
	  mlx-community/DeepSeek-R1-0528-4bitText Generation • 105B • Updated • 223 • 17
- 
	
	
	  mlx-community/DeepSeek-R1-0528-Qwen3-8B-4bitText Generation • 1B • Updated • 688 • 4
- 
	
	
	  mlx-community/DeepSeek-R1-0528-Qwen3-8B-4bit-DWQText Generation • 1B • Updated • 151 • 8
- 
	
	
	  mlx-community/DeepSeek-R1-0528-Qwen3-8B-8bitText Generation • 2B • Updated • 44 • 1
- 
	
	
	  mlx-community/AceReason-Nemotron-7B-4bitText Generation • 1B • Updated • 61
- 
	
	
	  mlx-community/AceReason-Nemotron-7B-8bitText Generation • 2B • Updated
- 
	
	
	  mlx-community/AceReason-Nemotron-7B-bf16Text Generation • 8B • Updated • 3
- 
	
	
	  mlx-community/AceReason-Nemotron-14B-4bitText Generation • 2B • Updated • 6
- 
	
	
	  mlx-community/Devstral-Small-2505-3bitText Generation • 3B • Updated • 21 • 1
- 
	
	
	  mlx-community/Devstral-Small-2505-4bitText Generation • 4B • Updated • 23 • 2
- 
	
	
	  mlx-community/Devstral-Small-2505-6bitText Generation • Updated • 12 • 1
- 
	
	
	  mlx-community/Devstral-Small-2505-8bitText Generation • Updated • 22 • 1
Collection of Gemma 3 variants for performance on medical text and image comprehension to accelerate building healthcare-based AI applications.
			
	
	- 
	
	
	  mlx-community/medgemma-4b-it-4bitImage-Text-to-Text • 0.9B • Updated • 55 • 2
- 
	
	
	  mlx-community/medgemma-4b-it-6bitImage-Text-to-Text • 1B • Updated • 24 • 1
- 
	
	
	  mlx-community/medgemma-4b-it-8bitImage-Text-to-Text • 1B • Updated • 62 • 1
- 
	
	
	  mlx-community/medgemma-4b-it-bf16Image-Text-to-Text • 5B • Updated • 47 • 1
- 
	
	
	  mlx-community/Llama-OuteTTS-1.0-1B-fp16Text-to-Speech • 1B • Updated • 28 • 3
- 
	
	
	  mlx-community/Llama-OuteTTS-1.0-1B-4bitText-to-Speech • 0.2B • Updated • 116 • 1
- 
	
	
	  mlx-community/Llama-OuteTTS-1.0-1B-8bitText-to-Speech • 0.4B • Updated • 15 • 1
- 
	
	
	  mlx-community/Llama-OuteTTS-1.0-1B-6bitText-to-Speech • 0.3B • Updated • 6
Gemma 3 distilled weight quantized (DWQ) models
			
	
	- 
	
	
	  mlx-community/gemma-3-4b-it-4bit-DWQText Generation • 0.7B • Updated • 119 • 1
- 
	
	
	  mlx-community/gemma-3-12b-it-4bit-DWQText Generation • 2B • Updated • 110 • 2
- 
	
	
	  mlx-community/gemma-3-1b-it-4bit-DWQText Generation • 0.2B • Updated • 65
- 
	
	
	  mlx-community/gemma-3-27b-it-4bit-DWQText Generation • 4B • Updated • 116 • 3
Nvidia's ASR models, now in MLX!
			
	
	- 
	
	
	  mlx-community/parakeet-ctc-0.6bAutomatic Speech Recognition • 0.6B • Updated • 457 • 2
- 
	
	
	  mlx-community/parakeet-rnnt-0.6bAutomatic Speech Recognition • 0.6B • Updated • 1.09k
- 
	
	
	  mlx-community/parakeet-ctc-1.1bAutomatic Speech Recognition • 1B • Updated • 15 • 1
- 
	
	
	  mlx-community/parakeet-rnnt-1.1bAutomatic Speech Recognition • 1B • Updated • 25 • 1
Abliterated, and further fine-tuned to be the most uncensored models available. Now in MLX
			
	
	- 
	
	
	  mlx-community/Josiefied-Qwen3-30B-A3B-abliterated-v2-bf16Text Generation • 31B • Updated • 28
- 
	
	
	  mlx-community/Josiefied-Qwen3-30B-A3B-abliterated-v2-8bitText Generation • 31B • Updated • 115
- 
	
	
	  mlx-community/Josiefied-Qwen3-30B-A3B-abliterated-v2-6bitText Generation • 31B • Updated • 34
- 
	
	
	  mlx-community/Josiefied-Qwen3-30B-A3B-abliterated-v2-4bitText Generation • 31B • Updated • 160 • 2
The GLM-4 and Z1 series are powerful open-source language models excelling in reasoning, code, and complex tasks.
			
	
	- 
	
	
	  mlx-community/GLM-Z1-32B-0414-4bitText Generation • 5B • Updated • 142 • 2
- 
	
	
	  mlx-community/GLM-4-32B-0414-4bitText Generation • 5B • Updated • 255 • 5
- 
	
	
	  mlx-community/GLM-4-32B-Base-0414-8bitText Generation • 9B • Updated • 22
- 
	
	
	  mlx-community/GLM-4-32B-Base-0414-6bitText Generation • 7B • Updated • 26
Quantization Aware Trained (QAT) Gemma 3 checkpoints. The model preserves similar quality as half precision while using 3x less memory.
			
	
	- 
	
	
	  mlx-community/gemma-3-27b-it-qat-bf16Image-Text-to-Text • Updated • 123 • 5
- 
	
	
	  mlx-community/gemma-3-27b-it-qat-8bitImage-Text-to-Text • Updated • 171 • 9
- 
	
	
	  mlx-community/gemma-3-27b-it-qat-6bitImage-Text-to-Text • Updated • 25
- 
	
	
	  mlx-community/gemma-3-27b-it-qat-4bitImage-Text-to-Text • Updated • 83.6k • 20
- 
	
	
	  mlx-community/Llama-4-Scout-17B-16E-Instruct-4bitImage-Text-to-Text • Updated • 369 • 9
- 
	
	
	  mlx-community/Llama-4-Scout-17B-16E-Instruct-6bitImage-Text-to-Text • Updated • 210 • 5
- 
	
	
	  mlx-community/Llama-4-Scout-17B-16E-Instruct-8bitImage-Text-to-Text • Updated • 222 • 3
- 
	
	
	  mlx-community/Llama-4-Maverick-17B-16E-Instruct-4bitText Generation • 63B • Updated • 283 • 7
- 
	
	
	  mlx-community/answerdotai-ModernBERT-base-8bitFill-Mask • 53M • Updated • 3
- 
	
	
	  mlx-community/answerdotai-ModernBERT-base-4bitFill-Mask • 29.5M • Updated • 10
- 
	
	
	  mlx-community/answerdotai-ModernBERT-base-bf16Fill-Mask • 0.2B • Updated • 18 • 1
- 
	
	
	  mlx-community/answerdotai-ModernBERT-Large-Instruct-4bitFill-Mask • 70M • Updated • 4
A collection of lightweight, state-of-the-art open models built from the same research and technology that powers the Gemini 2.0 models
			
	
	- 
	
	
	  mlx-community/gemma-3-4b-it-8bitImage-Text-to-Text • 2B • Updated • 625 • 5
- 
	
	
	  mlx-community/gemma-3-4b-pt-4bitImage-Text-to-Text • 1B • Updated • 45 • 3
- 
	
	
	  mlx-community/gemma-3-4b-it-bf16Image-Text-to-Text • 5B • Updated • 157 • 1
- 
	
	
	  mlx-community/gemma-3-4b-pt-6bitImage-Text-to-Text • 1B • Updated • 9
- 
	
	
	  mlx-community/OLMoE-1B-7B-0125-InstructText Generation • 7B • Updated • 4
- 
	
	
	  mlx-community/OLMoE-1B-7B-0125-Instruct-8bitText Generation • 2B • Updated • 5
- 
	
	
	  mlx-community/OLMoE-1B-7B-0125-Instruct-6bitText Generation • 2B • Updated
- 
	
	
	  mlx-community/OLMoE-1B-7B-0125-Instruct-4bitText Generation • 1B • Updated • 28 • 2
Kokoro is an open-weight TTS model with 82 million parameters. Despite its lightweight architecture, it delivers amazing quality.
			
	
	FuseAI is attempting to merge CoT models to achieve newer models that are more than the sum of their parts.
			
	
	- 
	
	
	  mlx-community/Qwen2.5-VL-72B-Instruct-8bitImage-Text-to-Text • 21B • Updated • 59 • 2
- 
	
	
	  mlx-community/Qwen2.5-VL-72B-Instruct-6bitImage-Text-to-Text • 16B • Updated • 21 • 1
- 
	
	
	  mlx-community/Qwen2.5-VL-72B-Instruct-4bitImage-Text-to-Text • 12B • Updated • 228 • 7
- 
	
	
	  mlx-community/Qwen2.5-VL-72B-Instruct-3bitImage-Text-to-Text • 10B • Updated • 38 • 5
- 
	
	
	  mlx-community/Qwen2.5-7B-Instruct-1M-4bitText Generation • 1B • Updated • 131 • 10
- 
	
	
	  mlx-community/Qwen2.5-7B-Instruct-1M-6bitText Generation • 2B • Updated • 1 • 2
- 
	
	
	  mlx-community/Qwen2.5-7B-Instruct-1M-3bitText Generation • 1.0B • Updated • 2
- 
	
	
	  mlx-community/Qwen2.5-7B-Instruct-1M-8bitText Generation • 2B • Updated • 13 • 3
Convert HTML content to LLM-friendly Markdown/JSON content
			
	
	Kyutai's Helium-1 2B Model, outperforming other state of the art small models.
			
	
	- 
	
	
	  mlx-community/helium-1-preview-2b-float32Text Generation • 2B • Updated • 2
- 
	
	
	  mlx-community/helium-1-preview-2bText Generation • 2B • Updated
- 
	
	
	  mlx-community/helium-1-preview-2b-8bitText Generation • 0.6B • Updated • 8 • 1
- 
	
	
	  mlx-community/helium-1-preview-2b-4bitText Generation • 0.3B • Updated • 2 • 1
- 
	
	
	  mlx-community/QVQ-72B-Preview-4bitImage-Text-to-Text • 11B • Updated • 8 • 7
- 
	
	
	  mlx-community/QVQ-72B-Preview-6bitImage-Text-to-Text • 16B • Updated • 2 • 2
- 
	
	
	  mlx-community/QVQ-72B-Preview-3bitImage-Text-to-Text • 9B • Updated • 2 • 5
- 
	
	
	  mlx-community/QVQ-72B-Preview-8bitImage-Text-to-Text • 21B • Updated • 1 • 3
- 
	
	
	  mlx-community/deepseek-vl2-6bitImage-Text-to-Text • 6B • Updated • 45 • 1
- 
	
	
	  mlx-community/deepseek-vl2-small-4bitImage-Text-to-Text • 3B • Updated • 41
- 
	
	
	  mlx-community/deepseek-vl2-4bitImage-Text-to-Text • 4B • Updated • 69 • 1
- 
	
	
	  mlx-community/deepseek-vl2-small-6bitImage-Text-to-Text • 4B • Updated • 20
The best uncensored models
			
	
	- 
	
	
	  mlx-community/Josiefied-Qwen2.5-Coder-7B-Instruct-abliterated-v1Text Generation • 8B • Updated • 12
- 
	
	
	  mlx-community/Josiefied-Qwen2.5-Coder-7B-Instruct-abliterated-v1-8bitText Generation • 2B • Updated • 11
- 
	
	
	  mlx-community/Josiefied-Qwen2.5-Coder-7B-Instruct-abliterated-v1-6bitText Generation • 2B • Updated • 8
- 
	
	
	  mlx-community/Josiefied-Qwen2.5-Coder-7B-Instruct-abliterated-v1-4bitText Generation • 1B • Updated • 31 • 1
EXAONE 3.5, a collection of instruction-tuned bilingual generative models ranging from 2.4B to 32B parameters, developed by LG AI.
			
	
	- 
	
	
	  mlx-community/Llama-3.3-70B-Instruct-8bitText Generation • 20B • Updated • 373 • 14
- 
	
	
	  mlx-community/Llama-3.3-70B-Instruct-6bitText Generation • 15B • Updated • 69 • 5
- 
	
	
	  mlx-community/Llama-3.3-70B-Instruct-3bitText Generation • 9B • Updated • 106 • 7
- 
	
	
	  mlx-community/Llama-3.3-70B-Instruct-4bitText Generation • 11B • Updated • 1.07k • 30
- 
	
	
	  mlx-community/paligemma2-3b-ft-docci-448-8bitImage-Text-to-Text • 0.9B • Updated
- 
	
	
	  mlx-community/paligemma2-3b-ft-docci-448-6bitImage-Text-to-Text • 0.7B • Updated
- 
	
	
	  mlx-community/paligemma2-3b-ft-docci-448-bf16Image-Text-to-Text • 3B • Updated • 8 • 1
- 
	
	
	  mlx-community/paligemma2-10b-ft-docci-448-bf16Image-Text-to-Text • 10B • Updated • 25 • 3
- 
	
	
	  mlx-community/SmolVLM-Instruct-4bitImage-Text-to-Text • 0.5B • Updated • 228 • 5
- 
	
	
	  mlx-community/SmolVLM-Instruct-6bitImage-Text-to-Text • 0.6B • Updated • 8
- 
	
	
	  mlx-community/SmolVLM-Instruct-8bitImage-Text-to-Text • 0.7B • Updated • 23 • 9
- 
	
	
	  mlx-community/SmolVLM-Instruct-bf16Image-Text-to-Text • 2B • Updated • 13 • 5
- 
	
	
	  mlx-community/Florence-2-base-ft-4bitImage-Text-to-Text • 48.8M • Updated • 64 • 1
- 
	
	
	  mlx-community/Florence-2-large-ft-bf16Image-Text-to-Text • 0.8B • Updated • 52 • 1
- 
	
	
	  mlx-community/Florence-2-base-ft-bf16Image-Text-to-Text • 0.3B • Updated • 10 • 1
- 
	
	
	  mlx-community/Florence-2-base-ft-8bitImage-Text-to-Text • 81.7M • Updated • 29 • 1
Falcon Mamba models compatible with MLX
			
	
	Code-specific model series based on Qwen2.5
			
	
	- 
	
	
	  mlx-community/Qwen2.5-Coder-32B-Instruct-8bitText Generation • 9B • Updated • 90 • 11
- 
	
	
	  mlx-community/Qwen2.5-Coder-14B-Instruct-4bitText Generation • 2B • Updated • 165 • 4
- 
	
	
	  mlx-community/Qwen2.5-Coder-14B-Instruct-bf16Text Generation • 15B • Updated • 15 • 2
- 
	
	
	  mlx-community/Qwen2.5-Coder-3B-Instruct-8bitText Generation • 0.9B • Updated • 8
A collection of Neversleep's RP focused Lumimaid LLMs.
			
	
	Qwen1.5 is the improved version of Qwen, the large language model series developed by Alibaba Cloud.
			
	
	- 
	
	
	  mlx-community/Qwen1.5-1.8B-Chat-4bitText Generation • 0.5B • Updated • 9 • 2
- 
	
	
	  mlx-community/Qwen1.5-0.5B-Chat-4bitText Generation • 72.6M • Updated • 3.47k • 4
- 
	
	
	  mlx-community/Qwen1.5-14B-Chat-4bitText Generation • 3B • Updated • 7 • 1
- 
	
	
	  mlx-community/Qwen1.5-7B-Chat-4bitText Generation • 2B • Updated • 7 • 2
Google’s Code-Gemma
			
	
	- 
	
	
	  mlx-community/Meta-Llama-3-8B-Instruct-4bitText Generation • 2B • Updated • 1.68k • 79
- 
	
	
	  mlx-community/Meta-Llama-3-8B-4bitText Generation • 2B • Updated • 42 • 8
- 
	
	
	  mlx-community/Meta-Llama-Guard-2-8B-4bitText Generation • 2B • Updated • 5
- 
	
	
	  mlx-community/Meta-Llama-3-70B-4bitText Generation • 11B • Updated • 58 • 9
The Qwen 2.5 models are a series of AI models trained on 18 trillion tokens, supporting 29 languages and offering advanced features such as instructio
			
	
	- 
	
	
	  mlx-community/Qwen2.5-72B-Instruct-bf16Text Generation • 73B • Updated • 8
- 
	
	
	  mlx-community/Qwen2.5-72B-Instruct-8bitText Generation • 20B • Updated • 21 • 3
- 
	
	
	  mlx-community/Qwen2.5-72B-Instruct-4bitText Generation • 11B • Updated • 101 • 5
- 
	
	
	  mlx-community/Qwen2.5-32B-Instruct-bf16Text Generation • 33B • Updated • 12
- 
	
	
	  mlx-community/Phi-3-mini-4k-instruct-4bitText Generation • 0.6B • Updated • 669 • 12
- 
	
	
	  mlx-community/Phi-3-mini-128k-instruct-4bitText Generation • 0.6B • Updated • 179 • 12
- 
	
	
	  mlx-community/Phi-3-mini-128k-instruct-8bitText Generation • 1B • Updated • 38 • 10
- 
	
	
	  mlx-community/Phi-3-mini-4k-instruct-8bitText Generation • 1B • Updated • 19 • 2
OpenAI Whisper speech recognition models in MLX format
			
	
	A family of Open-source Efficient Language Models from Apple.
			
	
	Mamba is a new LLM architecture that integrates the Structured State Space sequence model to manage lengthy data sequences.
			
	
	A series of smol LLMs: 135M, 360M and 1.7B.
			
	
	- 
	
	
	  mlx-community/Meta-Llama-3.1-70B-bf16Text Generation • 71B • Updated • 18 • 4
- 
	
	
	  mlx-community/Meta-Llama-3.1-70B-Instruct-bf16Text Generation • 71B • Updated • 9 • 2
- 
	
	
	  mlx-community/Meta-Llama-3.1-8B-Instruct-bf16Text Generation • 8B • Updated • 235 • 3
- 
	
	
	  mlx-community/Meta-Llama-3.1-8B-Instruct-8bitText Generation • 2B • Updated • 318 • 10
EnCodec models in MLX
			
	
	Meta goes small with Llama3.2, both text only 1B and 3B, and the 11B Vision models.
			
	
	- 
	
	
	  mlx-community/Llama-3.2-11B-Vision-Instruct-abliteratedImage-Text-to-Text • 11B • Updated • 667 • 7
- 
	
	
	  mlx-community/Llama-3.2-11B-Vision-Instruct-abliterated-8-bitImage-Text-to-Text • 3B • Updated • 74
- 
	
	
	  mlx-community/Llama-3.2-11B-Vision-Instruct-abliterated-4-bitImage-Text-to-Text • 2B • Updated • 91 • 1
- 
	
	
	  mlx-community/Llama-3.2-11B-Vision-Instruct-8bitImage-to-Text • 3B • Updated • 581 • 10
