UNPKG

llm-checker

Version:

Intelligent CLI tool with AI-powered model selection that analyzes your hardware and recommends optimal LLM models for your system

2,014 lines (2,013 loc) 373 kB
{ "models": [ { "model_identifier": "gpt-oss", "model_name": "gpt-oss", "description": "", "labels": [], "pulls": 0, "tags": [ "gpt-oss:latest", "gpt-oss:20b", "gpt-oss:120b" ], "last_updated": "Unknown", "url": "https://ollama.com/library/gpt-oss", "namespace": null, "model_type": "official", "variants": [ { "tag": "gpt-oss:latest", "size": "unknown", "quantization": "Q4_0", "command": "ollama pull gpt-oss:latest", "estimated_size_gb": 1, "real_size_gb": 14, "categories": [ "chat", "general", "reasoning" ] }, { "tag": "gpt-oss:20b", "size": "20b", "quantization": "Q4_0", "command": "ollama pull gpt-oss:20b", "estimated_size_gb": 20, "real_size_gb": 14, "categories": [ "chat", "general", "reasoning" ] }, { "tag": "gpt-oss:120b", "size": "120b", "quantization": "Q4_0", "command": "ollama pull gpt-oss:120b", "estimated_size_gb": 120, "real_size_gb": 65, "categories": [ "chat", "general", "reasoning" ] } ], "detailed_description": "OpenAI’s open-weight models designed for powerful reasoning, agentic tasks, and versatile developer use cases.", "parameters": {}, "quantizations": [], "model_sizes": [ "20b", "120b", "14gb", "65gb", "16gb", "80gb" ], "category": "general", "use_cases": [ "general", "assistant" ], "main_size": "20b", "actual_pulls": 0, "context_length": "128K", "input_types": [ "text", "image", "code" ], "detailed_scraped_at": "2025-08-12T21:10:33.720Z", "categories": [ "chat", "general", "reasoning" ], "primary_category": "reasoning" }, { "model_identifier": "deepseek-r1", "model_name": "deepseek-r1", "description": "", "labels": [], "pulls": 0, "tags": [ "deepseek-r1:671b", "deepseek-r1:8b", "deepseek-r1:1.5b", "deepseek-r1:7b", "deepseek-r1:14b", "deepseek-r1:32b", "deepseek-r1:70b", "deepseek-r1:latest" ], "last_updated": "Unknown", "url": "https://ollama.com/library/deepseek-r1", "namespace": null, "model_type": "official", "variants": [ { "tag": "deepseek-r1:671b", "size": "671b", "quantization": "Q4_0", "command": "ollama pull deepseek-r1:671b", "estimated_size_gb": 671, "real_size_gb": 404, "categories": [ "chat", "reasoning" ] }, { "tag": "deepseek-r1:8b", "size": "8b", "quantization": "Q4_0", "command": "ollama pull deepseek-r1:8b", "estimated_size_gb": 8, "real_size_gb": 5.2, "categories": [ "chat", "reasoning" ] }, { "tag": "deepseek-r1:1.5b", "size": "1.5b", "quantization": "Q4_0", "command": "ollama pull deepseek-r1:1.5b", "estimated_size_gb": 1.5, "real_size_gb": 1.1, "categories": [ "chat", "reasoning" ] }, { "tag": "deepseek-r1:7b", "size": "7b", "quantization": "Q4_0", "command": "ollama pull deepseek-r1:7b", "estimated_size_gb": 7, "real_size_gb": 4.7, "categories": [ "chat", "reasoning" ] }, { "tag": "deepseek-r1:14b", "size": "14b", "quantization": "Q4_0", "command": "ollama pull deepseek-r1:14b", "estimated_size_gb": 14, "real_size_gb": 9, "categories": [ "chat", "reasoning" ] }, { "tag": "deepseek-r1:32b", "size": "32b", "quantization": "Q4_0", "command": "ollama pull deepseek-r1:32b", "estimated_size_gb": 32, "real_size_gb": 20, "categories": [ "chat", "reasoning" ] }, { "tag": "deepseek-r1:70b", "size": "70b", "quantization": "Q4_0", "command": "ollama pull deepseek-r1:70b", "estimated_size_gb": 70, "real_size_gb": 43, "categories": [ "chat", "reasoning" ] }, { "tag": "deepseek-r1:latest", "size": "unknown", "quantization": "Q4_0", "command": "ollama pull deepseek-r1:latest", "estimated_size_gb": 1, "real_size_gb": 5.2, "categories": [ "chat", "reasoning" ] } ], "detailed_description": "DeepSeek-R1 is a family of open reasoning models with performance approaching that of leading models, such as O3 and Gemini 2.5 Pro.", "parameters": {}, "quantizations": [], "model_sizes": [ "1.5b", "7b", "8b", "14b", "32b", "70b", "671b", "5.2gb", "1.1gb", "4.7gb", "9.0gb", "20gb", "43gb", "404gb" ], "category": "reasoning", "use_cases": [ "reasoning", "mathematics", "logic" ], "main_size": "1.5b", "actual_pulls": 0, "context_length": "160K", "input_types": [ "text", "image", "code" ], "detailed_scraped_at": "2025-08-12T21:10:33.437Z", "categories": [ "chat", "reasoning" ], "primary_category": "reasoning" }, { "model_identifier": "gemma3", "model_name": "gemma3", "description": "", "labels": [], "pulls": 0, "tags": [ "gemma3:1b", "gemma3:4b", "gemma3:12b", "gemma3:27b", "gemma3:1b-it-qat", "gemma3:4b-it-qat", "gemma3:12b-it-qat", "gemma3:27b-it-qat", "gemma3:latest" ], "last_updated": "Unknown", "url": "https://ollama.com/library/gemma3", "namespace": null, "model_type": "official", "variants": [ { "tag": "gemma3:1b", "size": "1b", "quantization": "Q4_0", "command": "ollama pull gemma3:1b", "estimated_size_gb": 1, "real_size_gb": 0.7958984375, "categories": [ "chat" ] }, { "tag": "gemma3:4b", "size": "4b", "quantization": "Q4_0", "command": "ollama pull gemma3:4b", "estimated_size_gb": 4, "real_size_gb": 3.3, "categories": [ "chat" ] }, { "tag": "gemma3:12b", "size": "12b", "quantization": "Q4_0", "command": "ollama pull gemma3:12b", "estimated_size_gb": 12, "real_size_gb": 8.1, "categories": [ "chat" ] }, { "tag": "gemma3:27b", "size": "27b", "quantization": "Q4_0", "command": "ollama pull gemma3:27b", "estimated_size_gb": 27, "real_size_gb": 17, "categories": [ "chat" ] }, { "tag": "gemma3:1b-it-qat", "size": "1b", "quantization": "Q4_0", "command": "ollama pull gemma3:1b-it-qat", "estimated_size_gb": 1, "real_size_gb": 1, "categories": [ "chat" ] }, { "tag": "gemma3:4b-it-qat", "size": "4b", "quantization": "Q4_0", "command": "ollama pull gemma3:4b-it-qat", "estimated_size_gb": 4, "real_size_gb": 4, "categories": [ "chat" ] }, { "tag": "gemma3:12b-it-qat", "size": "12b", "quantization": "Q4_0", "command": "ollama pull gemma3:12b-it-qat", "estimated_size_gb": 12, "real_size_gb": 12, "categories": [ "chat" ] }, { "tag": "gemma3:27b-it-qat", "size": "27b", "quantization": "Q4_0", "command": "ollama pull gemma3:27b-it-qat", "estimated_size_gb": 27, "real_size_gb": 27, "categories": [ "chat" ] }, { "tag": "gemma3:latest", "size": "unknown", "quantization": "Q4_0", "command": "ollama pull gemma3:latest", "estimated_size_gb": 1, "real_size_gb": 3.3, "categories": [ "chat" ] } ], "detailed_description": "The current, most capable model that runs on a single GPU.", "parameters": {}, "quantizations": [], "model_sizes": [ "1b", "4b", "12b", "27b", "3.3gb", "8.1gb", "17gb" ], "category": "talking", "use_cases": [ "chat", "conversation", "assistant" ], "main_size": "1b", "actual_pulls": 0, "context_length": "128K", "input_types": [ "text", "image", "code" ], "detailed_scraped_at": "2025-08-12T21:10:33.379Z", "categories": [ "chat" ], "primary_category": "chat" }, { "model_identifier": "qwen3", "model_name": "qwen3", "description": "", "labels": [], "pulls": 0, "tags": [ "qwen3:30b", "qwen3:235b", "qwen3:latest", "qwen3:0.6b", "qwen3:1.7b", "qwen3:4b", "qwen3:8b", "qwen3:14b", "qwen3:32b" ], "last_updated": "Unknown", "url": "https://ollama.com/library/qwen3", "namespace": null, "model_type": "official", "variants": [ { "tag": "qwen3:30b", "size": "30b", "quantization": "Q4_0", "command": "ollama pull qwen3:30b", "estimated_size_gb": 30, "real_size_gb": 19, "categories": [ "chat" ] }, { "tag": "qwen3:235b", "size": "235b", "quantization": "Q4_0", "command": "ollama pull qwen3:235b", "estimated_size_gb": 235, "real_size_gb": 142, "categories": [ "chat" ] }, { "tag": "qwen3:latest", "size": "unknown", "quantization": "Q4_0", "command": "ollama pull qwen3:latest", "estimated_size_gb": 1, "real_size_gb": 5.2, "categories": [ "chat" ] }, { "tag": "qwen3:0.6b", "size": "0.6b", "quantization": "Q4_0", "command": "ollama pull qwen3:0.6b", "estimated_size_gb": 0.6, "real_size_gb": 0.5107421875, "categories": [ "chat" ] }, { "tag": "qwen3:1.7b", "size": "1.7b", "quantization": "Q4_0", "command": "ollama pull qwen3:1.7b", "estimated_size_gb": 1.7, "real_size_gb": 1.4, "categories": [ "chat" ] }, { "tag": "qwen3:4b", "size": "4b", "quantization": "Q4_0", "command": "ollama pull qwen3:4b", "estimated_size_gb": 4, "real_size_gb": 2.5, "categories": [ "chat" ] }, { "tag": "qwen3:8b", "size": "8b", "quantization": "Q4_0", "command": "ollama pull qwen3:8b", "estimated_size_gb": 8, "real_size_gb": 5.2, "categories": [ "chat" ] }, { "tag": "qwen3:14b", "size": "14b", "quantization": "Q4_0", "command": "ollama pull qwen3:14b", "estimated_size_gb": 14, "real_size_gb": 9.3, "categories": [ "chat" ] }, { "tag": "qwen3:32b", "size": "32b", "quantization": "Q4_0", "command": "ollama pull qwen3:32b", "estimated_size_gb": 32, "real_size_gb": 20, "categories": [ "chat" ] } ], "detailed_description": "Qwen3 is the latest generation of large language models in Qwen series, offering a comprehensive suite of dense and mixture-of-experts (MoE) models.", "parameters": {}, "quantizations": [], "model_sizes": [ "0.6b", "1.7b", "4b", "8b", "14b", "30b", "32b", "235b", "5.2gb", "1.4gb", "2.5gb", "9.3gb", "19gb", "20gb", "142gb", "72b" ], "category": "talking", "use_cases": [ "chat", "conversation", "assistant" ], "main_size": "0.6b", "actual_pulls": 0, "context_length": "256K", "input_types": [ "text", "image", "code" ], "detailed_scraped_at": "2025-08-12T21:10:33.427Z", "categories": [ "chat" ], "primary_category": "chat" }, { "model_identifier": "llama3.1", "model_name": "llama3.1", "description": "", "labels": [], "pulls": 0, "tags": [ "llama3.1:latest", "llama3.1:8b", "llama3.1:70b", "llama3.1:405b" ], "last_updated": "Unknown", "url": "https://ollama.com/library/llama3.1", "namespace": null, "model_type": "official", "variants": [ { "tag": "llama3.1:latest", "size": "unknown", "quantization": "Q4_0", "command": "ollama pull llama3.1:latest", "estimated_size_gb": 1, "real_size_gb": 4.9, "categories": [ "chat" ] }, { "tag": "llama3.1:8b", "size": "8b", "quantization": "Q4_0", "command": "ollama pull llama3.1:8b", "estimated_size_gb": 8, "real_size_gb": 4.9, "categories": [ "chat" ] }, { "tag": "llama3.1:70b", "size": "70b", "quantization": "Q4_0", "command": "ollama pull llama3.1:70b", "estimated_size_gb": 70, "real_size_gb": 43, "categories": [ "chat" ] }, { "tag": "llama3.1:405b", "size": "405b", "quantization": "Q4_0", "command": "ollama pull llama3.1:405b", "estimated_size_gb": 405, "real_size_gb": 243, "categories": [ "chat" ] } ], "detailed_description": "Llama 3.1 is a new state-of-the-art model from Meta available in 8B, 70B and 405B parameter sizes.", "parameters": {}, "quantizations": [], "model_sizes": [ "8b", "70b", "405b", "4.9gb", "43gb", "243gb", "952b" ], "category": "talking", "use_cases": [ "chat", "conversation", "assistant" ], "main_size": "8b", "actual_pulls": 0, "context_length": "128K", "input_types": [ "text", "image", "code" ], "detailed_scraped_at": "2025-08-12T21:10:33.423Z", "categories": [ "chat" ], "primary_category": "chat" }, { "model_identifier": "nomic-embed-text", "model_name": "nomic-embed-text", "description": "", "labels": [], "pulls": 0, "tags": [ "nomic-embed-text:latest", "nomic-embed-text:v1.5" ], "last_updated": "Unknown", "url": "https://ollama.com/library/nomic-embed-text", "namespace": null, "model_type": "official", "variants": [ { "tag": "nomic-embed-text:latest", "size": "unknown", "quantization": "Q4_0", "command": "ollama pull nomic-embed-text:latest", "estimated_size_gb": 1, "real_size_gb": 0.267578125, "categories": [ "chat", "embeddings" ] }, { "tag": "nomic-embed-text:v1.5", "size": "unknown", "quantization": "Q4_0", "command": "ollama pull nomic-embed-text:v1.5", "estimated_size_gb": 1, "real_size_gb": 0.267578125, "categories": [ "chat", "embeddings" ] } ], "detailed_description": "A high-performing open embedding model with a large token context window.", "parameters": {}, "quantizations": [], "model_sizes": [], "category": "embeddings", "use_cases": [ "embeddings", "search", "similarity" ], "main_size": "Unknown", "actual_pulls": 0, "context_length": "2K", "input_types": [ "text", "image", "code" ], "detailed_scraped_at": "2025-08-12T21:10:35.029Z", "categories": [ "chat", "embeddings" ], "primary_category": "embeddings" }, { "model_identifier": "llama3.2", "model_name": "llama3.2", "description": "", "labels": [], "pulls": 0, "tags": [ "llama3.2:1b", "llama3.2:latest", "llama3.2:3b" ], "last_updated": "Unknown", "url": "https://ollama.com/library/llama3.2", "namespace": null, "model_type": "official", "variants": [ { "tag": "llama3.2:1b", "size": "1b", "quantization": "Q4_0", "command": "ollama pull llama3.2:1b", "estimated_size_gb": 1, "real_size_gb": 1.3, "categories": [ "chat" ] }, { "tag": "llama3.2:latest", "size": "unknown", "quantization": "Q4_0", "command": "ollama pull llama3.2:latest", "estimated_size_gb": 1, "real_size_gb": 2, "categories": [ "chat" ] }, { "tag": "llama3.2:3b", "size": "3b", "quantization": "Q4_0", "command": "ollama pull llama3.2:3b", "estimated_size_gb": 3, "real_size_gb": 2, "categories": [ "chat" ] } ], "detailed_description": "Meta's Llama 3.2 goes small with 1B and 3B models.", "parameters": {}, "quantizations": [], "model_sizes": [ "1b", "3b", "2.0gb", "1.3gb", "929b", "2.6b" ], "category": "talking", "use_cases": [ "chat", "conversation", "assistant" ], "main_size": "1b", "actual_pulls": 0, "context_length": "128K", "input_types": [ "text", "image", "code" ], "detailed_scraped_at": "2025-08-12T21:10:35.041Z", "categories": [ "chat" ], "primary_category": "chat" }, { "model_identifier": "mistral", "model_name": "mistral", "description": "", "labels": [], "pulls": 0, "tags": [ "mistral:latest", "mistral:7b" ], "last_updated": "Unknown", "url": "https://ollama.com/library/mistral", "namespace": null, "model_type": "official", "variants": [ { "tag": "mistral:latest", "size": "unknown", "quantization": "Q4_0", "command": "ollama pull mistral:latest", "estimated_size_gb": 1, "real_size_gb": 4.4, "categories": [ "chat" ] }, { "tag": "mistral:7b", "size": "7b", "quantization": "Q4_0", "command": "ollama pull mistral:7b", "estimated_size_gb": 7, "real_size_gb": 4.4, "categories": [ "chat" ] } ], "detailed_description": "The 7B model released by Mistral AI, updated to version 0.3.", "parameters": {}, "quantizations": [], "model_sizes": [ "7b", "4.4gb", "417b", "13b", "34b" ], "category": "talking", "use_cases": [ "chat", "conversation", "assistant" ], "main_size": "7b", "actual_pulls": 0, "context_length": "32K", "input_types": [ "text", "image", "code" ], "detailed_scraped_at": "2025-08-12T21:10:35.022Z", "categories": [ "chat" ], "primary_category": "chat" }, { "model_identifier": "qwen2.5", "model_name": "qwen2.5", "description": "", "labels": [], "pulls": 0, "tags": [ "qwen2.5:latest", "qwen2.5:0.5b", "qwen2.5:1.5b", "qwen2.5:3b", "qwen2.5:7b", "qwen2.5:14b", "qwen2.5:32b", "qwen2.5:72b" ], "last_updated": "Unknown", "url": "https://ollama.com/library/qwen2.5", "namespace": null, "model_type": "official", "variants": [ { "tag": "qwen2.5:latest", "size": "unknown", "quantization": "Q4_0", "command": "ollama pull qwen2.5:latest", "estimated_size_gb": 1, "real_size_gb": 4.7, "categories": [ "chat" ] }, { "tag": "qwen2.5:0.5b", "size": "0.5b", "quantization": "Q4_0", "command": "ollama pull qwen2.5:0.5b", "estimated_size_gb": 0.5, "real_size_gb": 0.388671875, "categories": [ "chat" ] }, { "tag": "qwen2.5:1.5b", "size": "1.5b", "quantization": "Q4_0", "command": "ollama pull qwen2.5:1.5b", "estimated_size_gb": 1.5, "real_size_gb": 0.962890625, "categories": [ "chat" ] }, { "tag": "qwen2.5:3b", "size": "3b", "quantization": "Q4_0", "command": "ollama pull qwen2.5:3b", "estimated_size_gb": 3, "real_size_gb": 1.9, "categories": [ "chat" ] }, { "tag": "qwen2.5:7b", "size": "7b", "quantization": "Q4_0", "command": "ollama pull qwen2.5:7b", "estimated_size_gb": 7, "real_size_gb": 4.7, "categories": [ "chat" ] }, { "tag": "qwen2.5:14b", "size": "14b", "quantization": "Q4_0", "command": "ollama pull qwen2.5:14b", "estimated_size_gb": 14, "real_size_gb": 9, "categories": [ "chat" ] }, { "tag": "qwen2.5:32b", "size": "32b", "quantization": "Q4_0", "command": "ollama pull qwen2.5:32b", "estimated_size_gb": 32, "real_size_gb": 20, "categories": [ "chat" ] }, { "tag": "qwen2.5:72b", "size": "72b", "quantization": "Q4_0", "command": "ollama pull qwen2.5:72b", "estimated_size_gb": 72, "real_size_gb": 47, "categories": [ "chat" ] } ], "detailed_description": "Qwen2.5 models are pretrained on Alibaba's latest large-scale dataset, encompassing up to 18 trillion tokens. The model supports up to 128K tokens and has multilingual support.", "parameters": {}, "quantizations": [], "model_sizes": [ "0.5b", "1.5b", "3b", "7b", "14b", "32b", "72b", "4.7gb", "1.9gb", "9.0gb", "20gb", "47gb" ], "category": "talking", "use_cases": [ "chat", "conversation", "assistant" ], "main_size": "0.5b", "actual_pulls": 0, "context_length": "32K", "input_types": [ "text", "image", "code" ], "detailed_scraped_at": "2025-08-12T21:10:34.954Z", "categories": [ "chat" ], "primary_category": "chat" }, { "model_identifier": "llama3", "model_name": "llama3", "description": "", "labels": [], "pulls": 0, "tags": [ "llama3:70b", "llama3:text", "llama3:70b-text", "llama3:latest", "llama3:8b" ], "last_updated": "Unknown", "url": "https://ollama.com/library/llama3", "namespace": null, "model_type": "official", "variants": [ { "tag": "llama3:70b", "size": "70b", "quantization": "Q4_0", "command": "ollama pull llama3:70b", "estimated_size_gb": 70, "real_size_gb": 40, "categories": [ "chat" ] }, { "tag": "llama3:text", "size": "unknown", "quantization": "Q4_0", "command": "ollama pull llama3:text", "estimated_size_gb": 1, "real_size_gb": 1, "categories": [ "chat" ] }, { "tag": "llama3:70b-text", "size": "70b", "quantization": "Q4_0", "command": "ollama pull llama3:70b-text", "estimated_size_gb": 70, "real_size_gb": 70, "categories": [ "chat" ] }, { "tag": "llama3:latest", "size": "unknown", "quantization": "Q4_0", "command": "ollama pull llama3:latest", "estimated_size_gb": 1, "real_size_gb": 4.7, "categories": [ "chat" ] }, { "tag": "llama3:8b", "size": "8b", "quantization": "Q4_0", "command": "ollama pull llama3:8b", "estimated_size_gb": 8, "real_size_gb": 4.7, "categories": [ "chat" ] } ], "detailed_description": "Meta Llama 3: The most capable openly available LLM to date", "parameters": {}, "quantizations": [], "model_sizes": [ "8b", "70b", "4.7gb", "40gb", "876b" ], "category": "talking", "use_cases": [ "chat", "conversation", "assistant" ], "main_size": "8b", "actual_pulls": 0, "context_length": "8K", "input_types": [ "text", "image", "code" ], "detailed_scraped_at": "2025-08-12T21:10:35.060Z", "categories": [ "chat" ], "primary_category": "chat" }, { "model_identifier": "llava", "model_name": "llava", "description": "", "labels": [], "pulls": 0, "tags": [ "llava:latest", "llava:7b", "llava:13b", "llava:34b" ], "last_updated": "Unknown", "url": "https://ollama.com/library/llava", "namespace": null, "model_type": "official", "variants": [ { "tag": "llava:latest", "size": "unknown", "quantization": "Q4_0", "command": "ollama pull llava:latest", "estimated_size_gb": 1, "real_size_gb": 4.7, "categories": [ "chat", "multimodal" ] }, { "tag": "llava:7b", "size": "7b", "quantization": "Q4_0", "command": "ollama pull llava:7b", "estimated_size_gb": 7, "real_size_gb": 4.7, "categories": [ "chat", "multimodal" ] }, { "tag": "llava:13b", "size": "13b", "quantization": "Q4_0", "command": "ollama pull llava:13b", "estimated_size_gb": 13, "real_size_gb": 8, "categories": [ "chat", "multimodal" ] }, { "tag": "llava:34b", "size": "34b", "quantization": "Q4_0", "command": "ollama pull llava:34b", "estimated_size_gb": 34, "real_size_gb": 20, "categories": [ "chat", "multimodal" ] } ], "detailed_description": "🌋 LLaVA is a novel end-to-end trained large multimodal model that combines a vision encoder and Vicuna for general-purpose visual and language understanding. Updated to version 1.6.", "parameters": {}, "quantizations": [], "model_sizes": [ "7b", "13b", "34b", "4.7gb", "8.0gb", "20gb" ], "category": "multimodal", "use_cases": [ "vision", "multimodal", "image" ], "main_size": "7b", "actual_pulls": 0, "context_length": "32K", "input_types": [ "text", "image", "code" ], "detailed_scraped_at": "2025-08-12T21:10:36.304Z", "categories": [ "chat", "multimodal" ], "primary_category": "multimodal" }, { "model_identifier": "phi3", "model_name": "phi3", "description": "", "labels": [], "pulls": 0, "tags": [ "phi3:mini", "phi3:medium", "phi3:medium-128k", "phi3:latest", "phi3:3.8b", "phi3:14b" ], "last_updated": "Unknown", "url": "https://ollama.com/library/phi3", "namespace": null, "model_type": "official", "variants": [ { "tag": "phi3:mini", "size": "unknown", "quantization": "Q4_0", "command": "ollama pull phi3:mini", "estimated_size_gb": 1, "real_size_gb": 1, "categories": [ "chat" ] }, { "tag": "phi3:medium", "size": "unknown", "quantization": "Q4_0", "command": "ollama pull phi3:medium", "estimated_size_gb": 1, "real_size_gb": 1, "categories": [ "chat" ] }, { "tag": "phi3:medium-128k", "size": "unknown", "quantization": "Q4_0", "command": "ollama pull phi3:medium-128k", "estimated_size_gb": 1, "real_size_gb": 1, "categories": [ "chat" ] }, { "tag": "phi3:latest", "size": "unknown", "quantization": "Q4_0", "command": "ollama pull phi3:latest", "estimated_size_gb": 1, "real_size_gb": 2.2, "categories": [ "chat" ] }, { "tag": "phi3:3.8b", "size": "3.8b", "quantization": "Q4_0", "command": "ollama pull phi3:3.8b", "estimated_size_gb": 3.8, "real_size_gb": 2.2, "categories": [ "chat" ] }, { "tag": "phi3:14b", "size": "14b", "quantization": "Q4_0", "command": "ollama pull phi3:14b", "estimated_size_gb": 14, "real_size_gb": 7.9, "categories": [ "chat" ] } ], "detailed_description": "Phi-3 is a family of lightweight 3B (Mini) and 14B (Medium) state-of-the-art open models by Microsoft.", "parameters": {}, "quantizations": [], "model_sizes": [ "3b", "14b", "3.8b", "2.2gb", "7.9gb", "80g" ], "category": "talking", "use_cases": [ "chat", "conversation", "assistant" ], "main_size": "3b", "actual_pulls": 0, "context_length": "128K", "input_types": [ "text", "image", "code" ], "detailed_scraped_at": "2025-08-12T21:10:36.415Z", "categories": [ "chat" ], "primary_category": "chat" }, { "model_identifier": "gemma2", "model_name": "gemma2", "description": "", "labels": [], "pulls": 0, "tags": [ "gemma2:2b", "gemma2:27b", "gemma2:latest", "gemma2:9b" ], "last_updated": "Unknown", "url": "https://ollama.com/library/gemma2", "namespace": null, "model_type": "official", "variants": [ { "tag": "gemma2:2b", "size": "2b", "quantization": "Q4_0", "command": "ollama pull gemma2:2b", "estimated_size_gb": 2, "real_size_gb": 1.6, "categories": [ "chat" ] }, { "tag": "gemma2:27b", "size": "27b", "quantization": "Q4_0", "command": "ollama pull gemma2:27b", "estimated_size_gb": 27, "real_size_gb": 16, "categories": [ "chat" ] }, { "tag": "gemma2:latest", "size": "unknown", "quantization": "Q4_0", "command": "ollama pull gemma2:latest", "estimated_size_gb": 1, "real_size_gb": 5.4, "categories": [ "chat" ] }, { "tag": "gemma2:9b", "size": "9b", "quantization": "Q4_0", "command": "ollama pull gemma2:9b", "estimated_size_gb": 9, "real_size_gb": 5.4, "categories": [ "chat" ] } ], "detailed_description": "Google Gemma 2 is a high-performing and efficient model available in three sizes: 2B, 9B, and 27B.", "parameters": {}, "quantizations": [], "model_sizes": [ "2b", "9b", "27b", "5.4gb", "1.6gb", "16gb", "871b" ], "category": "talking", "use_cases": [ "chat", "conversation", "assistant" ], "main_size": "2b", "actual_pulls": 0, "context_length": "8K", "input_types": [ "text", "image", "code" ], "detailed_scraped_at": "2025-08-12T21:10:36.409Z", "categories": [ "chat" ], "primary_category": "chat" }, { "model_identifier": "qwen2.5-coder", "model_name": "qwen2.5-coder", "description": "", "labels": [], "pulls": 0, "tags": [ "qwen2.5-coder:32b", "qwen2.5-coder:14b", "qwen2.5-coder:7b", "qwen2.5-coder:3b", "qwen2.5-coder:1.5b", "qwen2.5-coder:0.5b", "qwen2.5-coder:latest" ], "last_updated": "Unknown", "url": "https://ollama.com/library/qwen2.5-coder", "namespace": null, "model_type": "official", "variants": [ { "tag": "qwen2.5-coder:32b", "size": "32b", "quantization": "Q4_0", "command": "ollama pull qwen2.5-coder:32b", "estimated_size_gb": 32, "real_size_gb": 20, "categories": [ "coding", "reasoning" ] }, { "tag": "qwen2.5-coder:14b", "size": "14b", "quantization": "Q4_0", "command": "ollama pull qwen2.5-coder:14b", "estimated_size_gb": 14, "real_size_gb": 9, "categories": [ "coding", "reasoning" ] }, { "tag": "qwen2.5-coder:7b", "size": "7b", "quantization": "Q4_0", "command": "ollama pull qwen2.5-coder:7b", "estimated_size_gb": 7, "real_size_gb": 4.7, "categories": [ "coding", "reasoning" ] }, { "tag": "qwen2.5-coder:3b", "size": "3b", "quantization": "Q4_0", "command": "ollama pull qwen2.5-coder:3b", "estimated_size_gb": 3, "real_size_gb": 1.9, "categories": [ "coding", "reasoning" ] }, { "tag": "qwen2.5-coder:1.5b", "size": "1.5b", "quantization": "Q4_0", "command": "ollama pull qwen2.5-coder:1.5b", "estimated_size_gb": 1.5, "real_size_gb": 0.962890625, "categories": [ "coding", "reasoning" ] }, { "tag": "qwen2.5-coder:0.5b", "size": "0.5b", "quantization": "Q4_0", "command": "ollama pull qwen2.5-coder:0.5b", "estimated_size_gb": 0.5, "real_size_gb": 0.388671875, "categories": [ "coding", "reasoning" ] }, { "tag": "qwen2.5-coder:latest", "size": "unknown", "quantization": "Q4_0", "command": "ollama pull qwen2.5-coder:latest", "estimated_size_gb": 1, "real_size_gb": 4.7, "categories": [ "coding", "reasoning" ] } ], "detailed_description": "The latest series of Code-Specific Qwen models, with significant improvements in code generation, code reasoning, and code fixing.", "parameters": {}, "quantizations": [], "model_sizes": [ "0.5b", "1.5b", "3b", "7b", "14b", "32b", "4.7gb", "1.9gb", "9.0gb", "20gb", "474b", "6436978b" ], "category": "coding", "use_cases": [ "coding", "programming", "development" ], "main_size": "0.5b", "actual_pulls": 0, "context_length": "32K", "input_types": [ "text", "image", "code" ], "detailed_scraped_at": "2025-08-12T21:10:36.428Z", "categories": [ "coding", "reasoning" ], "primary_category": "coding" }, { "model_identifier": "gemma", "model_name": "gemma", "description": "", "labels": [], "pulls": 0, "tags": [ "gemma:2b", "gemma:7b", "gemma:latest" ], "last_updated": "Unknown", "url": "https://ollama.com/library/gemma", "namespace": null, "model_type": "official", "variants": [ { "tag": "gemma:2b", "size": "2b", "quantization": "Q4_0", "command": "ollama pull gemma:2b", "estimated_size_gb": 2, "real_size_gb": 1.7, "categories": [ "chat" ] }, { "tag": "gemma:7b", "size": "7b", "quantization": "Q4_0", "command": "ollama pull gemma:7b", "estimated_size_gb": 7, "real_size_gb": 5, "categories": [ "chat" ] }, { "tag": "gemma:latest", "size": "unknown", "quantization": "Q4_0", "command": "ollama pull gemma:latest", "estimated_size_gb": 1, "real_size_gb": 5, "categories": [ "chat" ] } ], "detailed_description": "Gemma is a family of lightweight, state-of-the-art open models built by Google DeepMind. Updated to version 1.1", "parameters": {}, "quantizations": [], "model_sizes": [ "2b", "7b", "5.0gb", "1.7gb" ], "category": "talking", "use_cases": [ "chat", "conversation", "assistant" ], "main_size": "2b", "actual_pulls": 0, "context_length": "8K", "input_types": [ "text", "image", "code" ], "detailed_scraped_at": "2025-08-12T21:10:36.355Z", "categories": [ "chat" ], "primary_category": "chat" }, { "model_identifier": "qwen", "model_name": "qwen", "description": "", "labels": [], "pulls": 0, "tags": [ "qwen:0.5b", "qwen:1.8b", "qwen:4b", "qwen:7b", "qwen:14b", "qwen:32b", "qwen:72b", "qwen:110b", "qwen:latest" ], "last_updated": "Unknown", "url": "https://ollama.com/library/qwen", "namespace": null, "model_type": "official", "variants": [ { "tag": "qwen:0.5b", "size": "0.5b", "quantization": "Q4_0", "command": "ollama pull qwen:0.5b", "estimated_size_gb": 0.5, "real_size_gb": 0.3857421875, "categories": [ "chat" ] }, { "tag": "qwen:1.8b", "size": "1.8b", "quantization": "Q4_0", "command": "ollama pull qwen:1.8b", "estimated_size_gb": 1.8, "real_size_gb": 1.1, "categories": [ "chat" ] }, { "tag": "qwen:4b", "size": "4b", "quantization": "Q4_0", "command": "ollama pull qwen:4b", "estimated_size_gb": 4, "real_size_gb": 2.3, "categories": [ "chat" ] }, { "tag": "qwen:7b", "size": "7b", "quantization": "Q4_0", "command": "ollama pull qwen:7b", "estimated_size_gb": 7, "real_size_gb": 4.5, "categories": [ "chat" ] }, { "tag": "qwen:14b", "size": "14b", "quantization": "Q4_0", "command": "ollama pull qwen:14b", "estimated_size_gb": 14, "real_size_gb": 8.2, "categories": [ "chat" ] }, { "tag": "qwen:32b", "size": "32b", "quantization": "Q4_0", "command": "ollama pull qwen:32b", "estimated_size_gb": 32, "real_size_gb": 18, "categories": [ "chat" ] }, { "tag": "qwen:72b", "size": "72b", "quantization": "Q4_0", "command": "ollama pull qwen:72b", "estimated_size_gb": 72, "real_size_gb": 41, "categories": [ "chat" ] }, { "tag": "qwen:110b", "size": "110b", "quantization": "Q4_0", "command": "ollama pull qwen:110b", "estimated_size_gb": 110, "real_size_gb": 63, "categories": [ "chat" ] }, { "tag": "qwen:latest", "size": "unknown", "quantization": "Q4_0", "command": "ollama pull qwen:latest", "estimated_size_gb": 1, "real_size_gb": 2.3, "categories": [ "chat" ] } ], "detailed_description": "Qwen 1.5 is a series of large language models by Alibaba Cloud spanning from 0.5B to 110B parameters", "parameters": {}, "quantizations": [], "model_sizes": [ "0.5b", "110b", "1.8b", "4b", "7b", "14b", "32b", "72b", "2.3gb", "1.1gb", "4.5gb", "8.2gb", "18gb", "41gb", "63gb", "2gb" ], "category": "talking", "use_cases": [ "chat", "conversation", "assistant" ], "main_size": "0.5b", "actual_pulls": 0, "context_length": "32K", "input_types": [ "text", "image", "code" ], "detailed_scraped_at": "2025-08-12T21:10:37.681Z", "categories": [ "chat" ], "primary_category": "chat" }, { "model_identifier": "mxbai-embed-large", "model_name": "mxbai-embed-large", "description": "", "labels": [], "pulls": 0, "tags": [ "mxbai-embed-large:latest", "mxbai-embed-large:335m" ], "last_updated": "Unknown", "url": "https://ollama.com/library/mxbai-embed-large", "namespace": null, "model_type": "official", "variants": [ { "tag": "mxbai-embed-large:latest", "size": "unknown", "quantization": "Q4_0", "command": "ollama pull mxbai-embed-large:latest", "estimated_size_gb": 1, "real_size_gb": 0.654296875, "categories": [ "chat", "embeddings" ] }, { "tag": "mxbai-embed-large:335m", "size": "unknown", "quantization": "Q4_0", "command": "ollama pull mxbai-embed-large:335m", "estimated_size_gb": 1, "real_size_gb": 0.654296875, "categories": [ "chat", "embeddings" ] } ], "detailed_description": "State-of-the-art large embedding model from mixedbread.ai", "parameters": {}, "quantizations": [], "model_sizes": [], "category": "embeddings", "use_cases": [ "embeddings", "search", "similarity" ], "main_size": "Unknown", "actual_pulls": 0, "context_length": "512", "input_types": [ "text", "image", "code" ], "detailed_scraped_at": "2025-08-12T21:10:37.606Z", "categories": [ "chat", "embeddings" ], "primary_category": "embeddings" }, { "model_identifier": "qwen2", "model_name": "qwen2", "description": "", "labels": [], "pulls": 0, "tags": [ "qwen2:latest", "qwen2:0.5b", "qwen2:1.5b", "qwen2:7b", "qwen2:72b" ], "last_updated": "Unknown", "url": "https://ollama.com/library/qwen2", "namespace": null, "model_type": "official", "variants": [ { "tag": "qwen2:latest", "size": "unknown", "quantization": "Q4_0", "command": "ollama pull qwen2:latest", "estimated_size_gb": 1, "real_size_gb": 4.4, "categories": [ "chat" ] }, { "tag": "qwen2:0.5b", "size": "0.5b", "quantization": "Q4_0", "command": "ollama pull qwen2:0.5b", "estimated_size_gb": 0.5, "real_size_gb": 0.34375, "categories": [ "chat" ] }, { "tag": "qwen2:1.5b", "size": "1.5b", "quantization": "Q4_0", "command": "ollama pull qwen2:1.5b", "estimated_size_gb": 1.5, "real_size_gb": 0.9130859375, "categories": [ "chat" ] }, { "tag": "qwen2:7b", "size": "7b", "quantization": "Q4_0", "command": "ollama pull qwen2:7b", "estimated_size_gb": 7, "real_size_gb": 4.4, "categories": [ "chat" ] }, { "tag": "qwen2:72b", "size": "72b", "quantization": "Q4_0", "command": "ollama pull qwen2:72b", "estimated_size_gb": 72, "real_size_gb": 41, "categories": [ "chat" ] } ], "detailed_description": "Qwen2 is a new series of large language models from Alibaba group", "parameters": {}, "quantizations": [], "model_sizes": [ "0.5b", "1.5b", "7b", "72b", "4.4gb", "41gb", "0.49b", "1.54b", "7.07b", "72.71b", "0.35b", "1.31b", "5.98b", "70.21b" ], "category": "talking", "use_cases": [ "chat", "conversation", "assistant" ], "main_size": "0.5b", "actual_pulls": 0, "context_length": "32K", "input_types": [ "text", "image", "code" ], "detailed_scraped_at": "2025-08-12T21:10:37.590Z", "categories": [ "chat" ], "primary_category": "chat" }, { "model_identifier": "llama2", "model_name": "llama2", "description": "", "labels": [], "pulls": 0, "tags": [ "llama2:text", "llama2:latest", "llama2:7b", "llama2:13b", "llama2:70b" ], "last_updated": "Unknown", "url": "https://ollama.com/library/llama2", "namespace": null, "model_type": "official",