Model HQ
DocumentationApple Supported Models
Comprehensive list of 120+ AI models optimized for Apple Silicon and Apple Neural Engine
These models are optimized for Apple Silicon processors and leverage Metal Performance Shaders and the Apple Neural Engine for enhanced performance on macOS and iOS devices.
Apple Optimization Features
Key benefits of Apple-optimized models
Performance Benefits
- • Optimized for Apple Silicon architectures (M-series chips)
- • Enhanced inference speed with Metal Performance Shaders
- • Power-efficient execution on Apple Neural Engine
- • Unified memory architecture for fast data access
Supported Hardware
- • Apple Silicon M1, M2, M3, and M4 series processors
- • Apple Neural Engine (ANE)
- • Apple integrated GPUs
- • macOS and iOS devices
Apple Supported Models
Complete catalog of models optimized for Apple processors
| Model Type | Available Models |
|---|---|
Embedding Models | all-mini-lm-L6-v2all-mpnet-base-v2industry-bert-insuranceindustry-bert-contractsindustry-bert-asset-managementindustry-bert-secindustry-bert-loansnomic-ai/nomic-embed-text-v1 |
Jina Models | jina-reranker-tiny-pptjina-reranker-turbo-pptjina-reranker-tiny-onnxjina-reranker-turbo-onnxjina-reranker-v1-turbo-enjina-reranker-v1-tiny-en |
GPT Models | gpt-5.2-progpt-5.2gpt-5-minigpt-5-nanogpt-4.1 |
Claude Models | claude-opus-4-5claude-haiku-4-5claude-sonnet-4-5claude-sonnet-4-20250514claude-opus-4-20250514 |
Gemini Models | gemini-3-pro-previewgemini-3-flash-previewgemini-2.5-progemini-2.5-flashgemini-2.5-flash-lite |
Qwen Models | qwen2.5-vl-3b-instruct-ggufqwen3-vl-8b-ggufqwen3-vl-4b-ggufqwen3-vl-30b-ggufbling-qwen-mini-toolbling-qwen-0.5b-ggufdragon-qwen-7b-ggufqwen2-7B-instruct-ggufqwen3-1.7b-ggufqwen3-4b-instruct-ggufqwen3-8b-ggufqwen3-14b-ggufqwen-3.5-4b-ggufqwen-3.5-9b-ggufqwen-3.5-27b-ggufqwen-3.5-35b-a3b-ggufqwen2-1.5b-instruct-ggufqwen2-0.5b-instruct-ggufqwen-2.5-7b-coder-ggufqwen-2.5-14b-instruct-ggufqwen2.5-32b-ggufqwen2.5-72b-ggufdeepseek-qwen-14b-ggufdeepseek-qwen-7b-gguf |
Llama-Based Models | llama-3.1-instruct-ggufllama-2-7b-chat-ggufllama-3-8b-instruct-gguftiny-llama-chat-ggufllama-3.2-1b-instruct-ggufllama-3.2-3b-instruct-ggufdragon-llama-3.1-gguf |
Phi Models | bling-phi-3-ggufbling-phi-3.5-ggufphi-3.5-ggufphi-4-ggufphi-4-mini-ggufphi-4-mini-reasoning-ggufphi-3-ggufphi-3-ppt |
Mistral Models | dragon-mistral-0.3-ggufmistral-small-3.2-24b-ggufministral-3-14b-ggufopenhermes-2.5-mistral-7b-ggufzephyr-7b-beta-ggufstarling-lm-7b-alpha-ggufmistral-7b-instruct-v0.3-gguf |
Yi Models | dragon-yi-9b-ggufdragon-yi-answer-tool |
Gemma Models | gemma-3-4b-ggufgemma-3-12b-ggufgemma-4-4b-ggufgemma-4-2b-ggufgemma-4-26b-ggufgemma-2-9b-instruct-ggufgemma-2-27b-instruct-gguf |
StableLM Models | bling-stablelm-3b-gguf |
Dragon Models | dragon-llama-3.1-ggufdragon-mistral-0.3-ggufdragon-yi-9b-ggufdragon-qwen-7b-ggufbling-qwen-mini-toolbling-qwen-0.5b-ggufdragon-yi-answer-tooldragon-llama-answer-tooldragon-mistral-answer-tool |
Slim Models | slim-ner-toolslim-sentiment-toolslim-emotions-toolslim-ratings-toolslim-intent-toolslim-nli-toolslim-topics-toolslim-tags-toolslim-sql-toolbling-answer-toolslim-category-toolslim-xsum-toolslim-extract-toolslim-extract-phi-3-ggufslim-extract-qwen-1.5b-ggufslim-extract-qwen-nano-ggufslim-extract-tiny-toolslim-summary-tiny-toolslim-summary-phi-3-ggufslim-xsum-phi-3-ggufslim-boolean-toolslim-boolean-phi-3-ggufslim-sa-ner-phi-3-ggufslim-sa-ner-toolslim-tags-3b-toolslim-summary-toolslim-q-gen-phi-3-toolslim-q-gen-tiny-toolslim-qa-gen-tiny-toolslim-qa-gen-phi-3-tool |
Specialized Models | gpt-oss-20b-ggufolmo-13b-ggufgranite-4-micro-ggufliquidai-lfm2-2.6b-ggufminicpm-2.6-ggufwhisper-cpp-base-english |
Getting Started with Apple Models
To use Apple-optimized models in Model HQ:
- Ensure you have a device with Apple Silicon (M1 or later)
- Select models optimized for Apple from the Models section
- The system will automatically use Apple optimizations when available
- Monitor performance improvements and power efficiency
Technical Support
For Apple-specific optimization questions or issues, contact our technical support team at support@aibloks.com
🚀 Performance Tip
Apple-optimized models are designed for efficient on-device computing. They provide excellent performance while maintaining low power consumption thanks to Apple Silicon's unified memory architecture and Neural Engine.
Check System Requirements