Eclaire v0.4.0: Native Apple Silicon Support with MLX
•
TL;DR: Eclaire now runs AI models natively on Apple Silicon with MLX framework integration. New support for MLX-LM (text), MLX-VLM (vision), and LM Studio.
What is MLX?
MLX is Apple’s machine learning framework designed specifically for Apple Silicon. Built for efficient array operations on unified memory, MLX provides a NumPy-like interface with automatic differentiation and GPU acceleration through Metal.
Eclaire now leverages MLX to run AI models natively on Apple Silicon, taking full advantage of:
- Unified memory architecture: Models access the full memory pool without copying data between CPU and GPU
- Metal acceleration: Native GPU acceleration through Apple’s Metal framework delivers fast inference
- Energy efficiency: Run powerful models locally with significantly lower power consumption than traditional setups
- Privacy-first: All AI processing happens on your machine, no data leaves your device
Choose Your MLX LLM Backend
Eclaire now supports three LLM provider backends optimized for Apple Silicon. Choose the one that best fits your needs:
MLX-LM
Provides text generation capabilities using MLX-optimized models. Perfect for the assistant, chat, and content management.
MLX-VLM
Brings vision-language model support with multimodal capabilities. Ideal for photo analysis, OCR, document processing, and visual question answering. Handles both text and image inputs.
LM Studio
Offers an intuitive GUI and powerful CLI tools for model management. Browse, download, and run models with a user-friendly interface. Supports both MLX-optimized and GGUF format models, providing flexibility in model selection.
Eclaire has a backend service (for AI assistant functionality) and a workers service (for data processing, extraction, tagging, OCR, etc). Each service can use its own model.
- MLX-LM can be selected as the Eclaire backend
- MLX-VLM and LM Studio can be selected for either the Eclaire backend or workers
- Workers require vision for processing images and documents, which MLX-VLM and LM Studio are both capable of supporting.
Detailed guides for each backend are coming soon.
Enhanced Model Import Workflow
The Eclaire model CLI allows you to import model definitions directly from Hugging Face with support for identifying MLX models and their capabilities (vision, text, etc.):
- Hugging Face integration: Import models directly from Hugging Face repositories with automatic metadata parsing
- MLX model identification: Automatically detects and tags MLX-optimized models
- Capability detection: Identifies vision and text capabilities from model metadata
- Backend selection: Specific support for selecting MLX-LM, MLX-VLM, LM Studio, llama.cpp and other LLM backends
- Smart suggestions: Automatically recommends compatible backends based on model type and capabilities
- Context-aware validation: Alerts you when selecting incompatible backend/model combinations
./tools/model-cli/run.sh import https://huggingface.co/mlx-community/gemma-3-4b-it-qat-4bit
# Interactive prompt will guide you through backend selection
Imported Models - Click to view full size
This makes it much easier to import and configure models correctly on the first try.
Get Started
To try the new MLX support, you’ll need:
- macOS with Apple Silicon (M1, M2, M3, or newer)
- Eclaire v0.4.0 or above
- MLX-LM, MLX-VLM or LM Studio
Resources:
Happy building with local AI on Apple Silicon!