This guide covers the complete lifecycle of MLC models - from discovery and download to cleanup and removal.
The package includes a prebuilt runtime optimized for the following models:
Model ID | Size |
---|---|
Llama-3.2-3B-Instruct | ~2GB |
Phi-3-mini-4k-instruct | ~2.5GB |
Mistral-7B-Instruct | ~4.5GB (requires 8GB+ RAM) |
Qwen2.5-1.5B-Instruct | ~1GB |
Note: These are the only models supported for direct download. For other models, you'll need to build MLC from source (documentation coming soon).
Get the list of models included in the runtime:
Create a model instance using the mlc.languageModel()
method:
Models need to be downloaded to the device before use.
You can track download progress:
After downloading, prepare the model for inference:
Once prepared, use the model with AI SDK functions:
Unload the current model from memory to free resources:
Delete downloaded model files to free storage: