Smollm 135M Instruct
A lightweight instruction fine-tuned language model optimized for mobile deployment
Downloads 131
Release Time : 4/30/2025
Model Overview
This model is a variant of HuggingFaceTB/SmolLM-135M-Instruct, supporting efficient operation on Android devices through the LiteRT framework and MediaPipe LLM inference API.
Model Features
Mobile optimization
Optimized for Android/iOS/Web platforms, supporting efficient deployment
Quantization support
Provides dynamic_int8 and dynamic_int4 quantization versions, significantly reducing the model size
Efficient inference
Accelerated by LiteRT XNNPACK, supporting 4-thread CPU inference
Low memory usage
The memory usage of the quantized model is significantly reduced, suitable for mobile devices
Model Capabilities
Instruction following
Text generation
Mobile inference
Use Cases
Mobile applications
Device-side chat assistant
Deploy a localized chat application on Android devices
The quantized version can significantly reduce resource usage while maintaining performance
Featured Recommended AI Models
Š 2025AIbase