S

Smollm 135M Instruct

Developed by litert-community
A lightweight instruction fine-tuned language model optimized for mobile deployment
Downloads 131
Release Time : 4/30/2025

Model Overview

This model is a variant of HuggingFaceTB/SmolLM-135M-Instruct, supporting efficient operation on Android devices through the LiteRT framework and MediaPipe LLM inference API.

Model Features

Mobile optimization
Optimized for Android/iOS/Web platforms, supporting efficient deployment
Quantization support
Provides dynamic_int8 and dynamic_int4 quantization versions, significantly reducing the model size
Efficient inference
Accelerated by LiteRT XNNPACK, supporting 4-thread CPU inference
Low memory usage
The memory usage of the quantized model is significantly reduced, suitable for mobile devices

Model Capabilities

Instruction following
Text generation
Mobile inference

Use Cases

Mobile applications
Device-side chat assistant
Deploy a localized chat application on Android devices
The quantized version can significantly reduce resource usage while maintaining performance
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
Š 2025AIbase