P

Phi 4 Mini Instruct 8da4w

Developed by pytorch
Phi-4-mini is a quantized language model developed by the PyTorch team, featuring 8-bit embeddings, 8-bit dynamic activations, and a 4-bit weight linear layer (8da4w) quantization scheme, making it suitable for mobile deployment.
Downloads 780
Release Time : 4/7/2025

Model Overview

Phi-4-mini is a lightweight natural language processing model suitable for various tasks such as code generation, mathematical reasoning, and chat conversations.

Model Features

Efficient Quantization
Uses 8-bit embeddings, 8-bit dynamic activations, and a 4-bit weight linear layer (8da4w) quantization scheme, significantly reducing model size and memory usage.
Mobile Deployment
Supports running on mobile devices via ExecuTorch, making it ideal for resource-constrained environments.
High-Performance Inference
On an iPhone 15 Pro, the model achieves a speed of 17.3 tokens per second with a memory footprint of 3206 MB.

Model Capabilities

Text Generation
Code Generation
Mathematical Reasoning
Chat Conversations

Use Cases

Natural Language Processing
Chatbot
Used to build efficient chatbots that support multi-turn conversations.
Fast response times, ideal for mobile applications.
Code Assistance
Helps developers generate code snippets or solve programming problems.
Supports multiple programming languages with high-quality output.
Education
Math Tutoring
Used to solve math problems or provide problem-solving guidance.
Performs well on the GSM8K dataset.
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
© 2025AIbase