P

Phi 4 Mini Instruct Float8dq

Developed by pytorch
The Phi-4-mini-instruct model undergoes float8 dynamic activation and weight quantization via torchao, achieving 36% VRAM reduction and 15-20% speed improvement on H100 with minimal accuracy impact.
Downloads 1,006
Release Time : 4/8/2025

Model Overview

Quantized version based on Microsoft Phi-4-mini-instruct, suitable for text generation tasks, supporting multilingual interaction and mathematical reasoning.

Model Features

Efficient quantization
Utilizes float8 dynamic activation and weight quantization technology, significantly reducing VRAM usage
Performance optimization
Achieves 15-20% inference speed improvement on H100
Multi-task support
Supports code generation, mathematical reasoning, and dialogue tasks
Accuracy retention
Minimal accuracy loss after quantization (benchmarks show only 0.24% overall performance drop)

Model Capabilities

Text generation
Mathematical problem solving
Code generation
Multilingual dialogue
Logical reasoning

Use Cases

Educational assistance
Math problem solving
Helps students understand algebraic equation solutions
Can correctly solve equations like 2x+3=7
Creative generation
Recipe suggestions
Generates creative fruit combination recipes
Provides specific solutions like banana-dragonfruit smoothies
Technical Q&A
Programming assistance
Explains code logic or generates code snippets
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
© 2025AIbase