D

Deepseek Coder V2 Lite Instruct FP8

Developed by RedHatAI
FP8 quantized version of DeepSeek-Coder-V2-Lite-Instruct, suitable for commercial and research use in English, optimized for inference efficiency.
Downloads 11.29k
Release Time : 7/17/2024

Model Overview

This model is a quantized version of DeepSeek-Coder-V2-Lite-Instruct, optimized with FP8 weight and activation quantization, suitable for assistant-like chat scenarios.

Model Features

FP8 Quantization
Weights and activations quantized to FP8 data type, reducing disk size and GPU memory requirements by approximately 50%.
Efficient Inference
Compatible with vLLM >= 0.5.2 for efficient inference, optimizing inference speed.
High Accuracy
Excellent performance on the HumanEval+ benchmark, with accuracy comparable to the non-quantized model.

Model Capabilities

Text Generation
Code Generation
Chat Assistant

Use Cases

Commercial and Research
Code Generation Assistant
Helps developers generate code snippets, improving development efficiency.
Achieved an average score of 79.60 on the HumanEval+ benchmark.
Chatbot
Suitable for assistant-like chat scenarios, providing natural language interaction.
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
Š 2025AIbase