V

Vicuna 68m

Developed by double7
This is a 68-million-parameter Vicuna-like model, fine-tuned on ShareGPT data based on LLaMA-68m, primarily used for speculative model development.
Downloads 1,935
Release Time : 1/11/2024

Model Overview

This model is a small language model based on the LLaMA-68m architecture, fine-tuned on ShareGPT conversational data, designed to maintain alignment with LLaMA and Vicuna models, making it particularly suitable as a speculative model.

Model Features

Small and efficient
Only 68 million parameters, with low computational resource requirements, suitable for use as a speculative model
Dual alignment
Maintains alignment with both LLaMA and Vicuna models
Conversation optimization
Fine-tuned on ShareGPT conversational data, suitable for dialogue generation tasks

Model Capabilities

Text generation
Dialogue generation
Language understanding

Use Cases

Speculative decoding
Base model in the MCSD paper
Used as a small speculative model in conjunction with a large target model
Compared to the original LLaMA-68m, it achieves better alignment with Vicuna models
Dialogue systems
Lightweight dialogue agent
Suitable for dialogue generation in resource-constrained environments
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
Š 2025AIbase