L

Llama 3 8B Instruct 64k

Developed by MaziyarPanahi
An 8B parameter large language model developed based on winglian/Llama-3-8b-64k-PoSE, using PoSE technology to extend context length to 64k and optimized with DPO fine-tuning
Downloads 91
Release Time : 4/25/2024

Model Overview

This is an 8B parameter large language model based on the Meta Llama-3 architecture, extending context length to 64k via PoSE technology and optimized with DPO fine-tuning, suitable for long-text generation and dialogue tasks.

Model Features

64k long context support
Uses PoSE technology to extend context length from 8k to 64k, suitable for processing long documents and complex dialogues
DPO fine-tuning optimization
Fine-tuned with Intel/orca_dpo_pairs dataset using DPO to improve response quality
Efficient inference
Supports flash_attention_2 and bfloat16 inference for improved efficiency

Model Capabilities

Long text generation
Dialogue systems
Instruction following

Use Cases

Dialogue systems
Role-playing chatbot
Can be used to build chatbots with specific character traits, such as the pirate chatbot in the example
Capable of generating coherent dialogues that align with character settings
Long document processing
Long document summarization
Leverages 64k context length advantage to process long documents and generate summaries
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
Š 2025AIbase