L

Longalpaca 70B

Developed by Yukang
LongLoRA is an efficient fine-tuning technique for large language models with long context processing capabilities, achieving this through shifted short attention mechanisms, supporting context lengths from 8k to 100k.
Downloads 1,293
Release Time : 10/9/2023

Model Overview

LongLoRA proposes an efficient fine-tuning solution to extend the context processing capabilities of large language models, supporting models of various scales from 7B to 70B.

Model Features

Shifted Short Attention Mechanism
Innovative attention mechanism design, compatible with Flash-Attention technology, not required during inference
Long Context Support
Supports context lengths from 8k to 100k, covering various application scenarios
Full Series Models
Open-sourced models from 7B to 70B, meeting different computational resource needs
Dedicated Dataset
Constructed the LongAlpaca-12k long-context instruction dataset to enhance model instruction-following capabilities

Model Capabilities

Long Text Understanding
Long Text Generation
Instruction Following
Question Answering Systems

Use Cases

Document Processing
Long Document Summarization
Process documents up to tens of thousands of tokens and generate summaries
Effectively retains key information from long documents
Technical Document Analysis
Read and understand complex technical documents
Accurately answers detailed questions about document content
Academic Research
Paper Reading and Q&A
Read complete academic papers and answer related questions
Understands complex concepts and argumentation processes in papers
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
Š 2025AIbase