Tinyllama 1.1B 32k
A 32k-context fine-tuned version based on TinyLlama-1.1B, achieving long-context processing capability by increasing rope theta
Downloads 51
Release Time : 12/29/2023
Model Overview
This is a language model optimized for long contexts, supporting 32k context length by adjusting rope frequency base, suitable for use as a speculative decoding model
Model Features
Long-context support
Supports 32k context length by adjusting rope theta parameter
Efficient inference
Quantized version can run on a single A6000 GPU, suitable for speculative decoding
Optimized pretraining
Pretrained on RedPajama-Data-1T-Sample dataset with 32k context length
Model Capabilities
Long text generation
Code generation
Text understanding
Use Cases
Code generation
Programming assistance
Used for generating and completing code
HumanEval evaluation shows Pass@1 reaching 0.0829
Long text processing
Long document analysis
Processes text content up to 32k tokens
Perplexity of 7.1338 at 32768 length
Featured Recommended AI Models