L

Llmc Gpt2 774M 150B

Developed by mdouglas
This is a 774M-parameter language model based on the GPT-2 architecture, trained on 150 billion tokens from the FineWeb dataset.
Downloads 18
Release Time : 6/12/2024

Model Overview

This model is an implementation of the GPT-2 architecture, focusing on English text generation and comprehension tasks.

Model Features

Efficient training
Training completed in approximately 6 days on 8 A100 80GB SXM GPUs.
Large-scale data training
Trained on 150 billion tokens from the FineWeb dataset.
Lightweight implementation
Efficient implementation based on llm.c.

Model Capabilities

English text generation
Language understanding
Contextual learning

Use Cases

Text generation
Content creation
Generate articles, stories, or other creative texts
Dialogue systems
Build chatbots or virtual assistants
Education
Language learning
Serve as an auxiliary tool for English learning
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
Š 2025AIbase