G

Gpt Neo 125m

Developed by EleutherAI
GPT-Neo 125M is a Transformer model based on the GPT-3 architecture, developed by EleutherAI, with 125 million parameters, primarily used for English text generation tasks.
Downloads 150.96k
Release Time : 3/2/2022

Model Overview

GPT-Neo 125M is an autoregressive language model designed based on the Transformer architecture, mainly used for generating text from prompts. It was pre-trained on the large-scale dataset The Pile, enabling it to learn internal representations of the English language.

Model Features

Based on GPT-3 architecture
Replicates the architectural design of GPT-3, offering similar text generation capabilities.
Large-scale pre-training
Trained on 300 billion tokens from The Pile dataset, learning rich language representations.
Open-source model
Developed and open-sourced by EleutherAI, available for research and commercial use.

Model Capabilities

Text generation
Language modeling
Text completion

Use Cases

Text generation
Creative writing
Generates creative texts such as stories and poems based on prompts.
Content completion
Automatically completes subsequent content based on partial text input.
Research
Language model research
Used to study the behavior and performance of autoregressive language models.
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
Š 2025AIbase