M

Mdlm Owt

Developed by kuleshov-group
A masked language model based on the diffusion process, similar in scale to GPT2-medium, trained through a forward diffusion process capable of reconstructing original inputs from different masking levels.
Downloads 13.14k
Release Time : 6/6/2024

Model Overview

This model is a masked diffusion language model trained via a forward diffusion process, capable of handling inputs ranging from fully masked to fully unmasked and reconstructing the original input in the process.

Model Features

Diffusion Process Training
Trained through a forward diffusion process, generating different input levels from fully masked to fully unmasked.
Large-Scale Training
Trained for 1 million steps on the OpenWebText corpus, processing 33 billion tokens.
Efficient Reconstruction
Capable of effectively reconstructing original inputs from different masking levels and outputting logical values.

Model Capabilities

Masked Language Modeling
Text Reconstruction

Use Cases

Natural Language Processing
Text Completion
Completing and reconstructing partially masked text.
Language Model Pretraining
Used as a pretrained model for downstream NLP tasks.
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
Š 2025AIbase