M

Mrt5 Small

Developed by stanfordnlp
MrT5 is an efficient improved version of ByT5. It shortens the input sequence length through dynamic token merging technology to improve processing efficiency.
Downloads 65
Release Time : 3/23/2025

Model Overview

MrT5 is an improved model based on ByT5. By introducing a dynamic token deletion mechanism, it shortens the sequence length in the encoder while retaining key information, making it suitable for multilingual sequence-to-sequence tasks.

Model Features

Dynamic token merging
Dynamically shorten the input sequence length through the deletion gate mechanism to improve processing efficiency.
Multilingual support
Supports the processing of 15 typologically diverse languages.
Efficient improvement
Implement the deletion gate mechanism by only adding 3000 parameters on the basis of ByT5.

Model Capabilities

Multilingual text generation
Sequence-to-sequence conversion
Byte-level processing

Use Cases

Academic research
Research on language model efficiency
Study the impact of dynamic token merging technology on language model efficiency.
Text processing
Multilingual text summarization
Generate summaries for multilingual texts.
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
© 2025AIbase