L

Language Perceiver

Developed by deepmind
Pre-trained on BERT-style masked language modeling tasks, supports multimodal Transformer model processing UTF-8 byte inputs
Downloads 9,840
Release Time : 3/2/2022

Model Overview

Perceiver IO is a universal Transformer architecture applicable to multiple modalities such as text, images, and audio. This language model is trained directly on raw bytes without requiring a tokenizer and supports masked language modeling tasks.

Model Features

Byte-level Input Processing
Directly processes UTF-8 bytes without requiring a tokenizer or fixed vocabulary
Multimodal Architecture
Base architecture can be extended to other modalities like images and audio
Efficient Attention Mechanism
Achieves input-size-independent computational complexity through latent vectors

Model Capabilities

Text Feature Extraction
Masked Word Prediction
Downstream Task Fine-tuning

Use Cases

Natural Language Processing
Text Completion
Predicts masked portions of text
Successfully predicted masked words in 'missing portions' in examples
Text Classification
Can be fine-tuned for classification tasks like sentiment analysis
Average GLUE benchmark score of 81.8
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
Š 2025AIbase