Fairseq Dense 2.7B
A converted version of the 2.7 billion parameter dense model based on the paper 'Efficient Large-scale Language Modeling with Mixture of Experts' by Artetxe et al.
Downloads 18
Release Time : 3/2/2022
Model Overview
This model is a Hugging Face transformers-compatible version of the original dense 2.7 billion parameter model, focusing on large-scale language modeling tasks.
Model Features
Large-scale Parameters
Dense model architecture with 2.7 billion parameters, providing powerful language modeling capabilities.
Hugging Face Compatibility
Converted to Hugging Face transformers-compatible format for easy integration and use.
Efficient Modeling
Implements efficient large-scale language modeling using the Mixture of Experts approach.
Model Capabilities
Text Generation
Language Understanding
Contextual Learning
Use Cases
Natural Language Processing
Open-domain Dialogue
Can be used to build open-domain dialogue systems.
Text Completion
Generates coherent text completions based on context.
Featured Recommended AI Models