D

Distilroberta Base Finetuned Wikitext2

Developed by lamyae
This model is a fine-tuned version of distilroberta-base on the wikitext2 dataset, primarily used for text generation tasks.
Downloads 79
Release Time : 4/21/2022

Model Overview

This is a fine-tuned DistilRoBERTa model suitable for text generation and related tasks.

Model Features

Efficient Fine-tuning
Fine-tuned based on the DistilRoBERTa base model, retaining the efficiency of the original model while adapting to specific tasks.
Lightweight Architecture
Utilizes the DistilRoBERTa architecture, which is more lightweight compared to the full RoBERTa model while maintaining good performance.

Model Capabilities

Text Generation
Language Model Fine-tuning

Use Cases

Text Generation
Content Continuation
Generate coherent subsequent content based on given text fragments
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
Š 2025AIbase