D

Distilbert Base Nepali

Developed by Sakonii
This is a DistilBERT model pretrained on Nepali text, specifically optimized for downstream tasks in Nepali.
Downloads 109
Release Time : 3/2/2022

Model Overview

The model was pretrained on a dataset containing over 13 million Nepali text sequences using the Masked Language Modeling (MLM) objective, suitable for Nepali text processing tasks.

Model Features

Nepali Optimization
Specifically trained for Nepali using a dataset of 13 million text sequences.
Efficient Architecture
Utilizes the DistilBERT architecture to reduce model size and computational requirements while maintaining performance.
Custom Tokenizer
Uses a SentencePiece Model (SPM) for text tokenization with a vocabulary size of 24576.

Model Capabilities

Nepali Text Understanding
Masked Language Prediction
Downstream Task Fine-tuning

Use Cases

Text Processing
Text Completion
Predicts masked tokens in text
Achieved a perplexity of 10.479 on the evaluation set
Downstream Task Fine-tuning
Can be used for sequence classification, token classification, or question answering tasks
Featured Recommended AI Models
┬й 2025AIbase