🚀 bias_identificaiton45
This model is a fine - tuned version for bias identification in large language models, trained to classify text into 10 bias categories.
🚀 Quick Start
You can load the model using the Hugging Face transformers
library as follows:
from transformers import AutoTokenizer, AutoModelForSequenceClassification
tokenizer = AutoTokenizer.from_pretrained("PriyaPatel/bias_identificaiton45")
model = AutoModelForSequenceClassification.from_pretrained("PriyaPatel/bias_identificaiton45")
inputs = tokenizer("Your text here", return_tensors="tf")
outputs = model(**inputs)
✨ Features
- Bias Detection: Identifying and categorizing bias types in sentences or text fragments.
- Research: Analyzing and understanding biases in natural language processing models.
📚 Documentation
Model description
This model is a fine - tuned version of the cardiffnlp/twitter-roberta-base-sentiment-latest
on a custom dataset for bias identification in large language models. It is trained to classify input text into one of 10 bias categories.
Intended uses & limitations
Intended Uses:
- Bias Detection: Identifying and categorizing bias types in sentences or text fragments.
- Research: Analyzing and understanding biases in natural language processing models.
Limitations:
- Domain Specificity: The model's performance is optimized for detecting biases within the domains represented in the training data.
- Not for General Sentiment Analysis: This model is not designed for general sentiment analysis or other NLP tasks.
Dataset Used for Training
This dataset was compiled to analyze various types of stereotypical biases present in language models. It incorporates data from multiple publicly available datasets, each contributing to the identification of specific bias types.
Link of the dataset: PriyaPatel/Bias_identification
The biases are labeled as follows:
- Race/Color -
0
- Socioeconomic Status -
1
- Gender -
2
- Disability -
3
- Nationality -
4
- Sexual Orientation -
5
- Physical Appearance -
6
- Religion -
7
- Age -
8
- Profession -
9
Training procedure
- Base Model:
cardiffnlp/twitter-roberta-base-sentiment-latest
- Optimizer: Adam with a learning rate of 0.00001
- Loss Function: Sparse Categorical Crossentropy
- Batch Size: 20
- Epochs: 3
Training hyperparameters
- Learning Rate: 0.00001
- Optimizer: Adam
- Loss Function: Sparse Categorical Crossentropy
- Batch Size: 20
- Epochs: 3
Training Results
- Validation Loss: 0.0744
- Validation Accuracy: 0.9825
- Test Loss: 0.0715
- Test Accuracy: 0.9832
Property |
Details |
Model Type |
Fine - tuned text - classification model based on cardiffnlp/twitter-roberta-base-sentiment-latest |
Training Data |
PriyaPatel/Bias_identification |
Base Model |
cardiffnlp/twitter-roberta-base-sentiment-latest |
Optimizer |
Adam |
Learning Rate |
0.00001 |
Loss Function |
Sparse Categorical Crossentropy |
Batch Size |
20 |
Epochs |
3 |
Validation Loss |
0.0744 |
Validation Accuracy |
0.9825 |
Test Loss |
0.0715 |
Test Accuracy |
0.9832 |