đ EconoBert
This model, EconoBert, is a fine - tuned version of the pre - trained bert - base - uncased
model. It is specifically fine - tuned on an economics - related dataset, aiming to provide better performance for NLP tasks in the fields of economics, politics, and finance.
đ Quick Start
This model can be used as a backbone for NLP tasks in the domains of economics, politics, and finance. You can load it using relevant NLP libraries and start fine - tuning or making predictions.
⨠Features
- Fine - tuned on Economics Data: It is fine - tuned on a dataset specific to the economics domain, making it more suitable for related NLP tasks.
- Good Performance: Achieves an accuracy of 73% for the MLM task and 95% for the NSP task on the test set.
đĻ Installation
No specific installation steps are provided in the original document.
đģ Usage Examples
No code examples are provided in the original document.
đ Documentation
Model description
The model is a simple fine - tuning of a base BERT model on a dataset specific to the domain of economics. It follows the same architecture as the base model, and no resize_token_embeddings
operation was required.
Intended uses & limitations
This model should be used as a backbone for NLP tasks applied to the domains of economics, politics, and finance.
Training and evaluation data
The dataset used for fine - tuning is samchain/BIS_Speeches_97_23. It consists of 773k pairs of sentences, with half being negative pairs (sequence A and B are not related) and the other half being positive pairs (sequence B follows sequence A). The test set is made up of 136k pairs.
Training procedure
The model was fine - tuned for 2 epochs, with a batch size of 64 and a sequence length of 128. The Adam optimizer was used with a learning rate of 1e - 5.
Training hyperparameters
Property |
Details |
Optimizer |
{'name': 'Adam', 'weight_decay': None, 'clipnorm': None, 'global_clipnorm': None, 'clipvalue': None, 'use_ema': False, 'ema_momentum': 0.99, 'ema_overwrite_frequency': None, 'jit_compile': True, 'is_legacy_optimizer': False, 'learning_rate': 1e - 05, 'beta_1': 0.9, 'beta_2': 0.999, 'epsilon': 1e - 07, 'amsgrad': False} |
Training Precision |
float32 |
Training results
The training loss is 1.6046 on the training set and 1.47 on the test set.
Framework versions
Property |
Details |
Transformers |
4.31.0 |
TensorFlow |
2.12.0 |
Datasets |
2.13.1 |
Tokenizers |
0.13.3 |
đ§ Technical Details
The model is a fine - tuned version of bert - base - uncased
. It uses the same architecture as the base model and is trained on a specific economics - related dataset. The training process involves fine - tuning for 2 epochs with specific hyperparameters such as batch size, sequence length, and optimizer settings.
đ License
This model is released under the Apache 2.0 license.
Citing & Authors
Samuel Chaineau