đ Model Card for kobart-base-v2
KoBART-base-v2 is a Korean encoder-decoder language model based on the BART architecture, trained on over 40GB of Korean text.
đ Quick Start
Use the code below to get started with the model.
Click to expand
from transformers import PreTrainedTokenizerFast, BartModel
tokenizer = PreTrainedTokenizerFast.from_pretrained('gogamza/kobart-base-v2')
model = BartModel.from_pretrained('gogamza/kobart-base-v2')
⨠Features
- Feature Extraction: This model can be used for the task of feature extraction.
- Korean Language Support: Specifically designed for the Korean language, trained on a large amount of Korean text.
đĻ Installation
No specific installation steps are provided in the original document.
đģ Usage Examples
Basic Usage
from transformers import PreTrainedTokenizerFast, BartModel
tokenizer = PreTrainedTokenizerFast.from_pretrained('gogamza/kobart-base-v2')
model = BartModel.from_pretrained('gogamza/kobart-base-v2')
đ Documentation
Model Details
BART (Bidirectional and Auto-Regressive Transformers) is trained in the form of an autoencoder
by adding noise to parts of the input text and then restoring it to the original text. Korean BART (hereinafter KoBART) is a Korean encoder-decoder
language model trained on over 40GB of Korean text using the Text Infilling
noise function used in the paper. We are releasing the derived KoBART-base
.
- Developed by: More information needed
- Shared by [Optional]: Heewon(Haven) Jeon
- Model type: Feature Extraction
- Language(s) (NLP): Korean
- License: MIT
- Parent Model: BART
- Resources for more information:
Uses
Direct Use
This model can be used for the task of Feature Extraction.
Out-of-Scope Use
The model should not be used to intentionally create hostile or alienating environments for people.
Bias, Risks, and Limitations
Significant research has explored bias and fairness issues with language models (see, e.g., Sheng et al. (2021) and Bender et al. (2021)). Predictions generated by the model may include disturbing and harmful stereotypes across protected classes; identity characteristics; and sensitive, social, and occupational groups.
Recommendations
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
Training Details
Training Data
Data |
# of Sentences |
Korean Wiki |
5M |
Other corpus |
0.27B |
In addition to the Korean Wikipedia, various data such as news, books, Modu Corpus v1.0 (conversations, news, ...), and Blue House Petitions were used for model training.
The vocab
size is 30,000, and emoticons and emojis commonly used in conversations, such as đ, đ, đ, đ
, đ¤Ŗ, .. , :-)
, :)
, -)
, (-:
..., were added to improve the recognition ability of these tokens.
Training Procedure
Tokenizer
It was trained using the Character BPE tokenizer
from the tokenizers
package.
Speeds, Sizes, Times
Model |
# of params |
Type |
# of layers |
# of heads |
ffn_dim |
hidden_dims |
KoBART-base |
124M |
Encoder |
6 |
16 |
3072 |
768 |
|
|
Decoder |
6 |
16 |
3072 |
768 |
Evaluation
Testing Data, Factors & Metrics
More information needed for testing data, factors, and metrics.
Results
NSMC
The model authors also note in the GitHub Repo:
Model Examination
More information needed.
Environmental Impact
Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).
- Hardware Type: More information needed
- Hours used: More information needed
- Cloud Provider: More information needed
- Compute Region: More information needed
- Carbon Emitted: More information needed
Technical Specifications [optional]
More information needed for model architecture, objective, compute infrastructure (hardware and software).
Citation
More information needed for BibTeX citation.
Glossary [optional]
More information needed.
More Information [optional]
More information needed.
Model Card Authors [optional]
Heewon(Haven) Jeon in collaboration with Ezi Ozoani and the Hugging Face team
Model Card Contact
The model authors note in the GitHub Repo:
Please post issues related to KoBART
here.