đ Burmese-GPT
Burmese-GPT is a specialized large language model for the Burmese language, fine-tuned on the GPT-2 architecture. It serves as a foundational base for various NLP tasks in Burmese.
đ Quick Start
Installation
!pip install transformers
Loading the Model
from transformers import AutoTokenizer, AutoModelForCausalLM
tokenizer = AutoTokenizer.from_pretrained("WYNN747/Burmese-GPT")
model = AutoModelForCausalLM.from_pretrained("WYNN747/Burmese-GPT")
input_text = "ááŽá¸ááŊááēá¸ááŊá˛ááąáŦáēáááē ááŽ"
input_ids = tokenizer.encode(input_text, return_tensors='pt')
output = model.generate(input_ids, max_length=50)
print(tokenizer.decode(output[0], skip_special_tokens=True))
⨠Features
- Specialized for the Burmese language, based on the GPT-2 architecture (mGPT XL model).
- Primarily designed for text completion in Burmese, and can be a foundation for various NLP tasks.
đ Documentation
Intended Use
This model, primarily designed for text completion in Burmese, serves as a foundational tool for a variety of NLP tasks. While its current primary function is to assist in generating and completing text, it holds significant potential for further applications. Researchers and developers can fine-tune this model on specialized datasets to extend its capabilities to other NLP applications, such as summarization and instruction-based tasks. It is important to note, however, that for high-stakes decisions or understanding domain-specific jargon, additional specialized training of the model is recommended to ensure accuracy and reliability.
Training Data
Burmese-GPT was trained on a comprehensive dataset of Burmese texts, curated by the author. This dataset, which includes literature, news, online articles, and content from Burmese Wikipedia, has been meticulously compiled to ensure a wide representation of the linguistic diversity and styles found in the Burmese language. The dataset, created by the author, is available for academic and research purposes upon request. Interested parties should contact the author to gain access to this valuable resource.
Ethical Considerations
Users should be aware of the inherent limitations and biases of language models. This model should be used responsibly, especially in sensitive applications, and is not intended for generating misleading or harmful content.
Limitations
The Burmese GPT performs well with general Burmese text but may not be as effective with highly technical or niche content. Users are advised to conduct thorough testing for their specific use cases.
đ License
This project is licensed under the MIT license.
đ Contact Information
đ Acknowledgements
Credit and thanks to the creators of the mGPT-XL model for providing the foundational model. Their contributions have been instrumental in the development of the Burmese GPT.
â Frequently Asked Questions (FAQ) (In Burmese)
1) Is Burmese GPT the Burmese Chat-GPT?
- Burmese GPT is not a chat application designed for question - answering.
- It is a language model for text completion, which can continue writing based on the given text.
2) What is the purpose of the Burmese GPT (Text completion) model?
- It aims to build applications that can handle Burmese - related questions and topics, allowing users to ask questions in Burmese.
- Based on this model, applications such as Burmese Chat - GPT for question - answering, applications for summarizing Burmese text, and applications for writing Burmese articles and poems can be developed.
3) Why doesn't the full text appear when testing on the linked platform?
- The Hugging Face platform has a limit on the number of words it can display, so the full text may not appear.
- If the generated text is not complete, click the "Compute" button again. You can also test by calling the API if you want to see the full text.
4) What kind of data is used to train Burmese GPT?
- Burmese GPT is trained with data from the open - accessible Myanmar Wikipedia and open Myanmar databases, so it can understand most Burmese words.
- Since it does not use the intellectual property of writers and artists such as books, song lyrics, and information, when doing text completion related to them, it may not be accurate and may only generate AI - invented content.
- (Artists who want to contribute their works and information to Burmese GPT can contact the author.)
5) What is needed if I want to use the dataset used in Burmese GPT?
- The dataset containing 15K Burmese texts (corpus) is intended for those working on academic, research, or open - community projects. You can contact the author with information about your project, paper, or thesis.