🚀 mT5-m2o-russian-CrossSum
本倉庫包含在CrossSum數據集的所有跨語言對上微調的多對一(m2o)mT5檢查點,其中目標摘要為俄語,即該模型嘗試將任何語言的文本總結為俄語。有關微調細節和腳本,請參閱論文和官方倉庫。
🚀 快速開始
✨ 主要特性
- 支持多種語言:該模型支持阿姆哈拉語(am)、阿拉伯語(ar)、阿塞拜疆語(az)、孟加拉語(bn)、緬甸語(my)、中文(zh)、英語(en)、法語(fr)等眾多語言。
- 多對一總結:能夠將多種語言的文本總結為俄語。
📦 安裝指南
文檔未提及安裝步驟,暫不提供相關內容。
💻 使用示例
基礎用法
import re
from transformers import AutoTokenizer, AutoModelForSeq2SeqLM
WHITESPACE_HANDLER = lambda k: re.sub('\s+', ' ', re.sub('\n+', ' ', k.strip()))
article_text = """Videos that say approved vaccines are dangerous and cause autism, cancer or infertility are among those that will be taken down, the company said. The policy includes the termination of accounts of anti-vaccine influencers. Tech giants have been criticised for not doing more to counter false health information on their sites. In July, US President Joe Biden said social media platforms were largely responsible for people's scepticism in getting vaccinated by spreading misinformation, and appealed for them to address the issue. YouTube, which is owned by Google, said 130,000 videos were removed from its platform since last year, when it implemented a ban on content spreading misinformation about Covid vaccines. In a blog post, the company said it had seen false claims about Covid jabs "spill over into misinformation about vaccines in general". The new policy covers long-approved vaccines, such as those against measles or hepatitis B. "We're expanding our medical misinformation policies on YouTube with new guidelines on currently administered vaccines that are approved and confirmed to be safe and effective by local health authorities and the WHO," the post said, referring to the World Health Organization."""
model_name = "csebuetnlp/mT5_m2o_russian_crossSum"
tokenizer = AutoTokenizer.from_pretrained(model_name)
model = AutoModelForSeq2SeqLM.from_pretrained(model_name)
input_ids = tokenizer(
[WHITESPACE_HANDLER(article_text)],
return_tensors="pt",
padding="max_length",
truncation=True,
max_length=512
)["input_ids"]
output_ids = model.generate(
input_ids=input_ids,
max_length=84,
no_repeat_ngram_size=2,
num_beams=4
)[0]
summary = tokenizer.decode(
output_ids,
skip_special_tokens=True,
clean_up_tokenization_spaces=False
)
print(summary)
📚 詳細文檔
文檔未提及詳細說明,暫不提供相關內容。
🔧 技術細節
文檔未提及技術實現細節,暫不提供相關內容。
📄 許可證
本項目採用CC BY-NC-SA 4.0許可證。
📚 引用
如果您使用此模型,請引用以下論文:
@article{hasan2021crosssum,
author = {Tahmid Hasan and Abhik Bhattacharjee and Wasi Uddin Ahmad and Yuan-Fang Li and Yong-bin Kang and Rifat Shahriyar},
title = {CrossSum: Beyond English-Centric Cross-Lingual Abstractive Text Summarization for 1500+ Language Pairs},
journal = {CoRR},
volume = {abs/2112.08804},
year = {2021},
url = {https://arxiv.org/abs/2112.08804},
eprinttype = {arXiv},
eprint = {2112.08804}
}
📋 模型信息
屬性 |
詳情 |
模型類型 |
mT5多對一(m2o)微調模型 |
訓練數據 |
CrossSum數據集的所有跨語言對 |