🚀 mT5-large查詢生成模型
本項目是一個基於mT5-large的查詢生成模型,使用XOR QA數據進行訓練。該模型可用於生成不同語言的查詢問題,在跨語言密集檢索等領域具有重要價值。
🚀 快速開始
本模型是一個使用XOR QA數據訓練的mT5-large查詢生成模型。它被應用於以下兩篇論文的研究中:
💻 使用示例
基礎用法
from transformers import pipeline
lang2mT5 = dict(
ar='Arabic',
bn='Bengali',
fi='Finnish',
ja='Japanese',
ko='Korean',
ru='Russian',
te='Telugu'
)
PROMPT = 'Generate a {lang} question for this passage: {title} {passage}'
title = 'Transformer (machine learning model)'
passage = 'A transformer is a deep learning model that adopts the mechanism of self-attention, differentially ' \
'weighting the significance of each part of the input (which includes the recursive output) data.'
model_name_or_path = 'ielabgroup/xor-tydi-docTquery-mt5-large'
input_text = PROMPT.format_map({'lang': lang2mT5['ja'],
'title': title,
'passage': passage})
generator = pipeline(model=model_name_or_path,
task='text2text-generation',
device="cuda:0",
)
results = generator(input_text,
do_sample=True,
max_length=64,
num_return_sequences=10,
)
for i, result in enumerate(results):
print(f'{i + 1}. {result["generated_text"]}')
📄 許可證
本項目採用Apache-2.0許可證。
📚 引用信息
如果您在研究中使用了本模型,請參考以下BibTeX引用:
@article{zhuang2022bridging,
title={Bridging the gap between indexing and retrieval for differentiable search index with query generation},
author={Zhuang, Shengyao and Ren, Houxing and Shou, Linjun and Pei, Jian and Gong, Ming and Zuccon, Guido and Jiang, Daxin},
journal={arXiv preprint arXiv:2206.10128},
year={2022}
}
@inproceedings{zhuang2023augmenting,
title={Augmenting Passage Representations with Query Generation for Enhanced Cross-Lingual Dense Retrieval},
author={Zhuang, Shengyao and Shou, Linjun and Zuccon, Guido},
booktitle={Proceedings of the 46th international ACM SIGIR conference on research and development in information retrieval},
year={2023}
}