MEETING SUMMARY BART LARGE XSUM SAMSUM DIALOGSUM
M
MEETING SUMMARY BART LARGE XSUM SAMSUM DIALOGSUM
由knkarthick開發
基於BART-large架構的序列到序列模型,專門針對會議摘要任務進行微調,支持多數據集訓練
下載量 123
發布時間 : 3/2/2022
模型概述
該模型通過微調facebook/bart-large-xsum獲得,專門用於生成抽象文本摘要,特別擅長處理對話和會議場景的摘要任務
模型特點
多數據集訓練
在cnndaily/newyorkdaily/xsum/samsum/dialogsum等多個摘要數據集上進行微調
對話場景優化
特別針對會議記錄和對話摘要場景進行優化
抽象摘要能力
能夠生成非簡單抽取式的、語義連貫的摘要
模型能力
文本摘要生成
對話內容壓縮
會議記錄精簡
使用案例
會議記錄處理
自動會議紀要生成
將冗長的會議對話自動生成簡潔的要點摘要
生成符合人類寫作習慣的會議紀要
新聞摘要
新聞內容精簡
從長篇新聞報道中提取核心內容
生成保留關鍵信息的簡短新聞摘要
🚀 會議摘要BART大模型
本項目基於BART架構,通過微調facebook/bart-large-xsum
模型得到,用於文本摘要任務。它可以處理多種類型的文本,如會議記錄、新聞文章等,為用戶提供高效準確的摘要生成服務。
🚀 快速開始
本模型是通過微調facebook/bart-large-xsum
得到的。以下是使用本模型進行文本摘要的示例代碼:
💻 使用示例
基礎用法
from transformers import pipeline
summarizer = pipeline("summarization", model="knkarthick/MEETING-SUMMARY-BART-LARGE-XSUM-SAMSUM-DIALOGSUM")
text = '''The tower is 324 metres (1,063 ft) tall, about the same height as an 81-storey building, and the tallest structure in Paris. Its base is square, measuring 125 metres (410 ft) on each side. During its construction, the Eiffel Tower surpassed the Washington Monument to become the tallest man-made structure in the world, a title it held for 41 years until the Chrysler Building in New York City was finished in 1930. It was the first structure to reach a height of 300 metres. Due to the addition of a broadcasting aerial at the top of the tower in 1957, it is now taller than the Chrysler Building by 5.2 metres (17 ft). Excluding transmitters, the Eiffel Tower is the second tallest free-standing structure in France after the Millau Viaduct.
'''
summarizer(text)
高級用法
from transformers import pipeline
summarizer = pipeline("summarization", model="knkarthick/MEETING-SUMMARY-BART-LARGE-XSUM-SAMSUM-DIALOGSUM")
text = '''Bangalore is the capital and the largest city of the Indian state of Karnataka. It has a population of more than 8 million and a metropolitan population of around 11 million, making it the third most populous city and fifth most populous urban agglomeration in India. Located in southern India on the Deccan Plateau, at a height of over 900 m (3,000 ft) above sea level, Bangalore is known for its pleasant climate throughout the year. Its elevation is the highest among the major cities of India.The city's history dates back to around 890 CE, in a stone inscription found at the Nageshwara Temple in Begur, Bangalore. The Begur inscription is written in Halegannada (ancient Kannada), mentions 'Bengaluru Kalaga' (battle of Bengaluru). It was a significant turning point in the history of Bangalore as it bears the earliest reference to the name 'Bengaluru'. In 1537 CE, Kempé Gowdā – a feudal ruler under the Vijayanagara Empire – established a mud fort considered to be the foundation of modern Bangalore and its oldest areas, or petes, which exist to the present day.
After the fall of Vijayanagar empire in 16th century, the Mughals sold Bangalore to Chikkadevaraja Wodeyar (1673–1704), the then ruler of the Kingdom of Mysore for three lakh rupees. When Haider Ali seized control of the Kingdom of Mysore, the administration of Bangalore passed into his hands.
The city was captured by the British East India Company after victory in the Fourth Anglo-Mysore War (1799), who returned administrative control of the city to the Maharaja of Mysore. The old city developed in the dominions of the Maharaja of Mysore and was made capital of the Princely State of Mysore, which existed as a nominally sovereign entity of the British Raj. In 1809, the British shifted their cantonment to Bangalore, outside the old city, and a town grew up around it, which was governed as part of British India. Following India's independence in 1947, Bangalore became the capital of Mysore State, and remained capital when the new Indian state of Karnataka was formed in 1956. The two urban settlements of Bangalore – city and cantonment – which had developed as independent entities merged into a single urban centre in 1949. The existing Kannada name, Bengalūru, was declared the official name of the city in 2006.
Bangalore is widely regarded as the "Silicon Valley of India" (or "IT capital of India") because of its role as the nation's leading information technology (IT) exporter. Indian technological organisations are headquartered in the city. A demographically diverse city, Bangalore is the second fastest-growing major metropolis in India. Recent estimates of the metro economy of its urban area have ranked Bangalore either the fourth- or fifth-most productive metro area of India. As of 2017, Bangalore was home to 7,700 millionaires and 8 billionaires with a total wealth of $320 billion. It is home to many educational and research institutions. Numerous state-owned aerospace and defence organisations are located in the city. The city also houses the Kannada film industry. It was ranked the most liveable Indian city with a population of over a million under the Ease of Living Index 2020.
'''
summarizer(text)
📚 詳細文檔
模型信息
屬性 | 詳情 |
---|---|
模型類型 | BART大模型用於會議摘要 |
訓練數據 | cnndaily/newyorkdaily/xsum/samsum/dialogsum |
評估指標 | rouge |
模型結果
任務 | 數據集 | 評估指標 | 值 |
---|---|---|---|
抽象文本摘要 | cnndaily/newyorkdaily/xsum/samsum/dialogsum | Validation ROGUE - 1 | NA |
抽象文本摘要 | cnndaily/newyorkdaily/xsum/samsum/dialogsum | Validation ROGUE - 2 | NA |
抽象文本摘要 | cnndaily/newyorkdaily/xsum/samsum/dialogsum | Validation ROGUE - L | NA |
抽象文本摘要 | cnndaily/newyorkdaily/xsum/samsum/dialogsum | Validation ROGUE - Lsum | NA |
抽象文本摘要 | cnndaily/newyorkdaily/xsum/samsum/dialogsum | Test ROGUE - 1 | NA |
抽象文本摘要 | cnndaily/newyorkdaily/xsum/samsum/dialogsum | Test ROGUE - 2 | NA |
抽象文本摘要 | cnndaily/newyorkdaily/xsum/samsum/dialogsum | Test ROGUE - L | NA |
抽象文本摘要 | cnndaily/newyorkdaily/xsum/samsum/dialogsum | Test ROGUE - Lsum | NA |
📄 許可證
本項目採用Apache 2.0許可證。
Bart Large Cnn
MIT
基於英語語料預訓練的BART模型,專門針對CNN每日郵報數據集進行微調,適用於文本摘要任務
文本生成 英語
B
facebook
3.8M
1,364
Parrot Paraphraser On T5
Parrot是一個基於T5的釋義框架,專為加速訓練自然語言理解(NLU)模型而設計,通過生成高質量釋義實現數據增強。
文本生成
Transformers

P
prithivida
910.07k
152
Distilbart Cnn 12 6
Apache-2.0
DistilBART是BART模型的蒸餾版本,專門針對文本摘要任務進行了優化,在保持較高性能的同時顯著提升了推理速度。
文本生成 英語
D
sshleifer
783.96k
278
T5 Base Summarization Claim Extractor
基於T5架構的模型,專門用於從摘要文本中提取原子聲明,是摘要事實性評估流程的關鍵組件。
文本生成
Transformers 英語

T
Babelscape
666.36k
9
Unieval Sum
UniEval是一個統一的多維評估器,用於自然語言生成任務的自動評估,支持多個可解釋維度的評估。
文本生成
Transformers

U
MingZhong
318.08k
3
Pegasus Paraphrase
Apache-2.0
基於PEGASUS架構微調的文本複述模型,能夠生成語義相同但表達不同的句子。
文本生成
Transformers 英語

P
tuner007
209.03k
185
T5 Base Korean Summarization
這是一個基於T5架構的韓語文本摘要模型,專為韓語文本摘要任務設計,通過微調paust/pko-t5-base模型在多個韓語數據集上訓練而成。
文本生成
Transformers 韓語

T
eenzeenee
148.32k
25
Pegasus Xsum
PEGASUS是一種基於Transformer的預訓練模型,專門用於抽象文本摘要任務。
文本生成 英語
P
google
144.72k
198
Bart Large Cnn Samsum
MIT
基於BART-large架構的對話摘要模型,專為SAMSum語料庫微調,適用於生成對話摘要。
文本生成
Transformers 英語

B
philschmid
141.28k
258
Kobart Summarization
MIT
基於KoBART架構的韓語文本摘要模型,能夠生成韓語新聞文章的簡潔摘要。
文本生成
Transformers 韓語

K
gogamza
119.18k
12
精選推薦AI模型
Llama 3 Typhoon V1.5x 8b Instruct
專為泰語設計的80億參數指令模型,性能媲美GPT-3.5-turbo,優化了應用場景、檢索增強生成、受限生成和推理任務
大型語言模型
Transformers 支持多種語言

L
scb10x
3,269
16
Cadet Tiny
Openrail
Cadet-Tiny是一個基於SODA數據集訓練的超小型對話模型,專為邊緣設備推理設計,體積僅為Cosmo-3B模型的2%左右。
對話系統
Transformers 英語

C
ToddGoldfarb
2,691
6
Roberta Base Chinese Extractive Qa
基於RoBERTa架構的中文抽取式問答模型,適用於從給定文本中提取答案的任務。
問答系統 中文
R
uer
2,694
98