L

Llmlingua 2 Bert Base Multilingual Cased Meetingbank

Developed by microsoft
A prompt compression token classification model fine-tuned on the multilingual BERT base model, designed for task-agnostic prompt compression
Downloads 28.67k
Release Time : 3/17/2024

Model Overview

This model performs task-agnostic prompt compression token classification, where the retention probability of each token serves as a compression metric. Particularly suitable for compressing texts like meeting minutes.

Model Features

Task-Agnostic Prompt Compression
Capable of effective prompt compression without relying on specific downstream tasks
Multilingual Support
Based on a multilingual BERT model, supporting text compression in multiple languages
Data Distillation Training
Trained using the data distillation method proposed by LLMLingua-2, improving compression quality

Model Capabilities

Text Compression
Token Classification
Meeting Minutes Processing
Multilingual Text Processing

Use Cases

Meeting Minutes Processing
Meeting Minutes Compression
Compress lengthy meeting minutes while retaining key information
Significantly reduces text length while maintaining critical information
Downstream Task Preprocessing
Preprocess input text for downstream tasks like Q&A and summarization
Enhances downstream task efficiency without significantly affecting accuracy
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
Š 2025AIbase