Saute
S
Saute
Developed by JustinDuc
SAUTE is a lightweight Transformer architecture with speaker perception ability, designed for effectively modeling multi-speaker dialogues.
Downloads 229
Release Time : 6/9/2025
Model Overview
SAUTE combines EDU-level utterance embeddings, speaker-sensitive memory, and an efficient linear attention mechanism to encode rich dialogue contexts with minimal overhead, suitable for multi-turn dialogues, multi-speaker interactions, and long-distance dialogue dependencies.
Model Features
Speaker-aware memory
Represents the dialogue context for each speaker in a structured way
Linear attention mechanism
Efficient and scalable to long dialogues, avoiding the quadratic cost of full self-attention mechanism
Compatible with pre-trained Transformer
Can be connected to a frozen or fine-tuned BERT model
Lightweight design
Fewer parameters but better performance than traditional multi-layer Transformer
Model Capabilities
Multi-speaker dialogue modeling
Capturing long-distance dialogue dependencies
Masked language modeling
Generating utterance-level embeddings
Use Cases
Dialogue system
Multi-turn dialogue understanding
Track the context of different speakers in complex dialogues
Significant improvement in MLM accuracy on the SODA dataset
Meeting record analysis
Identify and distinguish the speech content of multiple participants
Featured Recommended AI Models
ยฉ 2025AIbase