Zero Shot Implicit Bi Encoder
Z
Zero Shot Implicit Bi Encoder
Developed by claritylab
A zero-shot text classification model based on sentence-transformers, achieving text classification without labeled data through implicit training
Downloads 31
Release Time : 5/15/2023
Model Overview
This model is specifically designed for zero-shot text classification, trained implicitly using the aspect-normalized UTCD dataset within a dual-encoder classification framework.
Model Features
Zero-shot learning capability
Capable of classification without task-specific labeled data
Implicit training
Trained implicitly using the aspect-normalized UTCD dataset
Dual-encoder framework
Employs a dual-encoder architecture to enhance classification performance
Model Capabilities
Zero-shot text classification
Intent recognition
Semantic similarity calculation
Use Cases
Natural language processing
Intent recognition
Identifying the underlying intent in user statements
Accurately recognized the 'play music' intent in the example
Text classification
Classifying text without labeled data
Featured Recommended AI Models
Š 2025AIbase