K

Ko Sroberta Multitask

Developed by jhgan
This is a Korean sentence embedding model based on sentence-transformers, capable of mapping sentences and paragraphs into a 768-dimensional dense vector space, suitable for tasks such as clustering or semantic search.
Downloads 162.23k
Release Time : 3/2/2022

Model Overview

This model is based on the RoBERTa architecture, trained with multi-task learning, specifically designed for Korean sentence embedding representation, supporting sentence similarity calculation and feature extraction.

Model Features

Multi-task learning
The model is trained using KorSTS and KorNLI datasets for multi-task learning, improving the quality of sentence embeddings.
Efficient semantic representation
Capable of efficiently mapping sentences and paragraphs into a 768-dimensional dense vector space while preserving semantic information.
Korean optimization
Specially optimized for Korean, suitable for Korean sentence embedding and similarity calculation.

Model Capabilities

Sentence embedding
Semantic search
Text clustering
Sentence similarity calculation

Use Cases

Natural Language Processing
Semantic search
Use sentence embeddings for efficient semantic search to find documents or paragraphs semantically similar to the query sentence.
Text clustering
Cluster large amounts of Korean text into groups with similar semantics for text classification or information organization.
Featured Recommended AI Models
ยฉ 2025AIbase