S

Sentence Bert Base Ja Mean Tokens V2

Developed by sonoisa
This is a Japanese-specific Sentence-BERT model, which uses an improved loss function for training optimization compared to Version 1, achieving a 1.5 to 2 percentage point increase in accuracy.
Downloads 108.15k
Release Time : 3/2/2022

Model Overview

Japanese Sentence-BERT model for generating sentence embeddings, suitable for tasks such as sentence similarity calculation and feature extraction.

Model Features

Optimized loss function
Trained using MultipleNegativesRankingLoss, achieving a 1.5-2% accuracy improvement over Version 1
Japanese-specific
Sentence-BERT model specifically optimized for Japanese text
Based on high-quality pre-trained model
Built on cl-tohoku/bert-base-japanese-whole-word-masking

Model Capabilities

Japanese sentence embedding
Sentence similarity calculation
Feature extraction

Use Cases

Text similarity
Semantic search
Implement semantic search by calculating sentence embedding similarity
Duplicate content detection
Identify sentences with similar semantics but different expressions
Information retrieval
Document clustering
Cluster documents based on sentence embeddings
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
© 2025AIbase