F

Finbert Pretrain Yiyanghkust

Developed by philschmid
FinBERT is a BERT model pretrained on financial texts, aiming to advance research and applications in financial natural language processing.
Downloads 580
Release Time : 3/2/2022

Model Overview

FinBERT is a pretrained language model based on the BERT architecture for the financial domain. It was trained on three major financial text corpora (corporate annual reports, earnings call transcripts, and analyst research reports), with a total corpus size of 4.9 billion tokens.

Model Features

Financial Domain Pretraining
Specifically pretrained on financial texts, including professional corpora such as corporate annual reports, earnings call transcripts, and analyst research reports.
Large-Scale Corpus
Training corpus size reaches 4.9 billion tokens, covering various types of financial texts.
Downstream Task Adaptation
Can serve as a base model for fine-tuning on various financial NLP tasks.

Model Capabilities

Financial Text Understanding
Financial Domain Masked Prediction
Financial Text Feature Extraction

Use Cases

Financial Text Analysis
Analyst Sentiment Classification
After fine-tuning, it can be used to analyze sentiment tendencies in financial texts.
Fine-tuned model has been released at https://huggingface.co/yiyanghkust/finbert-tone
Financial Document Information Extraction
Extract key information from documents such as annual reports and financial statements.
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
© 2025AIbase