G

Galactica 120b

Developed by facebook
GALACTICA is a large language model trained on a large-scale scientific corpus and designed specifically for scientific tasks, including citation prediction, scientific question answering, mathematical reasoning, etc.
Downloads 136
Release Time : 11/16/2022

Model Overview

The GALACTICA model was developed by the Papers with Code team at Meta AI, aiming to study the application of language models in the automatic organization of science. The parameter scale of this model series ranges from 125 million to 120 billion, and the 120B version is the largest specification.

Model Features

Scientific-specific training
Trained on 106 billion tokens of open-source scientific text and data, covering professional content such as papers, textbooks, and scientific websites
Multimodal support
Supports the processing of various scientific data types such as citations, mathematical formulas, and molecular structures through special markers
Low toxicity design
Performs better than other large language models in bias and toxicity assessments
Scalable
Provides model versions with different parameter scales from 125M to 120B

Model Capabilities

Scientific text generation
Citation prediction
Mathematical problem solving
Scientific question answering
Document summarization
Molecular property prediction
Scientific entity extraction

Use Cases

Academic research
Literature-assisted writing
Automatically generate academic text with correct citations
Improve the efficiency of academic writing, but manual verification of citation accuracy is required
Scientific question answering
Answer scientific questions in professional fields
Performs better than general language models in knowledge-intensive tasks
Education
Mathematical problem-solving tutoring
Solve physics and mathematical problems step by step
Can show the complete reasoning process
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
Š 2025AIbase