S

Siamese Smole Bert Muv 1x

Developed by UdS-LSV
A neural language model toolkit for pretraining and fine-tuning SMILES-based molecular language models, supporting semi-supervised learning
Downloads 33
Release Time : 8/17/2022

Model Overview

This model incorporates enumeration knowledge into pretrained language models through contrastive learning, multi-task regression, and masked language modeling as pretraining objectives, suitable for virtual screening and property prediction in the molecular domain.

Model Features

Enumeration-Aware Pretraining
Injects molecular enumeration knowledge into the pretraining process via contrastive learning and multi-task regression
Domain Adaptation Capability
Uses a Siamese BERT architecture for contrastive learning adaptation in the molecular domain
Semi-Supervised Learning Support
Provides semi-supervised fine-tuning solutions for low-data environments

Model Capabilities

Molecular Property Prediction
Virtual Screening
Molecular Representation Learning
Semi-Supervised Learning

Use Cases

Drug Discovery
Virtual Screening
Uses the model to predict molecular activity and screen potential drug candidates
Achieved 0.697 AUROC on the MUV dataset
Molecular Property Prediction
Predicts physicochemical properties of molecules
Pretrained on the Guacamol dataset
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
Š 2025AIbase