V

Videomae Base Short Finetuned Kinetics

Developed by MCG-NJU
VideoMAE is a video self-supervised pre-training model based on Masked Autoencoder (MAE), fine-tuned on the Kinetics-400 dataset for video classification tasks.
Downloads 62
Release Time : 8/2/2022

Model Overview

This model was pre-trained for 800 epochs in a self-supervised manner and then fine-tuned on the Kinetics-400 dataset, capable of classifying videos into one of 400 possible categories.

Model Features

Self-supervised Pre-training
Uses Masked Autoencoder (MAE) method for self-supervised pre-training, reducing reliance on labeled data
Efficient Video Representation Learning
Capable of learning internal video representations and extracting features needed for downstream tasks
High Accuracy
Achieves 79.4% top-1 accuracy and 94.1% top-5 accuracy on the Kinetics-400 test set

Model Capabilities

Video Classification
Video Feature Extraction

Use Cases

Video Content Analysis
Video Classification
Classify videos into one of the 400 categories in the Kinetics-400 dataset
Achieves 79.4% top-1 accuracy on the Kinetics-400 test set
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
Š 2025AIbase