Sac Pendulum V1
This is a reinforcement learning model based on the SAC algorithm, designed to solve control problems in the Pendulum-v1 environment.
Downloads 39
Release Time : 5/22/2022
Model Overview
The model is trained using the SAC algorithm from the Stable Baselines3 library, specifically for solving the inverted pendulum control problem in the Pendulum-v1 environment.
Model Features
Based on SAC Algorithm
Uses the Soft Actor-Critic algorithm, combining the advantages of policy gradients and Q-learning, suitable for control problems in continuous action spaces.
Stable Training
Implemented via the stable-baselines3 library, providing a stable training process and reliable performance.
Hyperparameter Optimization
The model has undergone hyperparameter optimization and performs well in the Pendulum-v1 environment.
Model Capabilities
Continuous Action Space Control
Reinforcement Learning Policy Optimization
Inverted Pendulum Balance Control
Use Cases
Control Problems
Inverted Pendulum Control
Control the inverted pendulum to maintain an upright position
Average reward -176.33 +/- 101.55
Reinforcement Learning Research
SAC Algorithm Benchmarking
Serves as a benchmark model for the SAC algorithm in continuous control tasks
Featured Recommended AI Models
Š 2025AIbase