๐ TransPixar: Advancing Text-to-Video Generation with Transparency
This repository presents a model that aims to address the challenge of generating RGBA videos with transparency. It extends existing text-to-video technology, enabling more diverse and consistent video outputs, which are highly valuable for visual effects and interactive content creation.
Links
Authors
Luozhou Wang*,
Yijun Li**,
Zhifei Chen,
Jui-Hsien Wang,
Zhifei Zhang,
He Zhang,
Zhe Lin,
Yingcong Chenโ
Affiliations: HKUST(GZ), HKUST, Adobe Research.
* Intership Project.
** Project Leader.
โ Corresponding Author.
Abstract
Text-to-video generative models have advanced significantly, finding applications in entertainment, advertising, and education. However, generating RGBA videos with alpha channels for transparency remains a challenge due to limited datasets and difficulties in adapting existing models. Alpha channels are essential for visual effects, allowing seamless integration of transparent elements. This paper introduces TransPixar, a method that extends pretrained video models for RGBA generation while maintaining RGB capabilities. It uses a diffusion transformer architecture, alpha-specific tokens, and LoRA-based fine-tuning to generate consistent RGB and alpha channels. By optimizing attention mechanisms, TransPixar preserves the strengths of the original RGB model and aligns RGB and alpha channels effectively, even with limited training data.
๐ Quick Start
๐ฐ News
- [2024.01.07] We have released the project page, arXiv paper, inference code, and Huggingface demo for TransPixar + CogVideoX.
๐ง Todo List
- [x] Release code, paper, and demo.
- [x] Release checkpoints of joint generation (RGB + Alpha).
Contents
๐ฆ Installation
conda create -n TransPixar python=3.10
conda activate TransPixar
pip install -r requirements.txt
โจ Features
TransPixar LoRA Hub
Our pipeline supports various video tasks, such as Text-to-RGBA Video and Image-to-RGBA Video. We offer pre-trained LoRA weights for different tasks:
๐ Documentation
Training - RGB + Alpha Joint Generation
We have open-sourced the training code for Mochi on RGBA joint generation. For detailed information, please refer to the Mochi README.
Inference
Gradio Demo
In addition to the Hugging Face online demo, you can launch a local inference demo based on CogVideoX-5B by running the following command:
python app.py
Command Line Interface (CLI)
To generate RGBA videos, navigate to the corresponding directory for the video model and execute the following command:
python cli.py \
--lora_path /path/to/lora \
--prompt "..." \
๐ Acknowledgement
- finetrainers: We followed their implementation of Mochi training and inference.
- CogVideoX: We followed their implementation of CogVideoX training and inference.
We appreciate their outstanding work and contributions to the open-source community.
๐ License
This project is licensed under the Apache-2.0 license.
๐ Citation
@misc{wang2025transpixar,
title={TransPixar: Advancing Text-to-Video Generation with Transparency},
author={Luozhou Wang and Yijun Li and Zhifei Chen and Jui-Hsien Wang and Zhifei Zhang and He Zhang and Zhe Lin and Yingcong Chen},
year={2025},
eprint={2501.03006},
archivePrefix={arXiv},
primaryClass={cs.CV},
url={https://arxiv.org/abs/2501.03006},
}