đ OmniGen: Unified Image Generation
OmniGen is a unified image generation model that can generate a wide range of images from multi-modal prompts. It simplifies the image generation process and offers flexibility for various tasks.
More information please refer to our repo: https://github.com/VectorSpaceLab/OmniGen
News |
Methodology |
Capabilities |
Quick Start |
Finetune |
License |
Citation
⨠Features
OmniGen is a unified image generation model capable of performing various tasks, including text-to-image generation, subject-driven generation, Identity-Preserving Generation, image editing, and image-conditioned generation. It can automatically identify features in input images according to text prompts without the need for additional plugins or operations.
đ Quick Start
Using OmniGen
Install via Github:
git clone https://github.com/staoxiao/OmniGen.git
cd OmniGen
pip install -e .
You also can create a new environment to avoid conflicts:
# Create a python 3.10.12 conda env (you could also use virtualenv)
conda create -n omnigen python=3.10.12
conda activate omnigen
# Install pytorch with your CUDA version, e.g.
pip install torch==2.3.1+cu118 torchvision --extra-index-url https://download.pytorch.org/whl/cu118
git clone https://github.com/staoxiao/OmniGen.git
cd OmniGen
pip install -e .
Here are some examples:
from OmniGen import OmniGenPipeline
pipe = OmniGenPipeline.from_pretrained("Shitao/OmniGen-v1")
images = pipe(
prompt="A curly-haired man in a red shirt is drinking tea.",
height=1024,
width=1024,
guidance_scale=2.5,
seed=0,
)
images[0].save("example_t2i.png")
images = pipe(
prompt="A man in a black shirt is reading a book. The man is the right man in <img><|image_1|></img>.",
input_images=["./imgs/test_cases/two_man.jpg"],
height=1024,
width=1024,
guidance_scale=2.5,
img_guidance_scale=1.6,
seed=0
)
images[0].save("example_ti2i.png")
- If out of memory, you can set
offload_model=True
. If the inference time is too long when inputting multiple images, you can reduce the max_input_image_size
. For the required resources and the method to run OmniGen efficiently, please refer to docs/inference.md#requiremented-resources.
- For more examples of image generation, you can refer to inference.ipynb and inference_demo.ipynb
- For more details about the argument in inference, please refer to docs/inference.md.
Using Diffusers
Coming soon.
Gradio Demo
We construct an online demo in Huggingface.
For the local gradio demo, you need to install pip install gradio spaces
, and then you can run:
pip install gradio spaces
python app.py
Use Google Colab
To use with Google Colab, please use the following command:
!git clone https://github.com/staoxiao/OmniGen.git
%cd OmniGen
!pip install -e .
!pip install gradio spaces
!python app.py --share
đ§ Technical Details
You can see details in our paper.
đĻ Finetune
We provide a training script train.py
to fine-tune OmniGen.
Here is a toy example about LoRA finetune:
accelerate launch --num_processes=1 train.py \
--model_name_or_path Shitao/OmniGen-v1 \
--batch_size_per_device 2 \
--condition_dropout_prob 0.01 \
--lr 1e-3 \
--use_lora \
--lora_rank 8 \
--json_file ./toy_data/toy_subject_data.jsonl \
--image_path ./toy_data/images \
--max_input_length_limit 18000 \
--keep_raw_resolution \
--max_image_size 1024 \
--gradient_accumulation_steps 1 \
--ckpt_every 10 \
--epochs 200 \
--log_every 1 \
--results_dir ./results/toy_finetune_lora
Please refer to docs/fine-tuning.md for more details (e.g. full finetune).
Contributors:
Thank all our contributors for their efforts and warmly welcome new members to join in!
đ License
This repo is licensed under the MIT License.
Citation
If you find this repository useful, please consider giving a star â and citation
@article{xiao2024omnigen,
title={Omnigen: Unified image generation},
author={Xiao, Shitao and Wang, Yueze and Zhou, Junjie and Yuan, Huaying and Xing, Xingrun and Yan, Ruiran and Wang, Shuting and Huang, Tiejun and Liu, Zheng},
journal={arXiv preprint arXiv:2409.11340},
year={2024}
}