DiscoverDaily Paper CastTransPixar: Advancing Text-to-Video Generation with Transparency
TransPixar: Advancing Text-to-Video Generation with Transparency

TransPixar: Advancing Text-to-Video Generation with Transparency

Update: 2025-01-08
Share

Description

🤗 Upvotes: 9 | cs.CV



Authors:

Luozhou Wang, Yijun Li, Zhifei Chen, Jui-Hsien Wang, Zhifei Zhang, He Zhang, Zhe Lin, Yingcong Chen



Title:

TransPixar: Advancing Text-to-Video Generation with Transparency



Arxiv:

http://arxiv.org/abs/2501.03006v1



Abstract:

Text-to-video generative models have made significant strides, enabling diverse applications in entertainment, advertising, and education. However, generating RGBA video, which includes alpha channels for transparency, remains a challenge due to limited datasets and the difficulty of adapting existing models. Alpha channels are crucial for visual effects (VFX), allowing transparent elements like smoke and reflections to blend seamlessly into scenes. We introduce TransPixar, a method to extend pretrained video models for RGBA generation while retaining the original RGB capabilities. TransPixar leverages a diffusion transformer (DiT) architecture, incorporating alpha-specific tokens and using LoRA-based fine-tuning to jointly generate RGB and alpha channels with high consistency. By optimizing attention mechanisms, TransPixar preserves the strengths of the original RGB model and achieves strong alignment between RGB and alpha channels despite limited training data. Our approach effectively generates diverse and consistent RGBA videos, advancing the possibilities for VFX and interactive content creation.

Comments 
In Channel
loading
00:00
00:00
x

0.5x

0.8x

1.0x

1.25x

1.5x

2.0x

3.0x

Sleep Timer

Off

End of Episode

5 Minutes

10 Minutes

15 Minutes

30 Minutes

45 Minutes

60 Minutes

120 Minutes

TransPixar: Advancing Text-to-Video Generation with Transparency

TransPixar: Advancing Text-to-Video Generation with Transparency

Jingwen Liang, Gengyu Wang