Search Results for author: Yuwei Zhou

Found 4 papers, 1 papers with code

DisenStudio: Customized Multi-subject Text-to-Video Generation with Disentangled Spatial Control

no code implementations21 May 2024 Hong Chen, Xin Wang, YiPeng Zhang, Yuwei Zhou, Zeyang Zhang, Siao Tang, Wenwu Zhu

To tackle the problems, in this paper, we propose DisenStudio, a novel framework that can generate text-guided videos for customized multiple subjects, given few images for each subject.

Attribute Text-to-Video Generation +1

LLM4VG: Large Language Models Evaluation for Video Grounding

no code implementations21 Dec 2023 Wei Feng, Xin Wang, Hong Chen, Zeyang Zhang, Zihan Song, Yuwei Zhou, Wenwu Zhu

Recently, researchers have attempted to investigate the capability of LLMs in handling videos and proposed several video LLM models.

Image Captioning Video Grounding +1

VideoDreamer: Customized Multi-Subject Text-to-Video Generation with Disen-Mix Finetuning

no code implementations2 Nov 2023 Hong Chen, Xin Wang, Guanning Zeng, YiPeng Zhang, Yuwei Zhou, Feilin Han, Wenwu Zhu

The video generator is further customized for the given multiple subjects by the proposed Disen-Mix Finetuning and Human-in-the-Loop Re-finetuning strategy, which can tackle the attribute binding problem of multi-subject generation.

Attribute Text-to-Video Generation +1

DisenBooth: Identity-Preserving Disentangled Tuning for Subject-Driven Text-to-Image Generation

1 code implementation5 May 2023 Hong Chen, YiPeng Zhang, Simin Wu, Xin Wang, Xuguang Duan, Yuwei Zhou, Wenwu Zhu

To tackle the problems, we propose DisenBooth, an identity-preserving disentangled tuning framework for subject-driven text-to-image generation.

Denoising Disentanglement +1

Cannot find the paper you are looking for? You can Submit a new open access paper.