11.14.3.7.2 Video Diffusion, Video Sysnthesis

Chapter Contents (Back)
Diffusion. Video Synthesis. Video Diffusion.
See also Diffusion for Description or Text to Image Generation.

Liu, D.[Daizong], Zhu, J.H.[Jia-Hao], Fang, X.[Xiang], Xiong, Z.[Zeyu], Wang, H.[Huan], Li, R.[Renfu], Zhou, P.[Pan],
Conditional Video Diffusion Network for Fine-Grained Temporal Sentence Grounding,
MultMed(26), 2024, pp. 5461-5476.
IEEE DOI 2404
Semantics, Grounding, Task analysis, Cognition, Visualization, Training, Stochastic processes, diffusion models BibRef

Xing, Z.[Zhen], Feng, Q.J.[Qi-Jun], Chen, H.R.[Hao-Ran], Dai, Q.[Qi], Hu, H.[Han], Xu, H.[Hang], Wu, Z.[Zuxuan], Jiang, Y.G.[Yu-Gang],
A Survey on Video Diffusion Models,
Surveys(57), No. 2, November 2024, pp. xx-yy.
DOI Link 2501
Survey, Video Diffusion. Survey, video diffusion model, video generation, video editing, AIGC BibRef


Yang, J.W.[Jing-Wen], Sun, J.M.[Jia-Mu], Yang, Y.L.[Yong-Liang], Yang, J.[Jie], Shan, Y.[Ying], Cao, Y.P.[Yan-Pei], Gao, L.[Lin],
DMiT: Deformable Mipmapped Tri-plane Representation for Dynamic Scenes,
ECCV24(LV: 436-453).
Springer DOI 2412
BibRef

Zatsarynna, O.[Olga], Bahrami, E.[Emad], Farha, Y.A.[Yazan Abu], Francesca, G.[Gianpiero], Gall, J.[Juergen],
Gated Temporal Diffusion for Stochastic Long-term Dense Anticipation,
ECCV24(LV: 454-472).
Springer DOI 2412
BibRef

Zhang, M.Y.[Ming-Yuan], Jin, D.[Daisheng], Gu, C.Y.[Chen-Yang], Hong, F.Z.[Fang-Zhou], Cai, Z.A.[Zhong-Ang], Huang, J.[Jingfang], Zhang, C.Z.[Chong-Zhi], Guo, X.[Xinying], Yang, L.[Lei], He, Y.[Ying], Liu, Z.W.[Zi-Wei],
Large Motion Model for Unified Multi-modal Motion Generation,
ECCV24(XIII: 397-421).
Springer DOI 2412
BibRef

Lu, J.C.[Jia-Chen], Huang, Z.[Ze], Yang, Z.[Zeyu], Zhang, J.[Jiahui], Zhang, L.[Li],
Wovogen: World Volume-aware Diffusion for Controllable Multi-camera Driving Scene Generation,
ECCV24(LXXX: 329-345).
Springer DOI 2412
BibRef

Zhu, Y.Z.[Yuan-Zhi], Liu, X.C.[Xing-Chao], Liu, Q.[Qiang],
Slimflow: Training Smaller One-step Diffusion Models with Rectified Flow,
ECCV24(LXXXII: 342-359).
Springer DOI 2412
BibRef

Zhong, L.[Lei], Xie, Y.M.[Yi-Ming], Jampani, V.[Varun], Sun, D.Q.[De-Qing], Jiang, H.[Huaizu],
Smoodi: Stylized Motion Diffusion Model,
ECCV24(I: 405-421).
Springer DOI 2412
BibRef

Zhou, W.Y.[Wen-Yang], Dou, Z.Y.[Zhi-Yang], Cao, Z.[Zeyu], Liao, Z.Y.C.[Zhou-Ying-Cheng], Wang, J.B.[Jing-Bo], Wang, W.J.[Wen-Jia], Liu, Y.[Yuan], Komura, T.[Taku], Wang, W.P.[Wen-Ping], Liu, L.J.[Ling-Jie],
EMDM: Efficient Motion Diffusion Model for Fast and High-quality Motion Generation,
ECCV24(II: 18-38).
Springer DOI 2412
BibRef

Gupta, A.[Agrim], Yu, L.J.[Li-Jun], Sohn, K.[Kihyuk], Gu, X.[Xiuye], Hahn, M.[Meera], Li, F.F.[Fei-Fei], Essa, I.[Irfan], Jiang, L.[Lu], Lezama, J.[José],
Photorealistic Video Generation with Diffusion Models,
ECCV24(LXXIX: 393-411).
Springer DOI 2412
BibRef

Han, J.L.[Jun-Lin], Kokkinos, F.[Filippos], Torr, P.H.S.[Philip H.S.],
Vfusion3d: Learning Scalable 3d Generative Models from Video Diffusion Models,
ECCV24(II: 333-350).
Springer DOI 2412
BibRef

Shi, F.Y.[Feng-Yuan], Gu, J.X.[Jia-Xi], Xu, H.[Hang], Xu, S.[Songcen], Zhang, W.[Wei], Wang, L.M.[Li-Min],
BIVDiff: A Training-Free Framework for General-Purpose Video Synthesis via Bridging Image and Video Diffusion Models,
CVPR24(7393-7402)
IEEE DOI 2410
Training, Smoothing methods, Image synthesis, Memory management, Text to image, Diffusion models, Video Synthesis, Diffusion models, General Framework BibRef

Ge, S.W.[Song-Wei], Nah, S.J.[Seung-Jun], Liu, G.L.[Gui-Lin], Poon, T.[Tyler], Tao, A.[Andrew], Catanzaro, B.[Bryan], Jacobs, D.[David], Huang, J.B.[Jia-Bin], Liu, M.Y.[Ming-Yu], Balaji, Y.[Yogesh],
Preserve Your Own Correlation: A Noise Prior for Video Diffusion Models,
ICCV23(22873-22884)
IEEE DOI Code:
WWW Link. 2401
BibRef

Zhu, Z.X.[Zi-Xin], Feng, X.[Xuelu], Chen, D.D.[Dong-Dong], Yuan, J.S.[Jun-Song], Qiao, C.M.[Chun-Ming], Hua, G.[Gang],
Exploring Pre-trained Text-to-video Diffusion Models for Referring Video Object Segmentation,
ECCV24(XII: 452-469).
Springer DOI 2412
BibRef

Guo, Y.W.[Yu-Wei], Yang, C.[Ceyuan], Rao, A.[Anyi], Agrawala, M.[Maneesh], Lin, D.[Dahua], Dai, B.[Bo],
Sparsectrl: Adding Sparse Controls to Text-to-video Diffusion Models,
ECCV24(XLII: 330-348).
Springer DOI 2412
BibRef

Zhao, R.[Rui], Gu, Y.C.[Yu-Chao], Wu, J.Z.J.[Jay Zhang-Jie], Zhang, D.J.H.[David Jun-Hao], Liu, J.W.[Jia-Wei], Wu, W.J.[Wei-Jia], Keppo, J.[Jussi], Shou, M.Z.[Mike Zheng],
Motiondirector: Motion Customization of Text-to-video Diffusion Models,
ECCV24(LVI: 273-290).
Springer DOI 2412
BibRef

Liang, J.Y.[Jing-Yun], Fan, Y.C.[Yu-Chen], Zhang, K.[Kai], Timofte, R.[Radu], Van Gool, L.J.[Luc J.], Ranjan, R.[Rakesh],
Movideo: Motion-aware Video Generation with Diffusion Model,
ECCV24(XLIV: 56-74).
Springer DOI 2412
BibRef

Huang, T.[Tao], Jiang, G.Q.[Guang-Qi], Ze, Y.J.[Yan-Jie], Xu, H.Z.[Hua-Zhe],
Diffusion Reward: Learning Rewards via Conditional Video Diffusion,
ECCV24(XLII: 478-495).
Springer DOI 2412
BibRef

Niu, M.[Muyao], Cun, X.D.[Xiao-Dong], Wang, X.[Xintao], Zhang, Y.[Yong], Shan, Y.[Ying], Zheng, Y.Q.[Yin-Qiang],
MOFA-Video: Controllable Image Animation via Generative Motion Field Adaptions in Frozen Image-to-Video Diffusion Model,
ECCV24(XIX: 111-128).
Springer DOI 2412
BibRef

Wu, T.X.[Tian-Xing], Si, C.Y.[Chen-Yang], Jiang, Y.M.[Yu-Ming], Huang, Z.Q.[Zi-Qi], Liu, Z.W.[Zi-Wei],
Freeinit: Bridging Initialization Gap in Video Diffusion Models,
ECCV24(III: 378-394).
Springer DOI 2412
BibRef

Xing, J.[Jinbo], Xia, M.[Menghan], Zhang, Y.[Yong], Chen, H.X.[Hao-Xin], Yu, W.[Wangbo], Liu, H.[Hanyuan], Liu, G.[Gongye], Wang, X.[Xintao], Shan, Y.[Ying], Wong, T.T.[Tien-Tsin],
Dynamicrafter: Animating Open-domain Images with Video Diffusion Priors,
ECCV24(XLVI: 399-417).
Springer DOI 2412
BibRef

Kim, K.[Kihong], Lee, H.[Haneol], Park, J.[Jihye], Kim, S.[Seyeon], Lee, K.[Kwanghee], Kim, S.[Seungryong], Yoo, J.[Jaejun],
Hybrid Video Diffusion Models with 2d Triplane and 3d Wavelet Representation,
ECCV24(LII: 148-165).
Springer DOI 2412
BibRef

Yuan, H.J.[Hang-Jie], Zhang, S.W.[Shi-Wei], Wang, X.[Xiang], Wei, Y.J.[Yu-Jie], Feng, T.[Tao], Pan, Y.[Yining], Zhang, Y.[Yingya], Liu, Z.W.[Zi-Wei], Albanie, S.[Samuel], Ni, D.[Dong],
InstructVideo: Instructing Video Diffusion Models with Human Feedback,
CVPR24(6463-6474)
IEEE DOI Code:
WWW Link. 2410
Degradation, Visualization, Image segmentation, Costs, Computational modeling, Diffusion processes, video generation, reward fine-tuning BibRef

Ni, H.[Haomiao], Egger, B.[Bernhard], Lohit, S.[Suhas], Cherian, A.[Anoop], Wang, Y.[Ye], Koike-Akino, T.[Toshiaki], Huang, S.X.[Sharon X.], Marks, T.K.[Tim K.],
TI2V-Zero: Zero-Shot Image Conditioning for Text-to-Video Diffusion Models,
CVPR24(9015-9025)
IEEE DOI 2410
Training, Visualization, Gaussian noise, Noise reduction, Diffusion models, Sampling methods BibRef

Jeong, H.[Hyeonho], Park, G.Y.[Geon Yeong], Ye, J.C.[Jong Chul],
VMC: Video Motion Customization Using Temporal Attention Adaption for Text-to-Video Diffusion Models,
CVPR24(9212-9221)
IEEE DOI Code:
WWW Link. 2410
Visualization, Adaptation models, Noise, Diffusion models, Vectors, Trajectory, Diffusion Models, Motion Customization BibRef

Motamed, S.[Saman], van Gansbeke, W.[Wouter], Van Gool, L.J.[Luc J.],
Investigating the Effectiveness of Cross-Attention to Unlock Zero-Shot Editing of Text-to-Video Diffusion Models,
GCV24(7406-7415)
IEEE DOI Code:
WWW Link. 2410
Codes, Shape, Computational modeling, Text to image, Diffusion models, Video Diffusion models, Video editing BibRef

Fei, H.[Hao], Wu, S.Q.[Sheng-Qiong], Ji, W.[Wei], Zhang, H.W.[Han-Wang], Chua, T.S.[Tat-Seng],
Dysen-VDM: Empowering Dynamics-Aware Text-to-Video Diffusion with LLMs,
CVPR24(7641-7653)
IEEE DOI Code:
WWW Link. 2410
Schedules, Codes, Dynamics, Dynamic scheduling, Diffusion models, Chatbots BibRef

Blattmann, A.[Andreas], Rombach, R.[Robin], Ling, H.[Huan], Dockhorn, T.[Tim], Kim, S.W.[Seung Wook], Fidler, S.[Sanja], Kreis, K.[Karsten],
Align Your Latents: High-Resolution Video Synthesis with Latent Diffusion Models,
CVPR23(22563-22575)
IEEE DOI 2309
BibRef

Rombach, R.[Robin], Blattmann, A.[Andreas], Lorenz, D.[Dominik], Esser, P.[Patrick], Ommer, B.[Björn],
High-Resolution Image Synthesis with Latent Diffusion Models,
CVPR22(10674-10685)
IEEE DOI 2210
Training, Visualization, Image synthesis, Computational modeling, Noise reduction, Superresolution, Process control, Image and video synthesis and generation BibRef

Chapter on 3-D Object Description and Computation Techniques, Surfaces, Deformable, View Generation, Video Conferencing continues in
Image Matting, Video Matting .


Last update:Jan 15, 2025 at 14:36:47