Liu, D.[Daizong],
Zhu, J.H.[Jia-Hao],
Fang, X.[Xiang],
Xiong, Z.[Zeyu],
Wang, H.[Huan],
Li, R.[Renfu],
Zhou, P.[Pan],
Conditional Video Diffusion Network for Fine-Grained Temporal
Sentence Grounding,
MultMed(26), 2024, pp. 5461-5476.
IEEE DOI
2404
Semantics, Grounding, Task analysis, Cognition, Visualization,
Training, Stochastic processes, diffusion models
BibRef
Xing, Z.[Zhen],
Feng, Q.J.[Qi-Jun],
Chen, H.R.[Hao-Ran],
Dai, Q.[Qi],
Hu, H.[Han],
Xu, H.[Hang],
Wu, Z.[Zuxuan],
Jiang, Y.G.[Yu-Gang],
A Survey on Video Diffusion Models,
Surveys(57), No. 2, November 2024, pp. xx-yy.
DOI Link
2501
Survey, Video Diffusion. Survey, video diffusion model, video generation, video editing, AIGC
BibRef
Zatsarynna, O.[Olga],
Bahrami, E.[Emad],
Farha, Y.A.[Yazan Abu],
Francesca, G.[Gianpiero],
Gall, J.[Juergen],
Gated Temporal Diffusion for Stochastic Long-term Dense Anticipation,
ECCV24(LV: 454-472).
Springer DOI
2412
BibRef
Zhang, M.Y.[Ming-Yuan],
Jin, D.[Daisheng],
Gu, C.Y.[Chen-Yang],
Hong, F.Z.[Fang-Zhou],
Cai, Z.A.[Zhong-Ang],
Huang, J.[Jingfang],
Zhang, C.Z.[Chong-Zhi],
Guo, X.[Xinying],
Yang, L.[Lei],
He, Y.[Ying],
Liu, Z.W.[Zi-Wei],
Large Motion Model for Unified Multi-modal Motion Generation,
ECCV24(XIII: 397-421).
Springer DOI
2412
BibRef
Lu, J.C.[Jia-Chen],
Huang, Z.[Ze],
Yang, Z.[Zeyu],
Zhang, J.[Jiahui],
Zhang, L.[Li],
Wovogen: World Volume-aware Diffusion for Controllable Multi-camera
Driving Scene Generation,
ECCV24(LXXX: 329-345).
Springer DOI
2412
BibRef
Zhu, Y.Z.[Yuan-Zhi],
Liu, X.C.[Xing-Chao],
Liu, Q.[Qiang],
Slimflow: Training Smaller One-step Diffusion Models with Rectified
Flow,
ECCV24(LXXXII: 342-359).
Springer DOI
2412
BibRef
Zhong, L.[Lei],
Xie, Y.M.[Yi-Ming],
Jampani, V.[Varun],
Sun, D.Q.[De-Qing],
Jiang, H.[Huaizu],
Smoodi: Stylized Motion Diffusion Model,
ECCV24(I: 405-421).
Springer DOI
2412
BibRef
Zhou, W.Y.[Wen-Yang],
Dou, Z.Y.[Zhi-Yang],
Cao, Z.[Zeyu],
Liao, Z.Y.C.[Zhou-Ying-Cheng],
Wang, J.B.[Jing-Bo],
Wang, W.J.[Wen-Jia],
Liu, Y.[Yuan],
Komura, T.[Taku],
Wang, W.P.[Wen-Ping],
Liu, L.J.[Ling-Jie],
EMDM: Efficient Motion Diffusion Model for Fast and High-quality Motion
Generation,
ECCV24(II: 18-38).
Springer DOI
2412
BibRef
Gupta, A.[Agrim],
Yu, L.J.[Li-Jun],
Sohn, K.[Kihyuk],
Gu, X.[Xiuye],
Hahn, M.[Meera],
Li, F.F.[Fei-Fei],
Essa, I.[Irfan],
Jiang, L.[Lu],
Lezama, J.[José],
Photorealistic Video Generation with Diffusion Models,
ECCV24(LXXIX: 393-411).
Springer DOI
2412
BibRef
Han, J.L.[Jun-Lin],
Kokkinos, F.[Filippos],
Torr, P.H.S.[Philip H.S.],
Vfusion3d: Learning Scalable 3d Generative Models from Video Diffusion
Models,
ECCV24(II: 333-350).
Springer DOI
2412
BibRef
Shi, F.Y.[Feng-Yuan],
Gu, J.X.[Jia-Xi],
Xu, H.[Hang],
Xu, S.[Songcen],
Zhang, W.[Wei],
Wang, L.M.[Li-Min],
BIVDiff: A Training-Free Framework for General-Purpose Video
Synthesis via Bridging Image and Video Diffusion Models,
CVPR24(7393-7402)
IEEE DOI
2410
Training, Smoothing methods, Image synthesis, Memory management,
Text to image, Diffusion models, Video Synthesis, Diffusion models,
General Framework
BibRef
Ge, S.W.[Song-Wei],
Nah, S.J.[Seung-Jun],
Liu, G.L.[Gui-Lin],
Poon, T.[Tyler],
Tao, A.[Andrew],
Catanzaro, B.[Bryan],
Jacobs, D.[David],
Huang, J.B.[Jia-Bin],
Liu, M.Y.[Ming-Yu],
Balaji, Y.[Yogesh],
Preserve Your Own Correlation:
A Noise Prior for Video Diffusion Models,
ICCV23(22873-22884)
IEEE DOI Code:
WWW Link.
2401
BibRef
Zhu, Z.X.[Zi-Xin],
Feng, X.[Xuelu],
Chen, D.D.[Dong-Dong],
Yuan, J.S.[Jun-Song],
Qiao, C.M.[Chun-Ming],
Hua, G.[Gang],
Exploring Pre-trained Text-to-video Diffusion Models for Referring
Video Object Segmentation,
ECCV24(XII: 452-469).
Springer DOI
2412
BibRef
Guo, Y.W.[Yu-Wei],
Yang, C.[Ceyuan],
Rao, A.[Anyi],
Agrawala, M.[Maneesh],
Lin, D.[Dahua],
Dai, B.[Bo],
Sparsectrl: Adding Sparse Controls to Text-to-video Diffusion Models,
ECCV24(XLII: 330-348).
Springer DOI
2412
BibRef
Zhao, R.[Rui],
Gu, Y.C.[Yu-Chao],
Wu, J.Z.J.[Jay Zhang-Jie],
Zhang, D.J.H.[David Jun-Hao],
Liu, J.W.[Jia-Wei],
Wu, W.J.[Wei-Jia],
Keppo, J.[Jussi],
Shou, M.Z.[Mike Zheng],
Motiondirector: Motion Customization of Text-to-video Diffusion Models,
ECCV24(LVI: 273-290).
Springer DOI
2412
BibRef
Liang, J.Y.[Jing-Yun],
Fan, Y.C.[Yu-Chen],
Zhang, K.[Kai],
Timofte, R.[Radu],
Van Gool, L.J.[Luc J.],
Ranjan, R.[Rakesh],
Movideo: Motion-aware Video Generation with Diffusion Model,
ECCV24(XLIV: 56-74).
Springer DOI
2412
BibRef
Huang, T.[Tao],
Jiang, G.Q.[Guang-Qi],
Ze, Y.J.[Yan-Jie],
Xu, H.Z.[Hua-Zhe],
Diffusion Reward: Learning Rewards via Conditional Video Diffusion,
ECCV24(XLII: 478-495).
Springer DOI
2412
BibRef
Niu, M.[Muyao],
Cun, X.D.[Xiao-Dong],
Wang, X.[Xintao],
Zhang, Y.[Yong],
Shan, Y.[Ying],
Zheng, Y.Q.[Yin-Qiang],
MOFA-Video: Controllable Image Animation via Generative Motion Field
Adaptions in Frozen Image-to-Video Diffusion Model,
ECCV24(XIX: 111-128).
Springer DOI
2412
BibRef
Wu, T.X.[Tian-Xing],
Si, C.Y.[Chen-Yang],
Jiang, Y.M.[Yu-Ming],
Huang, Z.Q.[Zi-Qi],
Liu, Z.W.[Zi-Wei],
Freeinit: Bridging Initialization Gap in Video Diffusion Models,
ECCV24(III: 378-394).
Springer DOI
2412
BibRef
Xing, J.[Jinbo],
Xia, M.[Menghan],
Zhang, Y.[Yong],
Chen, H.X.[Hao-Xin],
Yu, W.[Wangbo],
Liu, H.[Hanyuan],
Liu, G.[Gongye],
Wang, X.[Xintao],
Shan, Y.[Ying],
Wong, T.T.[Tien-Tsin],
Dynamicrafter: Animating Open-domain Images with Video Diffusion Priors,
ECCV24(XLVI: 399-417).
Springer DOI
2412
BibRef
Kim, K.[Kihong],
Lee, H.[Haneol],
Park, J.[Jihye],
Kim, S.[Seyeon],
Lee, K.[Kwanghee],
Kim, S.[Seungryong],
Yoo, J.[Jaejun],
Hybrid Video Diffusion Models with 2d Triplane and 3d Wavelet
Representation,
ECCV24(LII: 148-165).
Springer DOI
2412
BibRef
Yuan, H.J.[Hang-Jie],
Zhang, S.W.[Shi-Wei],
Wang, X.[Xiang],
Wei, Y.J.[Yu-Jie],
Feng, T.[Tao],
Pan, Y.[Yining],
Zhang, Y.[Yingya],
Liu, Z.W.[Zi-Wei],
Albanie, S.[Samuel],
Ni, D.[Dong],
InstructVideo: Instructing Video Diffusion Models with Human Feedback,
CVPR24(6463-6474)
IEEE DOI Code:
WWW Link.
2410
Degradation, Visualization, Image segmentation, Costs,
Computational modeling, Diffusion processes, video generation,
reward fine-tuning
BibRef
Ni, H.[Haomiao],
Egger, B.[Bernhard],
Lohit, S.[Suhas],
Cherian, A.[Anoop],
Wang, Y.[Ye],
Koike-Akino, T.[Toshiaki],
Huang, S.X.[Sharon X.],
Marks, T.K.[Tim K.],
TI2V-Zero: Zero-Shot Image Conditioning for Text-to-Video Diffusion
Models,
CVPR24(9015-9025)
IEEE DOI
2410
Training, Visualization, Gaussian noise, Noise reduction,
Diffusion models, Sampling methods
BibRef
Jeong, H.[Hyeonho],
Park, G.Y.[Geon Yeong],
Ye, J.C.[Jong Chul],
VMC: Video Motion Customization Using Temporal Attention Adaption for
Text-to-Video Diffusion Models,
CVPR24(9212-9221)
IEEE DOI Code:
WWW Link.
2410
Visualization, Adaptation models, Noise, Diffusion models, Vectors,
Trajectory, Diffusion Models,
Motion Customization
BibRef
Motamed, S.[Saman],
van Gansbeke, W.[Wouter],
Van Gool, L.J.[Luc J.],
Investigating the Effectiveness of Cross-Attention to Unlock
Zero-Shot Editing of Text-to-Video Diffusion Models,
GCV24(7406-7415)
IEEE DOI Code:
WWW Link.
2410
Codes, Shape, Computational modeling, Text to image,
Diffusion models, Video Diffusion models,
Video editing
BibRef
Fei, H.[Hao],
Wu, S.Q.[Sheng-Qiong],
Ji, W.[Wei],
Zhang, H.W.[Han-Wang],
Chua, T.S.[Tat-Seng],
Dysen-VDM: Empowering Dynamics-Aware Text-to-Video Diffusion with
LLMs,
CVPR24(7641-7653)
IEEE DOI Code:
WWW Link.
2410
Schedules, Codes, Dynamics, Dynamic scheduling, Diffusion models, Chatbots
BibRef
Blattmann, A.[Andreas],
Rombach, R.[Robin],
Ling, H.[Huan],
Dockhorn, T.[Tim],
Kim, S.W.[Seung Wook],
Fidler, S.[Sanja],
Kreis, K.[Karsten],
Align Your Latents: High-Resolution Video Synthesis with Latent
Diffusion Models,
CVPR23(22563-22575)
IEEE DOI
2309
BibRef
Rombach, R.[Robin],
Blattmann, A.[Andreas],
Lorenz, D.[Dominik],
Esser, P.[Patrick],
Ommer, B.[Björn],
High-Resolution Image Synthesis with Latent Diffusion Models,
CVPR22(10674-10685)
IEEE DOI
2210
Training, Visualization, Image synthesis, Computational modeling,
Noise reduction, Superresolution, Process control,
Image and video synthesis and generation
BibRef
Chapter on 3-D Object Description and Computation Techniques, Surfaces, Deformable, View Generation, Video Conferencing continues in
Image Matting, Video Matting .