Sun, G.[Gan],
Liang, W.Q.[Wen-Qi],
Dong, J.H.[Jia-Hua],
Li, J.[Jun],
Ding, Z.M.[Zheng-Ming],
Cong, Y.[Yang],
Create Your World: Lifelong Text-to-Image Diffusion,
PAMI(46), No. 9, September 2024, pp. 6454-6470.
IEEE DOI
2408
Task analysis, Dogs, Computational modeling, Semantics, Training,
Neural networks, Continual learning, image generation,
stable diffusion
BibRef
Chen, H.[Hong],
Zhang, Y.P.[Yi-Peng],
Wang, X.[Xin],
Duan, X.G.[Xu-Guang],
Zhou, Y.W.[Yu-Wei],
Zhu, W.W.[Wen-Wu],
DisenDreamer: Subject-Driven Text-to-Image Generation With
Sample-Aware Disentangled Tuning,
CirSysVideo(34), No. 8, August 2024, pp. 6860-6873.
IEEE DOI
2408
Noise reduction, Visualization, Tuning, Controllability,
Circuits and systems, Image synthesis, Training, Diffusion model,
disentangled finetuning
BibRef
Verma, A.[Ayushi],
Badal, T.[Tapas],
Bansal, A.[Abhay],
Advancing Image Generation with Denoising Diffusion Probabilistic
Model and ConvNeXt-V2:
A novel approach for enhanced diversity and quality,
CVIU(247), 2024, pp. 104077.
Elsevier DOI
2408
Deep learning, Diffusion model, Generative model, Image generation
BibRef
Xu, Y.F.[Yi-Fei],
Xu, X.L.[Xiao-Long],
Gao, H.H.[Hong-Hao],
Xiao, F.[Fu],
SGDM: An Adaptive Style-Guided Diffusion Model for Personalized Text
to Image Generation,
MultMed(26), 2024, pp. 9804-9813.
IEEE DOI
2410
Feature extraction, Adaptation models, Image synthesis,
Computational modeling, Training, Task analysis, Noise reduction,
image style similarity assessment
BibRef
Ramasinghe, S.[Sameera],
Shevchenko, V.[Violetta],
Avraham, G.[Gil],
Thalaiyasingam, A.[Ajanthan],
Accept the Modality Gap: An Exploration in the Hyperbolic Space,
CVPR24(27253-27262)
IEEE DOI
2410
Text to image, Machine learning, Linear programming,
multimodal learning, modality gap
BibRef
Luo, Y.M.[Yi-Min],
Yang, Q.[Qinyu],
Fan, Y.H.[Yu-Heng],
Qi, H.K.[Hai-Kun],
Xia, M.[Menghan],
Measurement Guidance in Diffusion Models:
Insight from Medical Image Synthesis,
PAMI(46), No. 12, December 2024, pp. 7983-7997.
IEEE DOI
2411
Task analysis, Medical diagnostic imaging, Uncertainty,
Image synthesis, Training, Reliability, Data models,
controllable generation
BibRef
Cao, J.H.[Jing-Hao],
Liu, S.[Sheng],
Yang, X.[Xiong],
Li, Y.[Yang],
Du, S.[Sidan],
ARES: Text-Driven Automatic Realistic Simulator for Autonomous
Traffic,
SPLetters(31), 2024, pp. 3049-3053.
IEEE DOI
2411
Trajectory, Rendering (computer graphics), Training,
Diffusion models, Accuracy, Logic, Turning, Predictive models
BibRef
Ren, J.X.[Jia-Xin],
Liu, W.[Wanzeng],
Chen, J.[Jun],
Yin, S.[Shunxi],
Tao, Y.[Yuan],
Word2Scene: Efficient remote sensing image scene generation with only
one word via hybrid intelligence and low-rank representation,
PandRS(218), 2024, pp. 231-257.
Elsevier DOI Code:
WWW Link.
2412
Intelligentized surveying and mapping, Hybrid intelligence,
Remote sensing image scene generation, Diffusion models, Zero-shot learning
BibRef
Salehi, S.[Sogand],
Shafiei, M.[Mahdi],
Yeo, T.[Teresa],
Bachmann, R.[Roman],
Zamir, A.[Amir],
Viper: Visual Personalization of Generative Models via Individual
Preference Learning,
ECCV24(LXXIV: 391-406).
Springer DOI
2412
Code:
WWW Link.
BibRef
Um, S.[Soobin],
Ye, J.C.[Jong Chul],
Self-guided Generation of Minority Samples Using Diffusion Models,
ECCV24(LXVIII: 414-430).
Springer DOI
2412
Code:
WWW Link.
BibRef
Mukhopadhyay, S.[Soumik],
Gwilliam, M.[Matthew],
Yamaguchi, Y.[Yosuke],
Agarwal, V.[Vatsal],
Padmanabhan, N.[Namitha],
Swaminathan, A.[Archana],
Zhou, T.Y.[Tian-Yi],
Ohya, J.[Jun],
Shrivastava, A.[Abhinav],
Do Text-free Diffusion Models Learn Discriminative Visual
Representations?,
ECCV24(LX: 253-272).
Springer DOI
2412
Project:
WWW Link. Code:
WWW Link.
BibRef
Wang, J.Y.[Jia-Yi],
Laube, K.A.[Kevin Alexander],
Li, Y.[Yumeng],
Metzen, J.H.[Jan Hendrik],
Cheng, S.I.[Shin-I],
Borges, J.[Julio],
Khoreva, A.[Anna],
Label-free Neural Semantic Image Synthesis,
ECCV24(LIII: 391-407).
Springer DOI
2412
BibRef
Xu, C.[Chen],
Song, T.[Tianhui],
Feng, W.X.[Wei-Xin],
Li, X.B.[Xu-Bin],
Ge, T.[Tiezheng],
Zheng, B.[Bo],
Wang, L.M.[Li-Min],
Accelerating Image Generation with Sub-path Linear Approximation Model,
ECCV24(LIII: 323-339).
Springer DOI
2412
BibRef
Zhang, S.[Shen],
Chen, Z.W.[Zhao-Wei],
Zhao, Z.Y.[Zhen-Yu],
Chen, Y.H.[Yu-Hao],
Tang, Y.[Yao],
Liang, J.J.[Jia-Jun],
Hidiffusion: Unlocking Higher-resolution Creativity and Efficiency in
Pretrained Diffusion Models,
ECCV24(LI: 145-161).
Springer DOI
2412
BibRef
Garibi, D.[Daniel],
Patashnik, O.[Or],
Voynov, A.[Andrey],
Averbuch-Elor, H.[Hadar],
Cohen-Or, D.[Daniel],
Renoise: Real Image Inversion Through Iterative Noising,
ECCV24(XIV: 395-413).
Springer DOI
2412
BibRef
Cao, Y.[Yu],
Gong, S.G.[Shao-Gang],
Few-shot Image Generation by Conditional Relaxing Diffusion Inversion,
ECCV24(LXXXIV: 20-37).
Springer DOI
2412
BibRef
Huang, R.[Runhui],
Cai, K.X.[Kai-Xin],
Han, J.H.[Jian-Hua],
Liang, X.D.[Xiao-Dan],
Pei, R.[Renjing],
Lu, G.S.[Guan-Song],
Xu, S.[Songcen],
Zhang, W.[Wei],
Xu, H.[Hang],
Layerdiff: Exploring Text-guided Multi-layered Composable Image
Synthesis via Layer-collaborative Diffusion Model,
ECCV24(LXXVI: 144-160).
Springer DOI
2412
BibRef
Brokman, J.[Jonathan],
Hofman, O.[Omer],
Vainshtein, R.[Roman],
Giloni, A.[Amit],
Shimizu, T.[Toshiya],
Singh, I.[Inderjeet],
Rachmil, O.[Oren],
Zolfi, A.[Alon],
Shabtai, A.[Asaf],
Unno, Y.[Yuki],
Kojima, H.[Hisashi],
Montrage: Monitoring Training for Attribution of Generative Diffusion
Models,
ECCV24(LXXV: 1-17).
Springer DOI
2412
BibRef
Desai, A.[Alakh],
Vasconcelos, N.M.[Nuno M.],
Improving Image Synthesis with Diffusion-negative Sampling,
ECCV24(LIII: 199-214).
Springer DOI
2412
BibRef
Zhang, M.Y.[Man-Yuan],
Song, G.L.[Guang-Lu],
Shi, X.Y.[Xiao-Yu],
Liu, Y.[Yu],
Li, H.S.[Hong-Sheng],
Three Things We Need to Know About Transferring Stable Diffusion to
Visual Dense Prediction Tasks,
ECCV24(XLII: 128-145).
Springer DOI
2412
BibRef
Huang, L.J.[Lin-Jiang],
Fang, R.Y.[Rong-Yao],
Zhang, A.[Aiping],
Song, G.L.[Guang-Lu],
Liu, S.[Si],
Liu, Y.[Yu],
Li, H.S.[Hong-Sheng],
FouriScale: A Frequency Perspective on Training-free High-resolution
Image Synthesis,
ECCV24(XII: 196-212).
Springer DOI
2412
Code:
WWW Link.
BibRef
Guan, S.[Shanyan],
Ge, Y.[Yanhao],
Tai, Y.[Ying],
Yang, J.[Jian],
Li, W.[Wei],
You, M.Y.[Ming-Yu],
Hybridbooth: Hybrid Prompt Inversion for Efficient Subject-driven
Generation,
ECCV24(IX: 403-419).
Springer DOI
2412
BibRef
Wu, Y.[Yi],
Li, Z.Q.[Zi-Qiang],
Zheng, H.L.[He-Liang],
Wang, C.Y.[Chao-Yue],
Li, B.[Bin],
Infinite-ID: Identity-preserved Personalization via ID-Semantics
Decoupling Paradigm,
ECCV24(VIII: 279-296).
Springer DOI
2412
BibRef
Butt, M.A.[Muhammad Atif],
Wang, K.[Kai],
Vazquez-Corral, J.[Javier],
van de Weijer, J.[Joost],
ColorPeel: Color Prompt Learning with Diffusion Models via Color and
Shape Disentanglement,
ECCV24(VII: 456-472).
Springer DOI
2412
Project:
WWW Link.
BibRef
Li, M.[Ming],
Yang, T.[Taojiannan],
Kuang, H.F.[Hua-Feng],
Wu, J.[Jie],
Wang, Z.N.[Zhao-Ning],
Xiao, X.F.[Xue-Feng],
Chen, C.[Chen],
Controlnet++: Improving Conditional Controls with Efficient Consistency
Feedback,
ECCV24(VII: 129-147).
Springer DOI
2412
Project:
WWW Link.
BibRef
Mu, J.[Jiteng],
Gharbi, M.[Michaël],
Zhang, R.[Richard],
Shechtman, E.[Eli],
Vasconcelos, N.M.[Nuno M.],
Wang, X.L.[Xiao-Long],
Park, T.[Taesung],
Editable Image Elements for Controllable Synthesis,
ECCV24(II: 39-56).
Springer DOI
2412
BibRef
Ning, W.X.[Wen-Xin],
Chang, D.L.[Dong-Liang],
Tong, Y.J.[Yu-Jun],
He, Z.J.[Zhong-Jiang],
Liang, K.M.[Kong-Ming],
Ma, Z.Y.[Zhan-Yu],
Hierarchical Prompting for Diffusion Classifiers,
ACCV24(VIII: 297-314).
Springer DOI
2412
Code:
WWW Link.
BibRef
Kim, G.[Gwanghyun],
Kim, H.[Hayeon],
Seo, H.[Hoigi],
Kang, D.U.[Dong Un],
Chun, S.Y.[Se Young],
Beyondscene: Higher-resolution Human-centric Scene Generation with
Pretrained Diffusion,
ECCV24(LXIV: 126-142).
Springer DOI
2412
BibRef
Wang, Y.L.[Yi-Lin],
Chen, Z.[Zeyuan],
Zhong, L.J.[Liang-Jun],
Ding, Z.[Zheng],
Tu, Z.W.[Zhuo-Wen],
Dolfin: Diffusion Layout Transformers Without Autoencoder,
ECCV24(LI: 326-343).
Springer DOI
2412
BibRef
Najdenkoska, I.[Ivona],
Sinha, A.[Animesh],
Dubey, A.[Abhimanyu],
Mahajan, D.[Dhruv],
Ramanathan, V.[Vignesh],
Radenovic, F.[Filip],
Context Diffusion: In-context Aware Image Generation,
ECCV24(LXXVII: 375-391).
Springer DOI
2412
BibRef
Ma, N.[Nanye],
Goldstein, M.[Mark],
Albergo, M.S.[Michael S.],
Boffi, N.M.[Nicholas M.],
Vanden-Eijnden, E.[Eric],
Xie, S.[Saining],
SIT: Exploring Flow and Diffusion-based Generative Models with Scalable
Interpolant Transformers,
ECCV24(LXXVII: 23-40).
Springer DOI
2412
BibRef
Zhang, D.J.H.[David Jun-Hao],
Xu, M.[Mutian],
Wu, J.Z.J.[Jay Zhang-Jie],
Xue, C.[Chuhui],
Zhang, W.Q.[Wen-Qing],
Han, X.G.[Xiao-Guang],
Bai, S.[Song],
Shou, M.Z.[Mike Zheng],
Free-atm: Harnessing Free Attention Masks for Representation Learning
on Diffusion-generated Images,
ECCV24(XL: 465-482).
Springer DOI
2412
BibRef
Yu, Z.M.[Zheng-Ming],
Dou, Z.Y.[Zhi-Yang],
Long, X.X.[Xiao-Xiao],
Lin, C.[Cheng],
Li, Z.K.[Ze-Kun],
Liu, Y.[Yuan],
Müller, N.[Norman],
Komura, T.[Taku],
Habermann, M.[Marc],
Theobalt, C.[Christian],
Li, X.[Xin],
Wang, W.P.[Wen-Ping],
SURF-D: Generating High-quality Surfaces of Arbitrary Topologies Using
Diffusion Models,
ECCV24(XXXIX: 419-438).
Springer DOI
2412
BibRef
Gandikota, R.[Rohit],
Materzynska, J.[Joanna],
Zhou, T.[Tingrui],
Torralba, A.[Antonio],
Bau, D.[David],
Concept Sliders: Lora Adaptors for Precise Control in Diffusion Models,
ECCV24(XL: 172-188).
Springer DOI
2412
BibRef
Iwai, S.[Shoma],
Osanai, A.[Atsuki],
Kitada, S.[Shunsuke],
Omachi, S.[Shinichiro],
Layout-corrector: Alleviating Layout Sticking Phenomenon in Discrete
Diffusion Model,
ECCV24(XXXIV: 92-110).
Springer DOI
2412
BibRef
Kong, Z.[Zhe],
Zhang, Y.[Yong],
Yang, T.Y.[Tian-Yu],
Wang, T.[Tao],
Zhang, K.H.[Kai-Hao],
Wu, B.[Bizhu],
Chen, G.Y.[Guan-Ying],
Liu, W.[Wei],
Luo, W.H.[Wen-Han],
OMG: Occlusion-friendly Personalized Multi-concept Generation in
Diffusion Models,
ECCV24(XXXI: 253-270).
Springer DOI
2412
BibRef
Lin, Z.H.[Zhi-Hang],
Lin, M.[Mingbao],
Zhao, M.[Meng],
Ji, R.R.[Rong-Rong],
Accdiffusion: An Accurate Method for Higher-resolution Image Generation,
ECCV24(VI: 38-53).
Springer DOI
2412
BibRef
Somepalli, G.[Gowthami],
Gupta, A.[Anubhav],
Gupta, K.[Kamal],
Palta, S.[Shramay],
Goldblum, M.[Micah],
Geiping, J.[Jonas],
Shrivastava, A.[Abhinav],
Goldstein, T.[Tom],
Investigating Style Similarity in Diffusion Models,
ECCV24(LXVI: 143-160).
Springer DOI
2412
BibRef
Qi, Z.[Zipeng],
Huang, G.X.[Guo-Xi],
Liu, C.Y.[Chen-Yang],
Ye, F.[Fei],
Layered Rendering Diffusion Model for Controllable Zero-Shot Image
Synthesis,
ECCV24(LXVI: 426-443).
Springer DOI
2412
BibRef
Chen, Z.K.[Zhe-Kai],
Wang, W.[Wen],
Yang, Z.[Zhen],
Yuan, Z.[Zeqing],
Chen, H.[Hao],
Shen, C.H.[Chun-Hua],
Freecompose: Generic Zero-Shot Image Composition with Diffusion Prior,
ECCV24(XVII: 70-87).
Springer DOI
2412
BibRef
Ju, X.[Xuan],
Liu, X.[Xian],
Wang, X.[Xintao],
Bian, Y.X.[Yu-Xuan],
Shan, Y.[Ying],
Xu, Q.[Qiang],
Brushnet: A Plug-and-play Image Inpainting Model with Decomposed
Dual-branch Diffusion,
ECCV24(XX: 150-168).
Springer DOI
2412
BibRef
Lin, C.H.[Chieh Hubert],
Kim, C.[Changil],
Huang, J.B.[Jia-Bin],
Li, Q.[Qinbo],
Ma, C.Y.[Chih-Yao],
Kopf, J.[Johannes],
Yang, M.H.[Ming-Hsuan],
Tseng, H.Y.[Hung-Yu],
Taming Latent Diffusion Model for Neural Radiance Field Inpainting,
ECCV24(III: 149-165).
Springer DOI
2412
BibRef
Gao, H.A.[Huan-Ang],
Gao, M.J.[Ming-Ju],
Li, J.[Jiaju],
Li, W.[Wenyi],
Zhi, R.[Rong],
Tang, H.[Hao],
Zhao, H.[Hao],
SCP-Diff: Spatial-categorical Joint Prior for Diffusion Based Semantic
Image Synthesis,
ECCV24(XXXII: 37-54).
Springer DOI
2412
BibRef
Le, M.Q.[Minh-Quan],
Graikos, A.[Alexandros],
Yellapragada, S.[Srikar],
Gupta, R.[Rajarsi],
Saltz, J.[Joel],
Samaras, D.[Dimitris],
inf-brush: Controllable Large Image Synthesis with Diffusion Models in
Infinite Dimensions,
ECCV24(XXXII: 385-401).
Springer DOI
2412
BibRef
Gong, C.[Chao],
Chen, K.[Kai],
Wei, Z.P.[Zhi-Peng],
Chen, J.J.[Jing-Jing],
Jiang, Y.G.[Yu-Gang],
Reliable and Efficient Concept Erasure of Text-to-image Diffusion
Models,
ECCV24(LIII: 73-88).
Springer DOI
2412
BibRef
Luo, J.J.[Jian-Jie],
Chen, J.W.[Jing-Wen],
Li, Y.[Yehao],
Pan, Y.W.[Ying-Wei],
Feng, J.L.[Jian-Lin],
Chao, H.Y.[Hong-Yang],
Yao, T.[Ting],
Unleashing Text-to-image Diffusion Prior for Zero-shot Image Captioning,
ECCV24(LVII: 237-254).
Springer DOI
2412
BibRef
Wang, L.Z.[Luo-Zhou],
Shen, G.[Guibao],
Ge, W.[Wenhang],
Chen, G.Y.[Guang-Yong],
Li, Y.J.[Yi-Jun],
Chen, Y.[Yingcong],
Text-anchored Score Composition: Tackling Condition Misalignment in
Text-to-image Diffusion Models,
ECCV24(XLVII: 21-37).
Springer DOI
2412
BibRef
Lu, G.S.[Guan-Song],
Guo, Y.F.[Yuan-Fan],
Han, J.H.[Jian-Hua],
Niu, M.Z.[Min-Zhe],
Zeng, Y.[Yihan],
Xu, S.[Songcen],
Huang, Z.Y.[Ze-Yi],
Zhong, Z.[Zhao],
Zhang, W.[Wei],
Xu, H.[Hang],
Pangu-draw: Advancing Resource-efficient Text-to-image Synthesis with
Time-decoupled Training and Reusable Coop-diffusion,
ECCV24(XLV: 159-176).
Springer DOI
2412
BibRef
Huang, C.P.[Chi-Pin],
Chang, K.P.[Kai-Po],
Tsai, C.T.[Chung-Ting],
Lai, Y.H.[Yung-Hsuan],
Yang, F.E.[Fu-En],
Wang, Y.C.A.F.[Yu-Chi-Ang Frank],
Receler: Reliable Concept Erasing of Text-to-image Diffusion Models via
Lightweight Erasers,
ECCV24(XL: 360-376).
Springer DOI
2412
BibRef
Zhang, Y.[Yasi],
Yu, P.[Peiyu],
Wu, Y.N.[Ying Nian],
Object-conditioned Energy-based Attention Map Alignment in
Text-to-image Diffusion Models,
ECCV24(XLII: 55-71).
Springer DOI
2412
BibRef
Chai, W.L.[Wei-Long],
Zheng, D.D.[Dan-Dan],
Cao, J.J.[Jia-Jiong],
Chen, Z.Q.[Zhi-Quan],
Wang, C.[Changbao],
Ma, C.G.[Chen-Guang],
Speedupnet: A Plug-and-play Adapter Network for Accelerating
Text-to-image Diffusion Models,
ECCV24(XLIII: 181-196).
Springer DOI
2412
BibRef
Zhang, Y.[Yi],
Tang, Y.[Yun],
Ruan, W.J.[Wen-Jie],
Huang, X.W.[Xiao-Wei],
Khastgir, S.[Siddartha],
Jennings, P.[Paul],
Zhao, X.Y.[Xing-Yu],
Protip: Probabilistic Robustness Verification on Text-to-image
Diffusion Models Against Stochastic Perturbation,
ECCV24(XXXII: 455-472).
Springer DOI
2412
BibRef
Nair, N.G.[Nithin Gopalakrishnan],
Valanarasu, J.M.J.[Jeya Maria Jose],
Patel, V.M.[Vishal M.],
Maxfusion: Plug&play Multi-modal Generation in Text-to-Image Diffusion
Models,
ECCV24(XXXVIII: 93-110).
Springer DOI
2412
BibRef
Zhang, Z.[Zhengbo],
Xu, L.[Li],
Peng, D.[Duo],
Rahmani, H.[Hossein],
Liu, J.[Jun],
Diff-tracker: Text-to-image Diffusion Models are Unsupervised Trackers,
ECCV24(XXVIII: 319-337).
Springer DOI
2412
BibRef
Motamed, S.[Saman],
Paudel, D.P.[Danda Pani],
Van Gool, L.J.[Luc J.],
Lego: Learning to Disentangle and Invert Personalized Concepts Beyond
Object Appearance in Text-to-image Diffusion Models,
ECCV24(XV: 116-133).
Springer DOI
2412
BibRef
Kong, H.Y.[Han-Yang],
Lian, D.Z.[Dong-Ze],
Mi, M.B.[Michael Bi],
Wang, X.C.[Xin-Chao],
Dreamdrone: Text-to-image Diffusion Models Are Zero-shot Perpetual View
Generators,
ECCV24(XIII: 324-341).
Springer DOI
2412
BibRef
Peng, D.[Duo],
Zhang, Z.[Zhengbo],
Hu, P.[Ping],
Ke, Q.H.[Qiu-Hong],
Yau, D.K.Y.[David K. Y.],
Liu, J.[Jun],
Harnessing Text-to-image Diffusion Models for Category-agnostic Pose
Estimation,
ECCV24(XIII: 342-360).
Springer DOI
2412
BibRef
Zhao, T.C.[Tian-Chen],
Ning, X.F.[Xue-Fei],
Fang, T.[Tongcheng],
Liu, E.[Enshu],
Huang, G.[Guyue],
Lin, Z.[Zinan],
Yan, S.[Shengen],
Dai, G.H.[Guo-Hao],
Wang, Y.[Yu],
Mixdq: Memory-efficient Few-step Text-to-image Diffusion Models with
Metric-decoupled Mixed Precision Quantization,
ECCV24(XIV: 285-302).
Springer DOI
2412
BibRef
Gao, Y.[Yi],
Psg-adapter: Controllable Planning Scene Graph for Improving
Text-to-image Diffusion,
ACCV24(V: 205-221).
Springer DOI
2412
BibRef
Gupta, P.[Parul],
Hayat, M.[Munawar],
Dhall, A.[Abhinav],
Do, T.T.[Thanh-Toan],
Conditional Distribution Modelling for Few-shot Image Synthesis with
Diffusion Models,
ACCV24(V: 3-20).
Springer DOI
2412
BibRef
Yang, D.[Danni],
Dong, R.H.[Ruo-Han],
Ji, J.Y.[Jia-Yi],
Ma, Y.W.[Yi-Wei],
Wang, H.[Haowei],
Sun, X.S.[Xiao-Shuai],
Ji, R.R.[Rong-Rong],
Exploring Phrase-level Grounding with Text-to-image Diffusion Model,
ECCV24(LIII: 161-180).
Springer DOI
2412
BibRef
Ren, J.[Jie],
Li, Y.X.[Ya-Xin],
Zeng, S.[Shenglai],
Xu, H.[Han],
Lyu, L.J.[Ling-Juan],
Xing, Y.[Yue],
Tang, J.[Jiliang],
Unveiling and Mitigating Memorization in Text-to-image Diffusion Models
Through Cross Attention,
ECCV24(LXXVII: 340-356).
Springer DOI
2412
BibRef
Zheng, W.[Wendi],
Teng, J.Y.[Jia-Yan],
Yang, Z.[Zhuoyi],
Wang, W.H.[Wei-Han],
Chen, J.[Jidong],
Gu, X.T.[Xiao-Tao],
Dong, Y.X.[Yu-Xiao],
Ding, M.[Ming],
Tang, J.[Jie],
Cogview3: Finer and Faster Text-to-image Generation via Relay Diffusion,
ECCV24(LXXVII: 1-22).
Springer DOI
2412
BibRef
Zhao, J.[Juntu],
Deng, J.Y.[Jun-Yu],
Ye, Y.X.[Yi-Xin],
Li, C.X.[Chong-Xuan],
Deng, Z.J.[Zhi-Jie],
Wang, D.[Dequan],
Lost in Translation: Latent Concept Misalignment in Text-to-image
Diffusion Models,
ECCV24(LXIX: 318-333).
Springer DOI
2412
BibRef
Hui, X.F.[Xiao-Fei],
Wu, Q.[Qian],
Rahmani, H.[Hossein],
Liu, J.[Jun],
Class-agnostic Object Counting with Text-to-image Diffusion Model,
ECCV24(LXIX: 1-18).
Springer DOI
2412
BibRef
Ma, J.[Jian],
Chen, C.[Chen],
Xie, Q.S.[Qing-Song],
Lu, H.[Haonan],
Pea-diffusion: Parameter-efficient Adapter with Knowledge Distillation
in Non-english Text-to-image Generation,
ECCV24(LXVIII: 89-105).
Springer DOI
2412
BibRef
Kim, S.[Sanghyun],
Jung, S.[Seohyeon],
Kim, B.[Balhae],
Choi, M.[Moonseok],
Shin, J.[Jinwoo],
Lee, J.H.[Ju-Ho],
Safeguard Text-to-image Diffusion Models with Human Feedback Inversion,
ECCV24(LXVII: 128-145).
Springer DOI
2412
BibRef
Biggs, B.[Benjamin],
Seshadri, A.[Arjun],
Zou, Y.[Yang],
Jain, A.[Achin],
Golatkar, A.[Aditya],
Xie, Y.S.[Yu-Sheng],
Achille, A.[Alessandro],
Swaminathan, A.[Ashwin],
Soatto, S.[Stefano],
Diffusion Soup: Model Merging for Text-to-image Diffusion Models,
ECCV24(LXIII: 257-274).
Springer DOI
2412
BibRef
Zhao, Y.[Yang],
Xu, Y.[Yanwu],
Xiao, Z.S.[Zhi-Sheng],
Jia, H.L.[Hao-Lin],
Hou, T.B.[Ting-Bo],
Mobilediffusion: Instant Text-to-image Generation on Mobile Devices,
ECCV24(LXII: 225-242).
Springer DOI
2412
BibRef
Zhang, Y.[Yang],
Tzun, T.T.[Teoh Tze],
Hern, L.W.[Lim Wei],
Kawaguchi, K.[Kenji],
Enhancing Semantic Fidelity in Text-to-image Synthesis: Attention
Regulation in Diffusion Models,
ECCV24(LXXXVI: 70-86).
Springer DOI
2412
BibRef
Wang, Z.Q.[Zhong-Qi],
Zhang, J.[Jie],
Shan, S.G.[Shi-Guang],
Chen, X.L.[Xi-Lin],
T2ishield: Defending Against Backdoors on Text-to-image Diffusion
Models,
ECCV24(LXXXV: 107-124).
Springer DOI
2412
BibRef
Kim, C.[Changhoon],
Min, K.[Kyle],
Yang, Y.Z.[Ye-Zhou],
R.A.C.E.: Robust Adversarial Concept Erasure for Secure Text-to-image
Diffusion Model,
ECCV24(LXXXIII: 461-478).
Springer DOI
2412
BibRef
Wu, X.[Xiaoshi],
Hao, Y.M.[Yi-Ming],
Zhang, M.[Manyuan],
Sun, K.Q.[Ke-Qiang],
Huang, Z.Y.[Zhao-Yang],
Song, G.L.[Guang-Lu],
Liu, Y.[Yu],
Li, H.S.[Hong-Sheng],
Deep Reward Supervisions for Tuning Text-to-image Diffusion Models,
ECCV24(LXXXIII: 108-124).
Springer DOI
2412
BibRef
Parihar, R.[Rishubh],
Sachidanand, V.S.,
Mani, S.[Sabariswaran],
Karmali, T.[Tejan],
Babu, R.V.[R. Venkatesh],
Precisecontrol: Enhancing Text-to-image Diffusion Models with
Fine-grained Attribute Control,
ECCV24(LXXXII: 469-487).
Springer DOI
2412
BibRef
Choi, D.W.[Dae-Won],
Jeong, J.[Jongheon],
Jang, H.[Huiwon],
Shin, J.[Jinwoo],
Adversarial Robustification via Text-to-image Diffusion Models,
ECCV24(LXXXI: 158-177).
Springer DOI
2412
BibRef
Zavadski, D.[Denis],
Feiden, J.F.[Johann-Friedrich],
Rother, C.[Carsten],
Controlnet-xs: Rethinking the Control of Text-to-image Diffusion Models
as Feedback-control Systems,
ECCV24(LXXXVIII: 343-362).
Springer DOI
2412
BibRef
Zhao, Y.P.[Ya-Ping],
Zhang, P.[Pei],
Wang, C.[Chutian],
Lam, E.Y.[Edmund Y.],
Controllable Unsupervised Event-Based Video Generation,
ICIP24(2278-2284)
IEEE DOI Code:
WWW Link.
2411
Training, Codes, Image edge detection, Cameras,
Diffusion models, neuromorphic imaging, computational imaging
BibRef
Qazi, T.[Tayeba],
Lall, B.[Brejesh],
Thermal Videodiff (TVD):
A Diffusion Architecture for Thermal Video Synthesis,
ICIP24(2438-2444)
IEEE DOI Code:
WWW Link.
2411
Deep learning, Temperature distribution, Costs, Infrared imaging,
Thermal sensors, Diffusion models, Synthetic Video Generation,
Visible Spectrum Context
BibRef
Maung-Maung, A.P.[April-Pyone],
Nguyen, H.H.[Huy H.],
Kiya, H.[Hitoshi],
Echizen, I.[Isao],
Fine-Tuning Text-To-Image Diffusion Models for Class-Wise Spurious
Feature Generation,
ICIP24(3910-3916)
IEEE DOI
2411
Text to image, Flowering plants, Diffusion models,
Feature extraction, Information filters, Internet, Testing, finetuning
BibRef
Hudson, D.A.[Drew A.],
Zoran, D.[Daniel],
Malinowski, M.[Mateusz],
Lampinen, A.K.[Andrew K.],
Jaegle, A.[Andrew],
McClelland, J.L.[James L.],
Matthey, L.[Loic],
Hill, F.[Felix],
Lerchner, A.[Alexander],
SODA: Bottleneck Diffusion Models for Representation Learning,
CVPR24(23115-23127)
IEEE DOI
2410
Representation learning, Training, Visualization, Image synthesis,
Semantics, Noise reduction, Self-supervised learning, classification
BibRef
Karras, T.[Tero],
Aittala, M.[Miika],
Lehtinen, J.[Jaakko],
Hellsten, J.[Janne],
Aila, T.[Timo],
Laine, S.[Samuli],
Analyzing and Improving the Training Dynamics of Diffusion Models,
CVPR24(24174-24184)
IEEE DOI
2410
Training, Systematics, Costs, Image synthesis, Computer architecture,
Network architecture
BibRef
Li, J.[Jing],
Wang, Z.[Zigan],
Li, J.L.[Jin-Liang],
AdvDenoise: Fast Generation Framework of Universal and Robust
Adversarial Patches Using Denoise,
SAIAD24(3481-3490)
IEEE DOI Code:
WWW Link.
2410
Visualization, Computational modeling, Noise reduction,
Diffusion models, Transformers, Robustness
BibRef
Tang, S.[Siao],
Wang, X.[Xin],
Chen, H.[Hong],
Guan, C.[Chaoyu],
Wu, Z.[Zewen],
Tang, Y.S.[Yan-Song],
Zhu, W.W.[Wen-Wu],
Post-training Quantization with Progressive Calibration and Activation
Relaxing for Text-to-image Diffusion Models,
ECCV24(LVI: 404-420).
Springer DOI
2412
BibRef
Wang, C.Y.[Chang-Yuan],
Wang, Z.W.[Zi-Wei],
Xu, X.W.[Xiu-Wei],
Tang, Y.S.[Yan-Song],
Zhou, J.[Jie],
Lu, J.W.[Ji-Wen],
Towards Accurate Post-Training Quantization for Diffusion Models,
CVPR24(16026-16035)
IEEE DOI Code:
WWW Link.
2410
Quantization (signal), Risk minimization, Accuracy, Tensors,
Image synthesis, Diffusion models, Minimization, diffusion model,
network quantization
BibRef
Islam, K.[Khawar],
Zaheer, M.Z.[Muhammad Zaigham],
Mahmood, A.[Arif],
Nandakumar, K.[Karthik],
Diffusemix: Label-Preserving Data Augmentation with Diffusion Models,
CVPR24(27611-27620)
IEEE DOI Code:
WWW Link.
2410
Training, Performance gain, Diffusion models, Data augmentation,
Robustness, Image augmentation, Fractals, data augmentation, cutmix
BibRef
Miao, Z.C.[Zi-Chen],
Wang, J.[Jiang],
Wang, Z.[Ze],
Yang, Z.Y.[Zheng-Yuan],
Wang, L.J.[Li-Juan],
Qiu, Q.[Qiang],
Liu, Z.C.[Zi-Cheng],
Training Diffusion Models Towards Diverse Image Generation with
Reinforcement Learning,
CVPR24(10844-10853)
IEEE DOI
2410
Training, Gradient methods, Limiting, Image synthesis, Estimation,
Diffusion processes, Reinforcement learning
BibRef
Shabani, M.A.[Mohammad Amin],
Wang, Z.W.[Zhao-Wen],
Liu, D.[Difan],
Zhao, N.X.[Nan-Xuan],
Yang, J.[Jimei],
Furukawa, Y.[Yasutaka],
Visual Layout Composer: Image-Vector Dual Diffusion Model for Design
Layout Generation,
CVPR24(9222-9231)
IEEE DOI Code:
WWW Link.
2410
Visualization, Computational modeling, Layout, Diffusion models,
Controllability, Vectors
BibRef
Qian, Y.R.[Yu-Rui],
Cai, Q.[Qi],
Pan, Y.W.[Ying-Wei],
Li, Y.[Yehao],
Yao, T.[Ting],
Sun, Q.[Qibin],
Mei, T.[Tao],
Boosting Diffusion Models with Moving Average Sampling in Frequency
Domain,
CVPR24(8911-8920)
IEEE DOI
2410
Schedules, Image synthesis, Frequency-domain analysis,
Noise reduction, Diffusion processes, Diffusion models, image generation
BibRef
Yang, K.[Kai],
Tao, J.[Jian],
Lyu, J.[Jiafei],
Ge, C.J.[Chun-Jiang],
Chen, J.X.[Jia-Xin],
Shen, W.H.[Wei-Han],
Zhu, X.L.[Xiao-Long],
Li, X.[Xiu],
Using Human Feedback to Fine-tune Diffusion Models without Any Reward
Model,
CVPR24(8941-8951)
IEEE DOI Code:
WWW Link.
2410
Training, Analytical models, Image coding, Computational modeling,
Noise reduction, Graphics processing units, Diffusion models,
Human feedback
BibRef
Zhu, R.[Rui],
Pan, Y.W.[Ying-Wei],
Li, Y.[Yehao],
Yao, T.[Ting],
Sun, Z.L.[Zheng-Long],
Mei, T.[Tao],
Chen, C.W.[Chang Wen],
SD-DiT: Unleashing the Power of Self-Supervised Discrimination in
Diffusion Transformer*,
CVPR24(8435-8445)
IEEE DOI
2410
Training, Image synthesis, Noise, Diffusion processes,
Ordinary differential equations, Transformers, self-supervised learning
BibRef
Zhou, Z.Y.[Zhen-Yu],
Chen, D.[Defang],
Wang, C.[Can],
Chen, C.[Chun],
Fast ODE-based Sampling for Diffusion Models in Around 5 Steps,
CVPR24(7777-7786)
IEEE DOI Code:
WWW Link.
2410
Degradation, Image resolution, Image synthesis,
Ordinary differential equations, Diffusion models, Fast Sampling
BibRef
Lee, H.Y.[Hsin-Ying],
Tseng, H.Y.[Hung-Yu],
Lee, H.Y.[Hsin-Ying],
Yang, M.H.[Ming-Hsuan],
Exploiting Diffusion Prior for Generalizable Dense Prediction,
CVPR24(7861-7871)
IEEE DOI Code:
WWW Link.
2410
Adaptation models, Visualization, Training data,
Stochastic processes, Estimation, Diffusion processes,
image generation
BibRef
Zhang, K.W.[Kai-Wen],
Zhou, Y.F.[Yi-Fan],
Xu, X.D.[Xu-Dong],
Dai, B.[Bo],
Pan, X.G.[Xin-Gang],
DiffMorpher: Unleashing the Capability of Diffusion Models for Image
Morphing,
CVPR24(7912-7921)
IEEE DOI
2410
Interpolation, Schedules, Image synthesis, Semantics, Image morphing,
Noise, Fitting, Diffusion models, Image morphing, video generation
BibRef
Li, M.Y.[Mu-Yang],
Cai, T.[Tianle],
Cao, J.X.[Jia-Xin],
Zhang, Q.S.[Qin-Sheng],
Cai, H.[Han],
Bai, J.J.[Jun-Jie],
Jia, Y.Q.[Yang-Qing],
Li, K.[Kai],
Han, S.[Song],
DistriFusion: Distributed Parallel Inference for High-Resolution
Diffusion Models,
CVPR24(7183-7193)
IEEE DOI
2410
Degradation, Computational modeling, Graphics processing units,
Diffusion processes, Parallel processing, Diffusion models, generative-ai
BibRef
Koley, S.[Subhadeep],
Bhunia, A.K.[Ayan Kumar],
Sekhri, D.[Deeptanshu],
Sain, A.[Aneeshan],
Chowdhury, P.N.[Pinaki Nath],
Xiang, T.[Tao],
Song, Y.Z.[Yi-Zhe],
It's All About Your Sketch: Democratising Sketch Control in Diffusion
Models,
CVPR24(7204-7214)
IEEE DOI
2410
Adaptation models, Adaptive systems, Navigation, Generative AI,
Image retrieval, Process control, Streaming media
BibRef
Wang, Y.[Yibo],
Gao, R.Y.[Rui-Yuan],
Chen, K.[Kai],
Zhou, K.Q.[Kai-Qiang],
Cai, Y.J.[Ying-Jie],
Hong, L.Q.[Lan-Qing],
Li, Z.G.[Zhen-Guo],
Jiang, L.H.[Li-Hui],
Yeung, D.Y.[Dit-Yan],
Xu, Q.[Qiang],
Zhang, K.[Kai],
DetDiffusion: Synergizing Generative and Perceptive Models for
Enhanced Data Generation and Perception,
CVPR24(7246-7255)
IEEE DOI
2410
Image segmentation, Image recognition, Image synthesis,
Training data, Object detection, Diffusion models, Data augmentation
BibRef
Zhang, P.Z.[Peng-Ze],
Yin, H.[Hubery],
Li, C.[Chen],
Xie, X.H.[Xiao-Hua],
Tackling the Singularities at the Endpoints of Time Intervals in
Diffusion Models,
CVPR24(6945-6954)
IEEE DOI
2410
Training, Brightness, Gaussian distribution, Diffusion models,
Diffusion Model, Generative Model, Singularity
BibRef
Hong, S.[Seongmin],
Lee, K.[Kyeonghyun],
Jeon, S.Y.[Suh Yoon],
Bae, H.[Hyewon],
Chun, S.Y.[Se Young],
On Exact Inversion of DPM-Solvers,
CVPR24(7069-7078)
IEEE DOI
2410
Noise, Noise reduction, Watermarking, Diffusion models, Robustness,
Diffusion, Inversion, DPM-Solver
BibRef
Fu, B.[Bin],
Yu, F.[Fanghua],
Liu, A.[Anran],
Wang, Z.X.[Zi-Xuan],
Wen, J.[Jie],
He, J.J.[Jun-Jun],
Qiao, Y.[Yu],
Generate Like Experts: Multi-Stage Font Generation by Incorporating
Font Transfer Process into Diffusion Models,
CVPR24(6892-6901)
IEEE DOI Code:
WWW Link.
2410
Costs, Noise, Diffusion processes, Transforms, Manuals,
Diffusion models, Generative adversarial networks,
Probabilistic Generative Model
BibRef
Deng, F.[Fei],
Wang, Q.F.[Qi-Fei],
Wei, W.[Wei],
Hou, T.B.[Ting-Bo],
Grundmann, M.[Matthias],
PRDP: Proximal Reward Difference Prediction for Large-Scale Reward
Finetuning of Diffusion Models,
CVPR24(7423-7433)
IEEE DOI
2410
Training, Technological innovation, Closed box,
Reinforcement learning, Diffusion models, RLHF
BibRef
Du, R.[Ruoyi],
Chang, D.L.[Dong-Liang],
Hospedales, T.[Timothy],
Song, Y.Z.[Yi-Zhe],
Ma, Z.Y.[Zhan-Yu],
DemoFusion: Democratising High-Resolution Image Generation With No $$,
CVPR24(6159-6168)
IEEE DOI
2410
Training, Image resolution, Image synthesis, Generative AI,
Semantics, Memory management, Image Generation, Diffusion Model, High-resolution
BibRef
Wang, H.J.[Hong-Jie],
Liu, D.[Difan],
Kang, Y.[Yan],
Li, Y.J.[Yi-Jun],
Lin, Z.[Zhe],
Jha, N.K.[Niraj K.],
Liu, Y.C.[Yu-Chen],
Attention-Driven Training-Free Efficiency Enhancement of Diffusion
Models,
CVPR24(16080-16089)
IEEE DOI Code:
WWW Link.
2410
Image quality, Schedules, Costs, Convolution, Computational modeling,
Noise reduction, diffusion model, training-free, efficiency,
attention map
BibRef
Chen, H.X.[Hao-Xin],
Zhang, Y.[Yong],
Cun, X.D.[Xiao-Dong],
Xia, M.H.[Meng-Han],
Wang, X.[Xintao],
Weng, C.[Chao],
Shan, Y.[Ying],
VideoCrafter2: Overcoming Data Limitations for High-Quality Video
Diffusion Models,
CVPR24(7310-7320)
IEEE DOI
2410
Training, Couplings, Degradation, Analytical models, Noise, Diffusion models
BibRef
Kang, J.[Junoh],
Choi, J.[Jinyoung],
Choi, S.[Sungik],
Han, B.H.[Bo-Hyung],
Observation-Guided Diffusion Probabilistic Models,
CVPR24(8323-8331)
IEEE DOI Code:
WWW Link.
2410
Training, Accuracy, Computational modeling, Noise reduction,
Quality control, Diffusion models, Robustness, generative models,
diffusion models
BibRef
Zhou, J.X.[Jin-Xin],
Ding, T.Y.[Tian-Yu],
Chen, T.Y.[Tian-Yi],
Jiang, J.C.[Jia-Chen],
Zharkov, I.[Ilya],
Zhu, Z.H.[Zhi-Hui],
Liang, L.[Luming],
DREAM: Diffusion Rectification and Estimation-Adaptive Models,
CVPR24(8342-8351)
IEEE DOI
2410
Training, Image quality, Navigation, Source coding, Superresolution,
Estimation, Distortion
BibRef
Chen, C.[Chen],
Liu, D.[Daochang],
Xu, C.[Chang],
Towards Memorization-Free Diffusion Models,
CVPR24(8425-8434)
IEEE DOI
2410
Image quality, Training, Measurement, Refining, Noise reduction,
Training data, Reliability theory, Diffusion Models, Memorization
BibRef
Qi, L.[Lu],
Yang, L.[Lehan],
Guo, W.D.[Wei-Dong],
Xu, Y.[Yu],
Du, B.[Bo],
Jampani, V.[Varun],
Yang, M.H.[Ming-Hsuan],
UniGS: Unified Representation for Image Generation and Segmentation,
CVPR24(6305-6315)
IEEE DOI
2410
Training, Image segmentation, Image synthesis,
Image color analysis, Pipelines, Training data, Transforms, diffusion
BibRef
Wang, L.Z.[Le-Zhong],
Frisvad, J.R.[Jeppe Revall],
Jensen, M.B.[Mark Bo],
Bigdeli, S.A.[Siavash Arjomand],
StereoDiffusion: Training-Free Stereo Image Generation Using Latent
Diffusion Models,
GCV24(7416-7425)
IEEE DOI
2410
Image quality, Image synthesis, Extended reality, Pipelines,
Noise reduction, Diffusion models, Deep Image/Video Synthesis,
Stable Diffusion
BibRef
Sharma, N.[Nakul],
Tripathi, A.[Aditay],
Chakraborty, A.[Anirban],
Mishra, A.[Anand],
Sketch-guided Image Inpainting with Partial Discrete Diffusion
Process,
NTIRE24(6024-6034)
IEEE DOI Code:
WWW Link.
2410
Visualization, Shape, Semantics, Diffusion processes, Text to image,
Transformers
BibRef
Guo, J.Y.[Jia-Yi],
Xu, X.Q.[Xing-Qian],
Pu, Y.F.[Yi-Fan],
Ni, Z.[Zanlin],
Wang, C.F.[Chao-Fei],
Vasu, M.[Manushree],
Song, S.[Shiji],
Huang, G.[Gao],
Shi, H.[Humphrey],
Smooth Diffusion: Crafting Smooth Latent Spaces in Diffusion Models,
CVPR24(7548-7558)
IEEE DOI Code:
WWW Link.
2410
Training, Measurement, Interpolation, Visualization, Fluctuations,
Perturbation methods, Text to image
BibRef
Lyu, M.Y.[Meng-Yao],
Yang, Y.H.[Yu-Hong],
Hong, H.[Haiwen],
Chen, H.[Hui],
Jin, X.[Xuan],
He, Y.[Yuan],
Xue, H.[Hui],
Han, J.G.[Jun-Gong],
Ding, G.[Guiguang],
One-dimensional Adapter to Rule Them All: Concepts, Diffusion Models
and Erasing Applications,
CVPR24(7559-7568)
IEEE DOI Code:
WWW Link.
2410
Deformable models, Adaptation models, Costs, Deformation,
Text to image, Diffusion models, Permeability, Diffusion Models, Concept Erasing
BibRef
Yang, L.[Ling],
Qian, H.T.[Hao-Tian],
Zhang, Z.L.[Zhi-Ling],
Liu, J.W.[Jing-Wei],
Cui, B.[Bin],
Structure-Guided Adversarial Training of Diffusion Models,
CVPR24(7256-7266)
IEEE DOI
2410
Training, Manifolds, Image synthesis, Noise reduction, Text to image,
Diffusion models, Data models, Diffusion models, generative models,
Image generation
BibRef
Yu, Y.Y.[Yu-Yang],
Liu, B.Z.[Bang-Zhen],
Zheng, C.X.[Chen-Xi],
Xu, X.M.[Xue-Miao],
He, S.F.[Sheng-Feng],
Zhang, H.D.[Huai-Dong],
Beyond Textual Constraints: Learning Novel Diffusion Conditions with
Fewer Examples,
CVPR24(7109-7118)
IEEE DOI Code:
WWW Link.
2410
Training, Adaptation models, Codes, Text to image,
Diffusion processes, Diffusion models,
diffusion model
BibRef
Xing, X.[Ximing],
Zhou, H.T.[Hai-Tao],
Wang, C.[Chuang],
Zhang, J.[Jing],
Xu, D.[Dong],
Yu, Q.[Qian],
SVGDreamer: Text Guided SVG Generation with Diffusion Model,
CVPR24(4546-4555)
IEEE DOI Code:
WWW Link.
2410
Visualization, Image color analysis, Shape, Text to image,
Process control, Diffusion models, vector graphics, SVG, text-to-svg,
Diffusion
BibRef
Parihar, R.[Rishubh],
Bhat, A.[Abhijnya],
Basu, A.[Abhipsa],
Mallick, S.[Saswat],
Kundu, J.N.[Jogendra Nath],
Babu, R.V.[R. Venkatesh],
Balancing Act: Distribution-Guided Debiasing in Diffusion Models,
CVPR24(6668-6678)
IEEE DOI
2410
Training, Image synthesis, Semantics, Noise reduction, Text to image,
Diffusion models, Data augmentation, Debiasing, diffusion models,
generative models
BibRef
Ren, J.W.[Jia-Wei],
Xu, M.M.[Meng-Meng],
Wu, J.C.[Jui-Chieh],
Liu, Z.W.[Zi-Wei],
Xiang, T.[Tao],
Toisoul, A.[Antoine],
Move Anything with Layered Scene Diffusion,
CVPR24(6380-6389)
IEEE DOI
2410
Codes, Layout, Noise reduction, Memory management, Text to image,
Process control
BibRef
Lu, Y.Z.[Yan-Zuo],
Zhang, M.[Manlin],
Ma, A.J.[Andy J.],
Xie, X.H.[Xiao-Hua],
Lai, J.H.[Jian-Huang],
Coarse-to-Fine Latent Diffusion for Pose-Guided Person Image
Synthesis,
CVPR24(6420-6429)
IEEE DOI Code:
WWW Link.
2410
Training, Image synthesis, Semantics, Text to image, Process control,
Diffusion models, Generators, Diffusion Model, Person Image Synthesis
BibRef
Liu, C.[Chang],
Wu, H.N.[Hao-Ning],
Zhong, Y.J.[Yu-Jie],
Zhang, X.Y.[Xiao-Yun],
Wang, Y.F.[Yan-Feng],
Xie, W.[Weidi],
Intelligent Grimm: Open-ended Visual Storytelling via Latent
Diffusion Models,
CVPR24(6190-6200)
IEEE DOI Code:
WWW Link.
2410
Visualization, Electronic publishing, Computational modeling,
Pipelines, Text to image, Image sequences,
Diffusion Models
BibRef
Wimbauer, F.[Felix],
Wu, B.[Bichen],
Schoenfeld, E.[Edgar],
Dai, X.L.[Xiao-Liang],
Hou, J.[Ji],
He, Z.J.[Zi-Jian],
Sanakoyeu, A.[Artsiom],
Zhang, P.Z.[Pei-Zhao],
Tsai, S.[Sam],
Kohler, J.[Jonas],
Rupprecht, C.[Christian],
Cremers, D.[Daniel],
Vajda, P.[Peter],
Wang, J.L.[Jia-Liang],
Cache Me if You Can: Accelerating Diffusion Models through Block
Caching,
CVPR24(6211-6220)
IEEE DOI
2410
Image quality, Visualization, Schedules, Image synthesis,
Computational modeling, Noise reduction, Noise, diffusion,
fid
BibRef
Dalva, Y.[Yusuf],
Yanardag, P.[Pinar],
NoiseCLR: A Contrastive Learning Approach for Unsupervised Discovery
of Interpretable Directions in Diffusion Models,
CVPR24(24209-24218)
IEEE DOI
2410
Image synthesis, Computational modeling, Semantics, Text to image,
Contrastive learning, Aerospace electronics, Diffusion models,
semantic discovery
BibRef
Sun, H.[Haoze],
Li, W.B.[Wen-Bo],
Liu, J.Z.[Jian-Zhuang],
Chen, H.Y.[Hao-Yu],
Pei, R.[Renjing],
Zou, X.[Xueyi],
Yan, Y.[Youliang],
Yang, Y.[Yujiu],
CoSeR: Bridging Image and Language for Cognitive Super-Resolution,
CVPR24(25868-25878)
IEEE DOI Code:
WWW Link.
2410
Computational modeling, Superresolution, Semantics, Text to image,
Benchmark testing, Diffusion models
BibRef
Wang, Z.C.[Zhi-Cai],
Wei, L.H.[Long-Hui],
Wang, T.[Tan],
Chen, H.Y.[He-Yu],
Hao, Y.B.[Yan-Bin],
Wang, X.[Xiang],
He, X.N.[Xiang-Nan],
Tian, Q.[Qi],
Enhance Image Classification via Inter-Class Image Mixup with
Diffusion Model,
CVPR24(17223-17233)
IEEE DOI Code:
WWW Link.
2410
Training, Computational modeling, Text to image, Data augmentation,
Diffusion models, diffusion model, data augmentation
BibRef
Hsiao, Y.T.[Yi-Ting],
Khodadadeh, S.[Siavash],
Duarte, K.[Kevin],
Lin, W.A.[Wei-An],
Qu, H.[Hui],
Kwon, M.[Mingi],
Kalarot, R.[Ratheesh],
Plug-and-Play Diffusion Distillation,
CVPR24(13743-13752)
IEEE DOI
2410
Training, Visualization, Image synthesis, Computational modeling,
Text to image, Diffusion processes, distillation, model efficiency,
diffusion model
BibRef
Zhan, C.[Chenlu],
Lin, Y.[Yu],
Wang, G.[Gaoang],
Wang, H.W.[Hong-Wei],
Wu, J.[Jian],
MedM2G: Unifying Medical Multi-Modal Generation via Cross-Guided
Diffusion with Visual Invariant,
CVPR24(11502-11512)
IEEE DOI
2410
Visualization, Adaptation models, Technological innovation,
Magnetic resonance imaging, Text to image, Medical services,
Diffusion Model
BibRef
Kant, Y.[Yash],
Siarohin, A.[Aliaksandr],
Wu, Z.[Ziyi],
Vasilkovsky, M.[Michael],
Qian, G.C.[Guo-Cheng],
Ren, J.[Jian],
Guler, R.A.[Riza Alp],
Ghanem, B.[Bernard],
Tulyakov, S.[Sergey],
Gilitschenski, I.[Igor],
SPAD: Spatially Aware Multi-View Diffusers,
CVPR24(10026-10038)
IEEE DOI
2410
Geometry, Text to image, Transforms, Cameras, Diffusion models,
Encoding, novel view synthesis, diffusion
BibRef
Starodubcev, N.[Nikita],
Baranchuk, D.[Dmitry],
Fedorov, A.[Artem],
Babenko, A.[Artem],
Your Student is Better than Expected: Adaptive Teacher-Student
Collaboration for Text-Conditional Diffusion Models,
CVPR24(9275-9285)
IEEE DOI
2410
Adaptation models, Computational modeling, Pipelines,
Text to image, Collaboration, Diffusion models,
Image and video synthesis and generation
BibRef
Mei, K.[Kangfu],
Delbracio, M.[Mauricio],
Talebi, H.[Hossein],
Tu, Z.Z.[Zheng-Zhong],
Patel, V.M.[Vishal M.],
Milanfar, P.[Peyman],
CoDi: Conditional Diffusion Distillation for Higher-Fidelity and
Faster Image Generation,
CVPR24(9048-9058)
IEEE DOI
2410
Image synthesis, Superresolution, Text to image,
Computer architecture, Predictive models, Diffusion models
BibRef
Ran, L.M.[Ling-Min],
Cun, X.D.[Xiao-Dong],
Liu, J.W.[Jia-Wei],
Zhao, R.[Rui],
Zijie, S.[Song],
Wang, X.[Xintao],
Keppo, J.[Jussi],
Shou, M.Z.[Mike Zheng],
X- Adapter: Universal Compatibility of Plugins for Upgraded Diffusion
Model,
CVPR24(8775-8784)
IEEE DOI Code:
WWW Link.
2410
Training, Connectors, Adaptation models, Noise reduction,
Text to image, Diffusion models, Data models
BibRef
Liu, Y.J.[Yu-Jian],
Zhang, Y.[Yang],
Jaakkola, T.[Tommi],
Chang, S.Y.[Shi-Yu],
Correcting Diffusion Generation Through Resampling,
CVPR24(8713-8723)
IEEE DOI Code:
WWW Link.
2410
Image quality, Image synthesis, Filtering, Computational modeling,
Text to image, Detectors, image generation, diffusion model, particle filtering
BibRef
Luo, G.[Grace],
Darrell, T.J.[Trevor J.],
Wang, O.[Oliver],
Goldman, D.B.[Dan B],
Holynski, A.[Aleksander],
Readout Guidance: Learning Control from Diffusion Features,
CVPR24(8217-8227)
IEEE DOI Code:
WWW Link.
2410
Training, Head, Image edge detection, Training data, Text to image,
Diffusion models, Image and video synthesis and generation
BibRef
Wallace, B.[Bram],
Dang, M.[Meihua],
Rafailov, R.[Rafael],
Zhou, L.Q.[Lin-Qi],
Lou, A.[Aaron],
Purushwalkam, S.[Senthil],
Ermon, S.[Stefano],
Xiong, C.M.[Cai-Ming],
Joty, S.[Shafiq],
Naik, N.[Nikhil],
Diffusion Model Alignment Using Direct Preference Optimization,
CVPR24(8228-8238)
IEEE DOI
2410
Training, Learning systems, Visualization, Pipelines, Text to image,
Reinforcement learning, Diffusion models, generative, diffusion,
dpo
BibRef
Yan, J.N.[Jing Nathan],
Gu, J.[Jiatao],
Rush, A.M.[Alexander M.],
Diffusion Models Without Attention,
CVPR24(8239-8249)
IEEE DOI
2410
Training, Image resolution, Computational modeling,
Noise reduction, Text to image, Computer architecture
BibRef
Gokaslan, A.[Aaron],
Cooper, A.F.[A. Feder],
Collins, J.[Jasmine],
Seguin, L.[Landan],
Jacobson, A.[Austin],
Patel, M.[Mihir],
Frankle, J.[Jonathan],
Stephenson, C.[Cory],
Kuleshov, V.[Volodymyr],
Common Canvas: Open Diffusion Models Trained on Creative-Commons Images,
CVPR24(8250-8260)
IEEE DOI
2410
Training, Computational modeling, Transfer learning, Text to image,
Diffusion models, Data models, diffusion, copyright, text2image,
dataset
BibRef
Habibian, A.[Amirhossein],
Ghodrati, A.[Amir],
Fathima, N.[Noor],
Sautiere, G.[Guillaume],
Garrepalli, R.[Risheek],
Porikli, F.M.[Fatih M.],
Petersen, J.[Jens],
Clockwork Diffusion: Efficient Generation With Model-Step
Distillation,
CVPR24(8352-8361)
IEEE DOI Code:
WWW Link.
2410
Training, Adaptation models, Runtime, Noise reduction, Semantics,
Layout, Text to image, diffusion, efficient diffusion,
distillation
BibRef
Wang, J.Y.[Jun-Yan],
Sun, Z.H.[Zhen-Hong],
Tan, Z.Y.[Zhi-Yu],
Chen, X.B.[Xuan-Bai],
Chen, W.H.[Wei-Hua],
Li, H.[Hao],
Zhang, C.[Cheng],
Song, Y.[Yang],
Towards Effective Usage of Human-Centric Priors in Diffusion Models
for Text-based Human Image Generation,
CVPR24(8446-8455)
IEEE DOI Code:
WWW Link.
2410
Accuracy, Image synthesis, Semantics, Text to image,
Diffusion processes, Diffusion models
BibRef
Lin, H.[Haonan],
DreamSalon: A Staged Diffusion Framework for Preserving
Identity-Context in Editable Face Generation,
CVPR24(8589-8598)
IEEE DOI
2410
Image quality, Face recognition, Semantics, Noise reduction, Noise,
Text to image, Stochastic processes,
staged diffusion framework
BibRef
Li, Z.[Zhen],
Cao, M.D.[Ming-Deng],
Wang, X.[Xintao],
Qi, Z.A.[Zhong-Ang],
Cheng, M.M.[Ming-Ming],
Shan, Y.[Ying],
PhotoMaker: Customizing Realistic Human Photos via Stacked ID
Embedding,
CVPR24(8640-8650)
IEEE DOI
2410
Training, Pipelines, Text to image, Training data, Controllability,
diffusion model, personalization,
face synthesis
BibRef
Feng, Y.T.[Yu-Tong],
Gong, B.[Biao],
Chen, D.[Di],
Shen, Y.J.[Yu-Jun],
Liu, Y.[Yu],
Zhou, J.[Jingren],
Ranni: Taming Text-to-Image Diffusion for Accurate Instruction
Following,
CVPR24(4744-4753)
IEEE DOI
2410
Visualization, Protocols, Semantics, Pipelines, Text to image,
Diffusion models, Generators, diffusion model, text-to-image
BibRef
Lu, S.L.[Shi-Lin],
Wang, Z.[Zilan],
Li, L.[Leyang],
Liu, Y.Z.[Yan-Zhu],
Kong, A.W.K.[Adams Wai-Kin],
MACE: Mass Concept Erasure in Diffusion Models,
CVPR24(6430-6440)
IEEE DOI Code:
WWW Link.
2410
Codes, Text to image, Interference, Diffusion models,
Generative AI, AI security, diffusion model,
concept editing
BibRef
Nam, J.[Jisu],
Kim, H.[Heesu],
Lee, D.[DongJae],
Jin, S.[Siyoon],
Kim, S.[Seungryong],
Chang, S.[Seunggyu],
DreamMatcher: Appearance Matching Self-Attention for
Semantically-Consistent Text-to-Image Personalization,
CVPR24(8100-8110)
IEEE DOI
2410
Visualization, Computational modeling, Semantics, Noise reduction,
Text to image, Diffusion models, Diffusion Models,
Semantic Correspondence
BibRef
Ham, C.[Cusuh],
Fisher, M.[Matthew],
Hays, J.[James],
Kolkin, N.[Nicholas],
Liu, Y.C.[Yu-Chen],
Zhang, R.[Richard],
Hinz, T.[Tobias],
Personalized Residuals for Concept-Driven Text-to-Image Generation,
CVPR24(8186-8195)
IEEE DOI
2410
Training, Measurement, Computational modeling, Text to image,
Graphics processing units, Diffusion models, personalization,
diffusion models
BibRef
Phung, Q.[Quynh],
Ge, S.W.[Song-Wei],
Huang, J.B.[Jia-Bin],
Grounded Text-to-Image Synthesis with Attention Refocusing,
CVPR24(7932-7942)
IEEE DOI
2410
Visualization, Large language models, Computational modeling,
Layout, Text to image, Benchmark testing, Diffusion models,
grounded text-to-image
BibRef
Nguyen, T.H.[Thuan Hoang],
Tran, A.[Anh],
SwiftBrush: One-Step Text-to-Image Diffusion Model with Variational
Score Distillation,
CVPR24(7807-7816)
IEEE DOI
2410
Training, Solid modeling, Text to image, Diffusion models,
Neural radiance field, Data models
BibRef
Cao, C.J.[Chen-Jie],
Cai, Y.[Yunuo],
Dong, Q.[Qiaole],
Wang, Y.K.[Yi-Kai],
Fu, Y.W.[Yan-Wei],
LeftRefill: Filling Right Canvas based on Left Reference through
Generalized Text-to-Image Diffusion Model,
CVPR24(7705-7715)
IEEE DOI Code:
WWW Link.
2410
Adaptation models, Image synthesis, Text to image, Diffusion models,
Filling, Diffusion Model, Image Inpainting
BibRef
Mo, S.C.[Si-Cheng],
Mu, F.Z.[Fang-Zhou],
Lin, K.H.[Kuan Heng],
Liu, Y.L.[Yan-Li],
Guan, B.[Bochen],
Li, Y.[Yin],
Zhou, B.[Bolei],
FreeControl: Training-Free Spatial Control of Any Text-to-Image
Diffusion Model with Any Condition,
CVPR24(7465-7475)
IEEE DOI Code:
WWW Link.
2410
Visualization, Text to image, Computer architecture,
Aerospace electronics, Diffusion models, Feature extraction,
Controllable generation
BibRef
Huang, M.Q.[Meng-Qi],
Mao, Z.D.[Zhen-Dong],
Liu, M.C.[Ming-Cong],
He, Q.[Qian],
Zhang, Y.D.[Yong-Dong],
RealCustom: Narrowing Real Text Word for Real-Time Open-Domain
Text-to-Image Customization,
CVPR24(7476-7485)
IEEE DOI
2410
Training, Visualization, Adaptive systems, Limiting, Navigation,
Text to image, text-to-image generation,
diffusion models
BibRef
Mahajan, S.[Shweta],
Rahman, T.[Tanzila],
Yi, K.M.[Kwang Moo],
Sigal, L.[Leonid],
Prompting Hard or Hardly Prompting: Prompt Inversion for
Text-to-Image Diffusion Models,
CVPR24(6808-6817)
IEEE DOI
2410
Vocabulary, Visualization, Image synthesis, Semantics, Text to image,
Diffusion processes, Diffusion models
BibRef
Zhou, D.[Dewei],
Li, Y.[You],
Ma, F.[Fan],
Zhang, X.T.[Xiao-Ting],
Yang, Y.[Yi],
MIGC: Multi-Instance Generation Controller for Text-to-Image
Synthesis,
CVPR24(6818-6828)
IEEE DOI Code:
WWW Link.
2410
Codes, Attention mechanisms, Aggregates, Pipelines, Layout,
Text to image, AIGC, Diffusion Models, Image Generation, Stable Diffusion
BibRef
Zeng, Y.[Yu],
Patel, V.M.[Vishal M.],
Wang, H.C.[Hao-Chen],
Huang, X.[Xun],
Wang, T.C.[Ting-Chun],
Liu, M.Y.[Ming-Yu],
Balaji, Y.[Yogesh],
JeDi: Joint-Image Diffusion Models for Finetuning-Free Personalized
Text-to-Image Generation,
CVPR24(6786-6795)
IEEE DOI
2410
Adaptation models, Computational modeling, Text to image,
Benchmark testing, Diffusion models, image generation
BibRef
Gong, B.[Biao],
Huang, S.[Siteng],
Feng, Y.T.[Yu-Tong],
Zhang, S.W.[Shi-Wei],
Li, Y.[Yuyuan],
Liu, Y.[Yu],
Check, Locate, Rectify: A Training-Free Layout Calibration System for
Text- to- Image Generation,
CVPR24(6624-6634)
IEEE DOI Code:
WWW Link.
2410
Image synthesis, Layout, Pipelines, Text to image, Benchmark testing,
Diffusion models, Generators, text-to-image generation, training-free
BibRef
Hoe, J.T.[Jiun Tian],
Jiang, X.D.[Xu-Dong],
Chan, C.S.[Chee Seng],
Tan, Y.P.[Yap-Peng],
Hu, W.P.[Wei-Peng],
InteractDiffusion: Interaction Control in Text-to-Image Diffusion
Models,
CVPR24(6180-6189)
IEEE DOI Code:
WWW Link.
2410
Location awareness, Visualization, Computational modeling, Layout,
Text to image, Diffusion models, image generation, generative ai
BibRef
Menon, S.[Sachit],
Misra, I.[Ishan],
Girdhar, R.[Rohit],
Generating Illustrated Instructions,
CVPR24(6274-6284)
IEEE DOI
2410
Measurement, Visualization, Large language models, Text to image,
Diffusion models, diffusion, multimodal,
text-to-image
BibRef
Yang, J.Y.[Jing-Yuan],
Feng, J.W.[Jia-Wei],
Huang, H.[Hui],
EmoGen: Emotional Image Content Generation with Text-to-Image
Diffusion Models,
CVPR24(6358-6368)
IEEE DOI Code:
WWW Link.
2410
Measurement, Visualization, Image color analysis, Image synthesis,
Semantics, Text to image
BibRef
Yang, Y.J.[Yi-Jun],
Gao, R.[Ruiyuan],
Wang, X.[Xiaosen],
Ho, T.Y.[Tsung-Yi],
Xu, N.[Nan],
xu, Q.[Qiang],
MMA-Diffusion: MultiModal Attack on Diffusion Models,
CVPR24(7737-7746)
IEEE DOI Code:
WWW Link.
2410
Visualization, Filters, Current measurement,
Computational modeling, Text to image, Diffusion models,
Adversarial attack
BibRef
Hedlin, E.[Eric],
Sharma, G.[Gopal],
Mahajan, S.[Shweta],
He, X.Z.[Xing-Zhe],
Isack, H.[Hossam],
Kar, A.[Abhishek],
Rhodin, H.[Helge],
Tagliasacchi, A.[Andrea],
Yi, K.M.[Kwang Moo],
Unsupervised Keypoints from Pretrained Diffusion Models,
CVPR24(22820-22830)
IEEE DOI
2410
Codes, Noise reduction, Neural networks, Text to image,
Computer architecture, Diffusion models, Diffusion models,
emergent understandings
BibRef
Sato, T.[Takami],
Yue, J.[Justin],
Chen, N.[Nanze],
Wang, N.[Ningfei],
Chen, Q.A.[Qi Alfred],
Intriguing Properties of Diffusion Models: An Empirical Study of the
Natural Attack Capability in Text-to-Image Generative Models,
CVPR24(24635-24644)
IEEE DOI
2410
Noise reduction, Text to image, Artificial neural networks,
Visual systems, Predictive models, Diffusion models, Safety
BibRef
Gandikota, K.V.[Kanchana Vaishnavi],
Chandramouli, P.[Paramanand],
Text-Guided Explorable Image Super-Resolution,
CVPR24(25900-25911)
IEEE DOI
2410
Training, Degradation, Superresolution, Semantics, Text to image,
Diffusion models, diffusion, text-to-image, super-resolution
BibRef
Mo, W.[Wenyi],
Zhang, T.Y.[Tian-Yu],
Bai, Y.[Yalong],
Su, B.[Bing],
Wen, J.R.[Ji-Rong],
Yang, Q.[Qing],
Dynamic Prompt Optimizing for Text-to-Image Generation,
CVPR24(26617-26626)
IEEE DOI
2410
Uniform resource locators, Training, Image synthesis, Semantics,
Refining, Text to image, Reinforcement learning,
Diffusion Model
BibRef
Smith, J.S.[James Seale],
Hsu, Y.C.[Yen-Chang],
Kira, Z.[Zsolt],
Shen, Y.L.[Yi-Lin],
Jin, H.X.[Hong-Xia],
Continual Diffusion with STAMINA: STack-And-Mask INcremental Adapters,
WhatNext24(1744-1754)
IEEE DOI
2410
Training, Continuing education, Costs, Text to image,
Benchmark testing, Diffusion models,
text-to-image customization
BibRef
Zhang, G.[Gong],
Wang, K.[Kai],
Xu, X.Q.[Xing-Qian],
Wang, Z.Y.[Zhang-Yang],
Shi, H.[Humphrey],
Forget-Me-Not: Learning to Forget in Text-to-Image Diffusion Models,
WhatNext24(1755-1764)
IEEE DOI
2410
Adaptation models, Privacy, Accuracy, Computational modeling,
Knowledge based systems, Text to image, Safety, text-to-image,
concept forgetting
BibRef
Tudosiu, P.D.[Petru-Daniel],
Yang, Y.X.[Yong-Xin],
Zhang, S.F.[Shi-Feng],
Chen, F.[Fei],
McDonagh, S.[Steven],
Lampouras, G.[Gerasimos],
Iacobacci, I.[Ignacio],
Parisot, S.[Sarah],
MULAN: A Multi Layer Annotated Dataset for Controllable Text-to-Image
Generation,
CVPR24(22413-22422)
IEEE DOI Code:
WWW Link.
2410
Training, Image segmentation, Annotations, Pipelines, Text to image,
Image decomposition, Software, Dataset, Text-to-Image Generation,
Diffusion Models
BibRef
Wang, F.F.[Fei-Fei],
Tan, Z.T.[Zhen-Tao],
Wei, T.Y.[Tian-Yi],
Wu, Y.[Yue],
Huang, Q.D.[Qi-Dong],
SimAC: A Simple Anti-Customization Method for Protecting Face Privacy
Against Text-to-Image Synthesis of Diffusion Models,
CVPR24(12047-12056)
IEEE DOI Code:
WWW Link.
2410
Training, Privacy, Adaptation models, Visualization,
Frequency-domain analysis, Noise reduction, Text to image,
face privacy
BibRef
Pang, L.[Lianyu],
Yin, J.[Jian],
Xie, H.R.[Hao-Ran],
Wang, Q.[Qiping],
Li, Q.[Qing],
Mao, X.D.[Xu-Dong],
Cross Initialization for Face Personalization of Text-to-Image Models,
CVPR24(8393-8403)
IEEE DOI Code:
WWW Link.
2410
Face recognition, Computational modeling, Text to image,
Diffusion models, Surges, Image reconstruction
BibRef
Xu, X.Q.[Xing-Qian],
Guo, J.Y.[Jia-Yi],
Wang, Z.Y.[Zhang-Yang],
Huang, G.[Gao],
Essa, I.[Irfan],
Shi, H.[Humphrey],
Prompt-Free Diffusion: Taking 'Text' Out of Text-to-Image Diffusion
Models,
CVPR24(8682-8692)
IEEE DOI
2410
Visualization, Pain, Image synthesis, Computational modeling,
Semantics, Noise, Text to image, Generative Model, Image Editing, Text-to-Image
BibRef
Qi, T.H.[Tian-Hao],
Fang, S.C.[Shan-Cheng],
Wu, Y.[Yanze],
Xie, H.T.[Hong-Tao],
Liu, J.W.[Jia-Wei],
Chen, L.[Lang],
He, Q.[Qian],
Zhang, Y.D.[Yong-Dong],
DEADiff: An Efficient Stylization Diffusion Model with Disentangled
Representations,
CVPR24(8693-8702)
IEEE DOI Code:
WWW Link.
2410
Learning systems, Visualization, Semantics, Text to image,
Feature extraction, Diffusion models
BibRef
Brack, M.[Manuel],
Friedrich, F.[Felix],
Kornmeier, K.[Katharina],
Tsaban, L.[Linoy],
Schramowski, P.[Patrick],
Kersting, K.[Kristian],
Passos, A.[Apolinário],
LEDITS++: Limitless Image Editing Using Text-to-Image Models,
CVPR24(8861-8870)
IEEE DOI
2410
Computational modeling, Text to image, Computer architecture,
Benchmark testing, Diffusion models
BibRef
Li, H.[Hang],
Shen, C.Z.[Cheng-Zhi],
Torr, P.[Philip],
Tresp, V.[Volker],
Gu, J.D.[Jin-Dong],
Self-Discovering Interpretable Diffusion Latent Directions for
Responsible Text-to-Image Generation,
CVPR24(12006-12016)
IEEE DOI Code:
WWW Link.
2410
Ethics, Prevention and mitigation, Semantics, Text to image,
Diffusion models, Vectors, Text-to-Image Generation,
Explainability and Transparency
BibRef
Li, H.[Hao],
Zou, Y.[Yang],
Wang, Y.[Ying],
Majumder, O.[Orchid],
Xie, Y.S.[Yu-Sheng],
Manmatha, R.,
Swaminathan, A.[Ashwin],
Tu, Z.W.[Zhuo-Wen],
Ermon, S.[Stefano],
Soatto, S.[Stefano],
On the Scalability of Diffusion-based Text-to-Image Generation,
CVPR24(9400-9409)
IEEE DOI
2410
Training, Costs, Systematics, Computational modeling, Scalability,
Noise reduction, Text to image, diffusion models, text-to-image,
Transformers
BibRef
Guo, X.[Xiefan],
Liu, J.L.[Jin-Lin],
Cui, M.M.[Miao-Miao],
Li, J.[Jiankai],
Yang, H.Y.[Hong-Yu],
Huang, D.[Di],
Initno: Boosting Text-to-Image Diffusion Models via Initial Noise
Optimization,
CVPR24(9380-9389)
IEEE DOI Code:
WWW Link.
2410
Navigation, Instruments, Noise, Pipelines, Text to image, Aerospace electronics
BibRef
Shen, D.[Dazhong],
Song, G.L.[Guang-Lu],
Xue, Z.[Zeyue],
Wang, F.Y.[Fu-Yun],
Liu, Y.[Yu],
Rethinking the Spatial Inconsistency in Classifier-Free Diffusion
Guidance,
CVPR24(9370-9379)
IEEE DOI Code:
WWW Link.
2410
Image quality, Training, Costs, Semantic segmentation, Semantics,
Noise reduction, Text-to-Image Diffusion Models,
Semantic Segmentation
BibRef
Zhou, Y.F.[Yu-Fan],
Zhang, R.[Ruiyi],
Gu, J.X.[Jiu-Xiang],
Sun, T.[Tong],
Customization Assistant for Text-to-image Generation,
CVPR24(9182-9191)
IEEE DOI
2410
Training, Large language models, Text to image, Diffusion models,
Testing
BibRef
Patel, M.[Maitreya],
Kim, C.[Changhoon],
Cheng, S.[Sheng],
Baral, C.[Chitta],
Yang, Y.Z.[Ye-Zhou],
ECLIPSE: A Resource-Efficient Text-to-Image Prior for Image
Generations,
CVPR24(9069-9078)
IEEE DOI Code:
WWW Link.
2410
Training, Image coding, Image synthesis, Computational modeling,
Text to image, Contrastive learning, Diffusion models,
ECLIPSE
BibRef
Meral, T.H.S.[Tuna Han Salih],
Simsar, E.[Enis],
Tombari, F.[Federico],
Yanardag, P.[Pinar],
CONFORM: Contrast is All You Need For High-Fidelity Text-to-Image
Diffusion Models,
CVPR24(9005-9014)
IEEE DOI
2410
Source coding, Computational modeling, Semantics, Text to image,
Benchmark testing, Diffusion models,
Semantic fidelity
BibRef
Jiang, Z.Z.[Zeyin-Zi],
Mao, C.J.[Chao-Jie],
Pan, Y.L.[Yu-Lin],
Han, Z.[Zhen],
Zhang, J.F.[Jing-Feng],
SCEdit: Efficient and Controllable Image Diffusion Generation via
Skip Connection Editing,
CVPR24(8995-9004)
IEEE DOI Code:
WWW Link.
2410
Training, Adaptation models, Tuners, Image synthesis, Text to image,
Diffusion models, Diffusion model, Text-to-image generation,
Efficient Tuning
BibRef
Kim, C.[Changhoon],
Min, K.[Kyle],
Patel, M.[Maitreya],
Cheng, S.[Sheng],
Yang, Y.Z.[Ye-Zhou],
WOUAF: Weight Modulation for User Attribution and Fingerprinting in
Text-to-Image Diffusion Models,
CVPR24(8974-8983)
IEEE DOI Code:
WWW Link.
2410
Solid modeling, Computational modeling,
Prevention and mitigation, Text to image, Modulation,
Generative Model
BibRef
Shirakawa, T.[Takahiro],
Uchida, S.[Seiichi],
NoiseCollage: A Layout-Aware Text-to-Image Diffusion Model Based on
Noise Cropping and Merging,
CVPR24(8921-8930)
IEEE DOI Code:
WWW Link.
2410
Image synthesis, Image edge detection, Noise, Layout,
Noise reduction, Merging, Text to image, diffusion model,
text-to-image generation
BibRef
Kwon, G.[Gihyun],
Jenni, S.[Simon],
Li, D.Z.[Ding-Zeyu],
Lee, J.Y.[Joon-Young],
Ye, J.C.[Jong Chul],
Heilbron, F.C.[Fabian Caba],
Concept Weaver: Enabling Multi-Concept Fusion in Text-to-Image Models,
CVPR24(8880-8889)
IEEE DOI
2410
Fuses, Semantics, Text to image, Diffusion models,
Optimization, Text-to-image Model,
Multi-concept
BibRef
Sueyoshi, K.[Kota],
Matsubara, T.[Takashi],
Predicated Diffusion: Predicate Logic-Based Attention Guidance for
Text-to-Image Diffusion Models,
CVPR24(8651-8660)
IEEE DOI
2410
Image quality, Image synthesis, Natural languages, Layout,
Text to image, Diffusion models, text-to-image generation,
attention guidance
BibRef
Wang, Z.[Zirui],
Sha, Z.Z.[Zhi-Zhou],
Ding, Z.[Zheng],
Wang, Y.L.[Yi-Lin],
Tu, Z.W.[Zhuo-Wen],
TokenCompose: Text-to-Image Diffusion with Token-Level Supervision,
CVPR24(8553-8564)
IEEE DOI
2410
Training, Photorealism, Pipelines, Noise reduction, Text to image,
Object segmentation, Benchmark testing, Diffusion Models,
Compositional Generation
BibRef
Kim, J.[Jimyeong],
Park, J.[Jungwon],
Rhee, W.[Wonjong],
Selectively Informative Description can Reduce Undesired Embedding
Entanglements in Text-to-Image Personalization,
CVPR24(8312-8322)
IEEE DOI
2410
Text to image, Reflection,
Text-to-Image Generation, Text-to-Image Diffusion, Text-to-image Personalization
BibRef
Koley, S.[Subhadeep],
Bhunia, A.K.[Ayan Kumar],
Sain, A.[Aneeshan],
Chowdhury, P.N.[Pinaki Nath],
Xiang, T.[Tao],
Song, Y.Z.[Yi-Zhe],
Text-to-Image Diffusion Models are Great Sketch-Photo Matchmakers,
CVPR24(16826-16837)
IEEE DOI
2410
Visualization, Adaptation models, Shape, Pipelines, Image retrieval,
Text to image, Benchmark testing
BibRef
Zhao, L.[Lin],
Zhao, T.C.[Tian-Chen],
Lin, Z.[Zinan],
Ning, X.F.[Xue-Fei],
Dai, G.H.[Guo-Hao],
Yang, H.Z.[Hua-Zhong],
Wang, Y.[Yu],
FlashEval: Towards Fast and Accurate Evaluation of Text-to-Image
Diffusion Generative Models,
CVPR24(16122-16131)
IEEE DOI Code:
WWW Link.
2410
Training, Schedules, Quantization (signal), Computational modeling,
Text to image, Training data, Diffusion models
BibRef
Liu, H.[Hanwen],
Sun, Z.C.[Zhi-Cheng],
Mu, Y.D.[Ya-Dong],
Countering Personalized Text-to-Image Generation with Influence
Watermarks,
CVPR24(12257-12267)
IEEE DOI
2410
Training, Visualization, Computational modeling, Semantics, Noise,
Text to image, Watermarking, diffusion models, watermarks
BibRef
Azarian, K.[Kambiz],
Das, D.[Debasmit],
Hou, Q.Q.[Qi-Qi],
Porikli, F.M.[Fatih M.],
Segmentation-Free Guidance for Text-to-Image Diffusion Models,
GCV24(7520-7529)
IEEE DOI
2410
Image segmentation, Costs, Image color analysis, Text to image,
Focusing, Switches
BibRef
Li, C.[Cheng],
Qi, Y.[Yali],
Zeng, Q.[Qingtao],
Lu, L.[Likun],
Comparison of Image Generation methods based on Diffusion Models,
CVIDL23(1-4)
IEEE DOI
2403
Training, Deep learning, Learning systems, Image synthesis,
Computational modeling, Diffusion models
BibRef
Xu, Y.[Yanwu],
Zhao, Y.[Yang],
Xiao, Z.S.[Zhi-Sheng],
Hou, T.B.[Ting-Bo],
UFOGen: You Forward Once Large Scale Text-to-Image Generation via
Diffusion GANs,
CVPR24(8196-8206)
IEEE DOI
2410
Image synthesis, Computational modeling, Text to image, Propulsion,
Diffusion models, Hybrid power systems, diffusion models,
GANs
BibRef
Huang, R.H.[Run-Hui],
Han, J.H.[Jian-Hua],
Lu, G.S.[Guan-Song],
Liang, X.D.[Xiao-Dan],
Zeng, Y.[Yihan],
Zhang, W.[Wei],
Xu, H.[Hang],
DiffDis: Empowering Generative Diffusion Model with Cross-Modal
Discrimination Capability,
ICCV23(15667-15677)
IEEE DOI
2401
BibRef
Yang, X.Y.[Xing-Yi],
Wang, X.C.[Xin-Chao],
Diffusion Model as Representation Learner,
ICCV23(18892-18903)
IEEE DOI Code:
WWW Link.
2401
BibRef
Nair, N.G.[Nithin Gopalakrishnan],
Cherian, A.[Anoop],
Lohit, S.[Suhas],
Wang, Y.[Ye],
Koike-Akino, T.[Toshiaki],
Patel, V.M.[Vishal M.],
Marks, T.K.[Tim K.],
Steered Diffusion: A Generalized Framework for Plug-and-Play
Conditional Image Synthesis,
ICCV23(20793-20803)
IEEE DOI
2401
BibRef
Wang, Z.D.[Zhen-Dong],
Bao, J.M.[Jian-Min],
Zhou, W.G.[Wen-Gang],
Wang, W.[Weilun],
Hu, H.[Hezhen],
Chen, H.[Hong],
Li, H.Q.[Hou-Qiang],
DIRE for Diffusion-Generated Image Detection,
ICCV23(22388-22398)
IEEE DOI Code:
WWW Link.
2401
BibRef
Hong, S.[Susung],
Lee, G.[Gyuseong],
Jang, W.[Wooseok],
Kim, S.[Seungryong],
Improving Sample Quality of Diffusion Models Using Self-Attention
Guidance,
ICCV23(7428-7437)
IEEE DOI
2401
BibRef
Feng, B.T.[Berthy T.],
Smith, J.[Jamie],
Rubinstein, M.[Michael],
Chang, H.[Huiwen],
Bouman, K.L.[Katherine L.],
Freeman, W.T.[William T.],
Score-Based Diffusion Models as Principled Priors for Inverse Imaging,
ICCV23(10486-10497)
IEEE DOI
2401
BibRef
Yang, B.B.[Bin-Bin],
Luo, Y.[Yi],
Chen, Z.L.[Zi-Liang],
Wang, G.R.[Guang-Run],
Liang, X.D.[Xiao-Dan],
Lin, L.[Liang],
LAW-Diffusion: Complex Scene Generation by Diffusion with Layouts,
ICCV23(22612-22622)
IEEE DOI
2401
BibRef
Levi, E.[Elad],
Brosh, E.[Eli],
Mykhailych, M.[Mykola],
Perez, M.[Meir],
DLT: Conditioned layout generation with Joint Discrete-Continuous
Diffusion Layout Transformer,
ICCV23(2106-2115)
IEEE DOI Code:
WWW Link.
2401
BibRef
Couairon, G.[Guillaume],
Careil, M.[Marlène],
Cord, M.[Matthieu],
Lathuilière, S.[Stéphane],
Verbeek, J.[Jakob],
Zero-shot spatial layout conditioning for text-to-image diffusion
models,
ICCV23(2174-2183)
IEEE DOI
2401
BibRef
Zhang, L.[Lvmin],
Rao, A.[Anyi],
Agrawala, M.[Maneesh],
Adding Conditional Control to Text-to-Image Diffusion Models,
ICCV23(3813-3824)
IEEE DOI
2401
Award, Marr Price, ICCV.
BibRef
Zhao, W.L.[Wen-Liang],
Rao, Y.M.[Yong-Ming],
Liu, Z.[Zuyan],
Liu, B.[Benlin],
Zhou, J.[Jie],
Lu, J.W.[Ji-Wen],
Unleashing Text-to-Image Diffusion Models for Visual Perception,
ICCV23(5706-5716)
IEEE DOI Code:
WWW Link.
2401
BibRef
Xie, J.[Jinheng],
Li, Y.X.[Yue-Xiang],
Huang, Y.W.[Ya-Wen],
Liu, H.Z.[Hao-Zhe],
Zhang, W.[Wentian],
Zheng, Y.F.[Ye-Feng],
Shou, M.Z.[Mike Zheng],
BoxDiff: Text-to-Image Synthesis with Training-Free Box-Constrained
Diffusion,
ICCV23(7418-7427)
IEEE DOI
2401
BibRef
Wu, Q.C.[Qiu-Cheng],
Liu, Y.J.[Yu-Jian],
Zhao, H.[Handong],
Bui, T.[Trung],
Lin, Z.[Zhe],
Zhang, Y.[Yang],
Chang, S.Y.[Shi-Yu],
Harnessing the Spatial-Temporal Attention of Diffusion Models for
High-Fidelity Text-to-Image Synthesis,
ICCV23(7732-7742)
IEEE DOI
2401
BibRef
Khachatryan, L.[Levon],
Movsisyan, A.[Andranik],
Tadevosyan, V.[Vahram],
Henschel, R.[Roberto],
Wang, Z.Y.[Zhang-Yang],
Navasardyan, S.[Shant],
Shi, H.[Humphrey],
Text2Video-Zero: Text-to-Image Diffusion Models are Zero-Shot Video
Generators,
ICCV23(15908-15918)
IEEE DOI Code:
WWW Link.
2401
BibRef
Zhao, J.[Jing],
Zheng, H.[Heliang],
Wang, C.[Chaoyue],
Lan, L.[Long],
Yang, W.J.[Wen-Jing],
MagicFusion: Boosting Text-to-Image Generation Performance by Fusing
Diffusion Models,
ICCV23(22535-22545)
IEEE DOI Code:
WWW Link.
2401
BibRef
Kumari, N.[Nupur],
Zhang, B.L.[Bing-Liang],
Wang, S.Y.[Sheng-Yu],
Shechtman, E.[Eli],
Zhang, R.[Richard],
Zhu, J.Y.[Jun-Yan],
Ablating Concepts in Text-to-Image Diffusion Models,
ICCV23(22634-22645)
IEEE DOI
2401
BibRef
Schwartz, I.[Idan],
Snæbjarnarson, V.[Vésteinn],
Chefer, H.[Hila],
Belongie, S.[Serge],
Wolf, L.[Lior],
Benaim, S.[Sagie],
Discriminative Class Tokens for Text-to-Image Diffusion Models,
ICCV23(22668-22678)
IEEE DOI Code:
WWW Link.
2401
BibRef
Patashnik, O.[Or],
Garibi, D.[Daniel],
Azuri, I.[Idan],
Averbuch-Elor, H.[Hadar],
Cohen-Or, D.[Daniel],
Localizing Object-level Shape Variations with Text-to-Image Diffusion
Models,
ICCV23(22994-23004)
IEEE DOI
2401
BibRef
Ge, S.W.[Song-Wei],
Park, T.[Taesung],
Zhu, J.Y.[Jun-Yan],
Huang, J.B.[Jia-Bin],
Expressive Text-to-Image Generation with Rich Text,
ICCV23(7511-7522)
IEEE DOI
2401
BibRef
Kim, Y.J.[Yun-Ji],
Lee, J.Y.[Ji-Young],
Kim, J.H.[Jin-Hwa],
Ha, J.W.[Jung-Woo],
Zhu, J.Y.[Jun-Yan],
Dense Text-to-Image Generation with Attention Modulation,
ICCV23(7667-7677)
IEEE DOI Code:
WWW Link.
2401
BibRef
Schramowski, P.[Patrick],
Brack, M.[Manuel],
Deiseroth, B.[Björn],
Kersting, K.[Kristian],
Safe Latent Diffusion: Mitigating Inappropriate Degeneration in
Diffusion Models,
CVPR23(22522-22531)
IEEE DOI
2309
BibRef
Chen, C.[Chen],
Liu, D.[Daochang],
Ma, S.Q.[Si-Qi],
Nepal, S.[Surya],
Xu, C.[Chang],
Private Image Generation with Dual-Purpose Auxiliary Classifier,
CVPR23(20361-20370)
IEEE DOI
2309
BibRef
Ni, H.[Haomiao],
Shi, C.[Changhao],
Li, K.[Kai],
Huang, S.X.[Sharon X.],
Min, M.R.[Martin Renqiang],
Conditional Image-to-Video Generation with Latent Flow Diffusion
Models,
CVPR23(18444-18455)
IEEE DOI
2309
BibRef
Zhang, Q.S.[Qin-Sheng],
Song, J.[JiaMing],
Huang, X.[Xun],
Chen, Y.X.[Yong-Xin],
Liu, M.Y.[Ming-Yu],
DiffCollage: Parallel Generation of Large Content with Diffusion
Models,
CVPR23(10188-10198)
IEEE DOI
2309
BibRef
Phung, H.[Hao],
Dao, Q.[Quan],
Tran, A.[Anh],
Wavelet Diffusion Models are fast and scalable Image Generators,
CVPR23(10199-10208)
IEEE DOI
2309
BibRef
Kim, S.W.[Seung Wook],
Brown, B.[Bradley],
Yin, K.X.[Kang-Xue],
Kreis, K.[Karsten],
Schwarz, K.[Katja],
Li, D.[Daiqing],
Rombach, R.[Robin],
Torralba, A.[Antonio],
Fidler, S.[Sanja],
NeuralField-LDM: Scene Generation with Hierarchical Latent Diffusion
Models,
CVPR23(8496-8506)
IEEE DOI
2309
BibRef
Luo, Z.X.[Zheng-Xiong],
Chen, D.[Dayou],
Zhang, Y.Y.[Ying-Ya],
Huang, Y.[Yan],
Wang, L.[Liang],
Shen, Y.J.[Yu-Jun],
Zhao, D.L.[De-Li],
Zhou, J.[Jingren],
Tan, T.N.[Tie-Niu],
VideoFusion: Decomposed Diffusion Models for High-Quality Video
Generation,
CVPR23(10209-10218)
IEEE DOI
2309
BibRef
Ruan, L.[Ludan],
Ma, Y.Y.[Yi-Yang],
Yang, H.[Huan],
He, H.G.[Hui-Guo],
Liu, B.[Bei],
Fu, J.L.[Jian-Long],
Yuan, N.J.[Nicholas Jing],
Jin, Q.[Qin],
Guo, B.[Baining],
MM-Diffusion: Learning Multi-Modal Diffusion Models for Joint Audio
and Video Generation,
CVPR23(10219-10228)
IEEE DOI
2309
BibRef
Zhu, Y.Z.[Yuan-Zhi],
Li, Z.H.[Zhao-Hai],
Wang, T.W.[Tian-Wei],
He, M.C.[Meng-Chao],
Yao, C.[Cong],
Conditional Text Image Generation with Diffusion Models,
CVPR23(14235-14244)
IEEE DOI
2309
BibRef
Zhou, Y.F.[Yu-Fan],
Liu, B.C.[Bing-Chen],
Zhu, Y.Z.[Yi-Zhe],
Yang, X.[Xiao],
Chen, C.Y.[Chang-You],
Xu, J.H.[Jin-Hui],
Shifted Diffusion for Text-to-image Generation,
CVPR23(10157-10166)
IEEE DOI
2309
BibRef
Li, M.[Muheng],
Duan, Y.[Yueqi],
Zhou, J.[Jie],
Lu, J.W.[Ji-Wen],
Diffusion-SDF: Text-to-Shape via Voxelized Diffusion,
CVPR23(12642-12651)
IEEE DOI
2309
BibRef
Chai, S.[Shang],
Zhuang, L.S.[Lian-Sheng],
Yan, F.Y.[Feng-Ying],
LayoutDM: Transformer-based Diffusion Model for Layout Generation,
CVPR23(18349-18358)
IEEE DOI
2309
BibRef
Wu, Q.C.[Qiu-Cheng],
Liu, Y.J.[Yu-Jian],
Zhao, H.[Handong],
Kale, A.[Ajinkya],
Bui, T.[Trung],
Yu, T.[Tong],
Lin, Z.[Zhe],
Zhang, Y.[Yang],
Chang, S.Y.[Shi-Yu],
Uncovering the Disentanglement Capability in Text-to-Image Diffusion
Models,
CVPR23(1900-1910)
IEEE DOI
2309
BibRef
Jain, A.[Ajay],
Xie, A.[Amber],
Abbeel, P.[Pieter],
VectorFusion: Text-to-SVG by Abstracting Pixel-Based Diffusion Models,
CVPR23(1911-1920)
IEEE DOI
2309
BibRef
Kumari, N.[Nupur],
Zhang, B.L.[Bing-Liang],
Zhang, R.[Richard],
Shechtman, E.[Eli],
Zhu, J.Y.[Jun-Yan],
Multi-Concept Customization of Text-to-Image Diffusion,
CVPR23(1931-1941)
IEEE DOI
2309
BibRef
Hui, M.[Mude],
Zhang, Z.Z.[Zhi-Zheng],
Zhang, X.Y.[Xiao-Yi],
Xie, W.X.[Wen-Xuan],
Wang, Y.W.[Yu-Wang],
Lu, Y.[Yan],
Unifying Layout Generation with a Decoupled Diffusion Model,
CVPR23(1942-1951)
IEEE DOI
2309
BibRef
Ruiz, N.[Nataniel],
Li, Y.Z.[Yuan-Zhen],
Jampani, V.[Varun],
Pritch, Y.[Yael],
Rubinstein, M.[Michael],
Aberman, K.[Kfir],
DreamBooth: Fine Tuning Text-to-Image Diffusion Models for
Subject-Driven Generation,
CVPR23(22500-22510)
IEEE DOI
2309
BibRef
Zheng, G.C.[Guang-Cong],
Zhou, X.P.[Xian-Pan],
Li, X.W.[Xue-Wei],
Qi, Z.A.[Zhong-Ang],
Shan, Y.[Ying],
Li, X.[Xi],
LayoutDiffusion: Controllable Diffusion Model for Layout-to-Image
Generation,
CVPR23(22490-22499)
IEEE DOI
2309
BibRef
Liu, X.H.[Xi-Hui],
Park, D.H.[Dong Huk],
Azadi, S.[Samaneh],
Zhang, G.[Gong],
Chopikyan, A.[Arman],
Hu, Y.X.[Yu-Xiao],
Shi, H.[Humphrey],
Rohrbach, A.[Anna],
Darrell, T.J.[Trevor J.],
More Control for Free! Image Synthesis with Semantic Diffusion
Guidance,
WACV23(289-299)
IEEE DOI
2302
Image synthesis, Annotations, Image matching, Semantics,
Noise reduction, Probabilistic logic,
Vision + language and/or other modalities
BibRef
Pan, Z.H.[Zhi-Hong],
Zhou, X.[Xin],
Tian, H.[Hao],
Arbitrary Style Guidance for Enhanced Diffusion-Based Text-to-Image
Generation,
WACV23(4450-4460)
IEEE DOI
2302
Graphics, Training, Technological innovation, Adaptation models,
Adaptive systems, Art, Navigation,
Vision + language and/or other modalities
BibRef
Gu, S.Y.[Shu-Yang],
Chen, D.[Dong],
Bao, J.M.[Jian-Min],
Wen, F.[Fang],
Zhang, B.[Bo],
Chen, D.D.[Dong-Dong],
Yuan, L.[Lu],
Guo, B.N.[Bai-Ning],
Vector Quantized Diffusion Model for Text-to-Image Synthesis,
CVPR22(10686-10696)
IEEE DOI
2210
Image quality, Image resolution, Image synthesis,
Computational modeling, Noise reduction,
Vision+language
BibRef
Jing, B.[Bowen],
Corso, G.[Gabriele],
Berlinghieri, R.[Renato],
Jaakkola, T.[Tommi],
Subspace Diffusion Generative Models,
ECCV22(XXIII:274-289).
Springer DOI
2211
BibRef
Han, L.G.[Li-Gong],
Li, Y.X.[Yin-Xiao],
Zhang, H.[Han],
Milanfar, P.[Peyman],
Metaxas, D.N.[Dimitris N.],
Yang, F.[Feng],
SVDiff: Compact Parameter Space for Diffusion Fine-Tuning,
ICCV23(7289-7300)
IEEE DOI
2401
BibRef
Nair, N.G.[Nithin Gopalakrishnan],
Bandara, W.G.C.[Wele Gedara Chaminda],
Patel, V.M.[Vishal M.],
Unite and Conquer: Plug and Play Multi-Modal Synthesis Using
Diffusion Models,
CVPR23(6070-6079)
IEEE DOI
2309
BibRef
Benny, Y.[Yaniv],
Wolf, L.B.[Lior B.],
Dynamic Dual-Output Diffusion Models,
CVPR22(11472-11481)
IEEE DOI
2210
Image quality, Image synthesis, Noise reduction,
Generative adversarial networks,
Image and video synthesis and generation
BibRef
Hu, M.H.[Ming-Hui],
Wang, Y.J.[Yu-Jie],
Cham, T.J.[Tat-Jen],
Yang, J.F.[Jian-Fei],
Suganthan, P.N.,
Global Context with Discrete Diffusion in Vector Quantised Modelling
for Image Generation,
CVPR22(11492-11501)
IEEE DOI
2210
Training, Visualization, Image resolution, Image synthesis,
Pipelines, Noise reduction, Probabilistic logic,
Image and video synthesis and generation
BibRef
Ma, H.Y.[Heng-Yuan],
Zhang, L.[Li],
Zhu, X.T.[Xia-Tian],
Feng, J.F.[Jian-Feng],
Accelerating Score-Based Generative Models with Preconditioned
Diffusion Sampling,
ECCV22(XXIII:1-16).
Springer DOI
2211
BibRef
Zheng, G.[Guangcong],
Li, S.[Shengming],
Wang, H.[Hui],
Yao, T.P.[Tai-Ping],
Chen, Y.[Yang],
Ding, S.H.[Shou-Hong],
Li, X.[Xi],
Entropy-Driven Sampling and Training Scheme for Conditional Diffusion
Generation,
ECCV22(XXII:754-769).
Springer DOI
2211
BibRef
Liu, N.[Nan],
Li, S.[Shuang],
Du, Y.L.[Yi-Lun],
Torralba, A.[Antonio],
Tenenbaum, J.B.[Joshua B.],
Compositional Visual Generation with Composable Diffusion Models,
ECCV22(XVII:423-439).
Springer DOI
2211
BibRef
Sehwag, V.[Vikash],
Hazirbas, C.[Caner],
Gordo, A.[Albert],
Ozgenel, F.[Firat],
Ferrer, C.C.[Cristian Canton],
Generating High Fidelity Data from Low-density Regions using
Diffusion Models,
CVPR22(11482-11491)
IEEE DOI
2210
Manifolds, Computational modeling, Diffusion processes,
Data models, Representation learning
BibRef
Chapter on 3-D Object Description and Computation Techniques, Surfaces, Deformable, View Generation, Video Conferencing continues in
Vision Transformers for Image Generation and Image Synthesis .