Cao, Z.[Ziang],
Hong, F.Z.[Fang-Zhou],
Wu, T.[Tong],
Pan, L.[Liang],
Liu, Z.W.[Zi-Wei],
DiffTF++: 3D-Aware Diffusion Transformer for Large-Vocabulary 3D
Generation,
PAMI(47), No. 4, April 2025, pp. 3018-3030.
IEEE DOI
2503
Transformers, Diffusion models, Fitting, Training, Solid modeling,
Neural radiance field, Image reconstruction, Topology, transformer
BibRef
Xu, H.F.[Hai-Feng],
Huai, Y.J.[Yong-Jian],
Nie, X.Y.[Xiao-Ying],
Meng, Q.[Qingkuo],
Zhao, X.[Xun],
Pei, X.[Xuanda],
Lu, H.[Hao],
Diff-Tree: A Diffusion Model for Diversified Tree Point Cloud
Generation with High Realism,
RS(17), No. 5, 2025, pp. 923.
DOI Link
2503
BibRef
Liu, Y.H.[Yu-Heng],
Li, X.K.[Xin-Ke],
Li, X.T.[Xue-Ting],
Qi, L.[Lu],
Li, C.[Chongshou],
Yang, M.H.[Ming-Hsuan],
Pyramid Diffusion for Fine 3d Large Scene Generation,
ECCV24(LXIX: 71-87).
Springer DOI
2412
BibRef
Liu, Y.[Yibo],
Yang, Z.[Zheyuan],
Wu, G.[Guile],
Ren, Y.[Yuan],
Lin, K.[Kejian],
Liu, B.B.[Bing-Bing],
Liu, Y.[Yang],
Shan, J.J.[Jin-Jun],
VQA-DIFF: Exploiting VQA and Diffusion for Zero-shot Image-to-3d
Vehicle Asset Generation in Autonomous Driving,
ECCV24(LXVI: 323-340).
Springer DOI
2412
BibRef
Burgess, J.[James],
Wang, K.C.[Kuan-Chieh],
Yeung-Levy, S.[Serena],
Viewpoint Textual Inversion: Discovering Scene Representations and 3d
View Control in 2d Diffusion Models,
ECCV24(LXIV: 416-435).
Springer DOI
2412
BibRef
Kalischek, N.[Nikolai],
Peters, T.[Torben],
Wegner, J.D.[Jan D.],
Schindler, K.[Konrad],
Tetradiffusion: Tetrahedral Diffusion Models for 3d Shape Generation,
ECCV24(LIII: 357-373).
Springer DOI
2412
BibRef
Shim, D.[Dongseok],
Kim, H.J.[H. Jin],
SEDIFF: Structure Extraction for Domain Adaptive Depth Estimation via
Denoising Diffusion Models,
ECCV24(XIX: 37-53).
Springer DOI
2412
BibRef
Chen, Y.W.[Yong-Wei],
Wang, T.F.[Teng-Fei],
Wu, T.[Tong],
Pan, X.G.[Xin-Gang],
Jia, K.[Kui],
Liu, Z.W.[Zi-Wei],
Comboverse: Compositional 3D Assets Creation Using Spatially-aware
Diffusion Guidance,
ECCV24(XXIV: 128-146).
Springer DOI
2412
BibRef
Cheng, F.[Feng],
Luo, M.[Mi],
Wang, H.Y.[Hui-Yu],
Dimakis, A.[Alex],
Torresani, L.[Lorenzo],
Bertasius, G.[Gedas],
Grauman, K.[Kristen],
4DIFF: 3D-Aware Diffusion Model for Third-to-first Viewpoint
Translation,
ECCV24(XXIV: 409-427).
Springer DOI
2412
BibRef
Yang, X.F.[Xiao-Feng],
Chen, Y.W.[Yi-Wen],
Chen, C.[Cheng],
Zhang, C.[Chi],
Xu, Y.[Yi],
Yang, X.[Xulei],
Liu, F.[Fayao],
Lin, G.S.[Guo-Sheng],
Learn to Optimize Denoising Scores: A Unified and Improved Diffusion
Prior for 3d Generation,
ECCV24(XLIV: 136-152).
Springer DOI
2412
BibRef
Yoshiyasu, Y.[Yusuke],
Sun, L.Y.[Le-Yuan],
Diffsurf: A Transformer-based Diffusion Model for Generating and
Reconstructing 3d Surfaces in Pose,
ECCV24(LXXXII: 246-264).
Springer DOI
2412
BibRef
Shim, H.J.[Ha-Jin],
Kim, C.H.[Chang-Hun],
Yang, E.[Eunho],
Cloudfixer: Test-time Adaptation for 3d Point Clouds via
Diffusion-guided Geometric Transformation,
ECCV24(XXXI: 454-471).
Springer DOI
2412
BibRef
Lan, Y.S.[Yu-Shi],
Hong, F.Z.[Fang-Zhou],
Yang, S.[Shuai],
Zhou, S.C.[Shang-Chen],
Meng, X.Y.[Xu-Yi],
Dai, B.[Bo],
Pan, X.G.[Xin-Gang],
Loy, C.C.[Chen Change],
Ln3diff: Scalable Latent Neural Fields Diffusion for Speedy 3d
Generation,
ECCV24(IV: 112-130).
Springer DOI
2412
BibRef
Zhu, X.Y.[Xiao-Yu],
Zhou, H.[Hao],
Xing, P.F.[Peng-Fei],
Zhao, L.[Long],
Xu, H.[Hao],
Liang, J.W.[Jun-Wei],
Hauptmann, A.[Alexander],
Liu, T.[Ting],
Gallagher, A.[Andrew],
Open-Vocabulary 3D Semantic Segmentation with Text-to-Image Diffusion
Models,
ECCV24(XXIX: 357-375).
Springer DOI
2412
BibRef
Gu, Y.M.[Yu-Ming],
Xu, H.Y.[Hong-Yi],
Xie, Y.[You],
Song, G.X.[Guo-Xian],
Shi, Y.C.[Yi-Chun],
Chang, D.[Di],
Yang, J.[Jing],
Luo, L.J.[Lin-Jie],
DiffPortrait3D: Controllable Diffusion for Zero-Shot Portrait View
Synthesis,
CVPR24(10456-10465)
IEEE DOI
2410
Training, Visualization, Noise reduction, Noise, Cameras,
Diffusion models, diffusion model, generative model, single to 3D
BibRef
Wang, Z.[Zhen],
Xu, Q.G.[Qian-Geng],
Tan, F.T.[Fei-Tong],
Chai, M.L.[Meng-Lei],
Liu, S.C.[Shi-Chen],
Pandey, R.[Rohit],
Fanello, S.[Sean],
Kadambi, A.[Achuta],
Zhang, Y.[Yinda],
MVDD: Multi-view Depth Diffusion Models,
ECCV24(XIII: 236-253).
Springer DOI
2412
BibRef
Zhai, G.Y.[Guang-Yao],
Örnek, E.P.[Evin Pinar],
Chen, D.Z.Y.[Dave Zhen-Yu],
Liao, R.[Ruotong],
Di, Y.[Yan],
Navab, N.[Nassir],
Tombari, F.[Federico],
Busam, B.[Benjamin],
Echoscene: Indoor Scene Generation via Information Echo Over Scene
Graph Diffusion,
ECCV24(XXI: 167-184).
Springer DOI
2412
BibRef
Tang, S.T.[Shi-Tao],
Chen, J.C.[Jia-Cheng],
Wang, D.[Dilin],
Tang, C.Z.[Cheng-Zhou],
Zhang, F.[Fuyang],
Fan, Y.C.[Yu-Chen],
Chandra, V.[Vikas],
Furukawa, Y.[Yasutaka],
Ranjan, R.[Rakesh],
Mvdiffusion++: A Dense High-resolution Multi-view Diffusion Model for
Single or Sparse-view 3d Object Reconstruction,
ECCV24(XVI: 175-191).
Springer DOI
2412
BibRef
Yan, Z.Z.[Zi-Zheng],
Zhou, J.P.[Jia-Peng],
Meng, F.[Fanpeng],
Wu, Y.[Yushuang],
Qiu, L.[Lingteng],
Ye, Z.[Zisheng],
Cui, S.G.[Shu-Guang],
Chen, G.Y.[Guan-Ying],
Han, X.G.[Xiao-Guang],
Dreamdissector: Learning Disentangled Text-to-3d Generation from 2d
Diffusion Priors,
ECCV24(XII: 124-141).
Springer DOI
2412
BibRef
Liu, Z.X.[Ze-Xiang],
Li, Y.G.[Yang-Guang],
Lin, Y.T.[You-Tian],
Yu, X.[Xin],
Peng, S.[Sida],
Cao, Y.P.[Yan-Pei],
Qi, X.J.[Xiao-Juan],
Huang, X.S.[Xiao-Shui],
Liang, D.[Ding],
Ouyang, W.L.[Wan-Li],
Unidream: Unifying Diffusion Priors for Relightable Text-to-3d
Generation,
ECCV24(V: 74-91).
Springer DOI
2412
BibRef
Liu, Q.H.[Qi-Hao],
Zhang, Y.[Yi],
Bai, S.[Song],
Kortylewski, A.[Adam],
Yuilie, A.[Alan],
DIRECT-3D: Learning Direct Text-to-3D Generation on Massive Noisy 3D
Data,
CVPR24(6881-6891)
IEEE DOI Code:
WWW Link.
2410
Geometry, Training, Solid modeling, Technological innovation,
Diffusion models, Data models
BibRef
Zhou, L.Q.[Lin-Qi],
Shih, A.[Andy],
Meng, C.L.[Chen-Lin],
Ermon, S.[Stefano],
DreamPropeller: Supercharge Text-to-3D Generation with Parallel
Sampling,
CVPR24(4610-4619)
IEEE DOI
2410
Pipelines, Graphics processing units, Diffusion models,
User experience, Text-to-3D, Diffusion models, Acceleration
BibRef
Chen, Y.[Yang],
Pant, Y.W.[Ying-Wei],
Yang, H.B.[Hai-Bo],
Yao, T.[Ting],
Meit, T.[Tao],
VP3D: Unleashing 2D Visual Prompt for Text-to-3D Generation,
CVPR24(4896-4905)
IEEE DOI Code:
WWW Link.
2410
Visualization, Solid modeling, Technological innovation,
Zero-shot learning, Semantics, Diffusion models, diffusion model,
image generation
BibRef
Ding, L.[Lihe],
Dong, S.C.[Shao-Cong],
Huang, Z.P.[Zhan-Peng],
Wang, Z.[Zibin],
Zhang, Y.Y.[Yi-Yuan],
Gong, K.X.[Kai-Xiong],
Xu, D.[Dan],
Xue, T.F.[Tian-Fan],
Text-to-3D Generation with Bidirectional Diffusion Using Both 2D and
3D Priors,
CVPR24(5115-5124)
IEEE DOI Code:
WWW Link.
2410
Training, Geometry, Bridges, Solid modeling, Costs, 3D generation
BibRef
Ling, H.[Huan],
Kim, S.W.[Seung Wook],
Torralba, A.[Antonio],
Fidler, S.[Sanja],
Kreis, K.[Karsten],
Align Your Gaussians: Text-to-4D with Dynamic 3D Gaussians and
Composed Diffusion Models,
CVPR24(8576-8588)
IEEE DOI
2410
Training, Visualization, Deformation, Dynamics, Diffusion models, Animation,
text-to-4d, generative models, 3d generation, score distillation sampling
BibRef
Qiu, L.T.[Ling-Teng],
Chen, G.Y.[Guan-Ying],
Gu, X.D.[Xiao-Dong],
Zuo, Q.[Qi],
Xu, M.[Mutian],
Wu, Y.[Yushuang],
Yuan, W.H.[Wei-Hao],
Dong, Z.L.[Zi-Long],
Bo, L.[Liefeng],
Han, X.G.[Xiao-Guang],
RichDreamer: A Generalizable Normal-Depth Diffusion Model for Detail
Richness in Text-to-3D,
CVPR24(9914-9925)
IEEE DOI Code:
WWW Link.
2410
Geometry, Training, Computational modeling, Pipelines, Lighting,
Diffusion models
BibRef
Liu, Y.T.[Ying-Tian],
Guo, Y.C.[Yuan-Chen],
Luo, G.[Guan],
Sun, H.[Heyi],
Yin, W.[Wei],
Zhang, S.H.[Song-Hai],
PI3D: Efficient Text-to-3D Generation with Pseudo-Image Diffusion,
CVPR24(19915-19924)
IEEE DOI
2410
Training, Solid modeling, Adaptation models, Shape, Image synthesis,
Text to image
BibRef
Liu, F.F.[Fang-Fu],
Wu, D.K.[Dian-Kun],
Wei, Y.[Yi],
Rao, Y.M.[Yong-Ming],
Duan, Y.Q.[Yue-Qi],
Sherpa3D: Boosting High-Fidelity Text-to-3D Generation via Coarse 3D
Prior,
CVPR24(20763-20774)
IEEE DOI Code:
WWW Link.
2410
Solid modeling, Computational modeling, Semantics, Coherence,
Diffusion models, 3D Generation, Diffusion Models
BibRef
Chen, C.[Cheng],
Yang, X.F.[Xiao-Feng],
Yang, F.[Fan],
Feng, C.Z.[Cheng-Zeng],
Fu, Z.[Zhoujie],
Foo, C.S.[Chuan-Sheng],
Lin, G.S.[Guo-Sheng],
Liu, F.[Fayao],
Sculpt3D: Multi-View Consistent Text-to-3D Generation with Sparse 3D
Prior,
CVPR24(10228-10237)
IEEE DOI Code:
WWW Link.
2410
Geometry, Legged locomotion, Accuracy, Shape, Pipelines, Text-to-3D,
2D diffusion, 3D Generation
BibRef
Wu, Z.[Zike],
Zhou, P.[Pan],
Yi, X.Y.[Xuan-Yu],
Yuan, X.D.[Xiao-Ding],
Zhang, H.W.[Han-Wang],
Consistent3D: Towards Consistent High-Fidelity Text-to-3D Generation
with Deterministic Sampling Prior,
CVPR24(9892-9902)
IEEE DOI Code:
WWW Link.
2410
Training, Geometry, Solid modeling, Stochastic processes,
Ordinary differential equations, Mathematical models,
Diffusion Model
BibRef
LU, Y.Y.X.[Yuan-Yan-Xun],
Zhang, J.Y.[Jing-Yang],
Li, S.W.[Shi-Wei],
Fang, T.[Tian],
McKinnon, D.[David],
Tsin, Y.H.[Yang-Hai],
Quarn, L.[Long],
Cao, X.[Xun],
Yao, Y.[Yao],
Direct2.5: Diverse Text-to-3D Generation via Multi-view 2.5D
Diffusion,
CVPR24(8744-8753)
IEEE DOI Code:
WWW Link.
2410
Geometry, Solid modeling, Image synthesis, Computational modeling,
Pipelines, Diffusion models
BibRef
Huang, T.Y.[Tian-Yu],
Zeng, Y.[Yihan],
Zhang, Z.[Zhilu],
Xu, W.[Wan],
Xu, H.[Hang],
Xu, S.[Songcen],
Lau, R.W.H.[Rynson W. H.],
Zuo, W.M.[Wang-Meng],
DreamControl: Control-Based Text-to-3D Generation with 3D Self-Prior,
CVPR24(5364-5373)
IEEE DOI Code:
WWW Link.
2410
Geometry, Measurement, Text to image, Neural radiance field, Diffusion models
BibRef
Liang, Y.X.[Yi-Xun],
Yang, X.[Xin],
Lin, J.T.[Jian-Tao],
Li, H.D.[Hao-Dong],
Xu, X.G.[Xiao-Gang],
Chen, Y.[Yingcong],
LucidDreamer: Towards High-Fidelity Text-to-3D Generation via
Interval Score Matching,
CVPR24(6517-6526)
IEEE DOI
2410
Training, Solid modeling, Pipelines, Rendering (computer graphics),
Neural radiance field, Distance measurement, 3D Generation, Diffusion Models
BibRef
Yi, T.[Taoran],
Fang, J.[Jiemin],
Wang, J.J.[Jun-Jie],
Wu, G.[Guanjun],
Xie, L.X.[Ling-Xi],
Zhang, X.P.[Xiao-Peng],
Liu, W.Y.[Wen-Yu],
Tian, Q.[Qi],
Wang, X.G.[Xing-Gang],
GaussianDreamer: Fast Generation from Text to 3D Gaussians by
Bridging 2D and 3D Diffusion Models,
CVPR24(6796-6807)
IEEE DOI Code:
WWW Link.
2410
Geometry, Solid modeling, Perturbation methods,
Graphics processing units, Diffusion models,Neural rendering
BibRef
Yang, J.[Jiayu],
Cheng, Z.[Ziang],
Duan, Y.F.[Yun-Fei],
Ji, P.[Pan],
Li, H.D.[Hong-Dong],
ConsistNet: Enforcing 3D Consistency for Multi-View Images Diffusion,
CVPR24(7079-7088)
IEEE DOI Code:
WWW Link.
2410
Solid modeling, Image synthesis, Computational modeling,
Graphics processing units, Diffusion models, latent diffusion model
BibRef
Wan, Z.Y.[Zi-Yu],
Paschalidou, D.[Despoina],
Huang, I.[Ian],
Liu, H.Y.[Hong-Yu],
Shen, B.[Bokui],
Xiang, X.Y.[Xiao-Yu],
Liao, J.[Jing],
Guibas, L.J.[Leonidas J.],
CAD: Photorealistic 3D Generation via Adversarial Distillation,
CVPR24(10194-10207)
IEEE DOI
2410
Training, Solid modeling, Interpolation, Pipelines, Diffusion models,
Rendering (computer graphics)
BibRef
Huang, X.[Xin],
Shao, R.Z.[Rui-Zhi],
Zhang, Q.[Qi],
Zhang, H.W.[Hong-Wen],
Feng, Y.[Ying],
Liu, Y.B.[Ye-Bin],
Wang, Q.[Qing],
HumanNorm: Learning Normal Diffusion Model for High-quality and
Realistic 3D Human Generation,
CVPR24(4568-4577)
IEEE DOI Code:
WWW Link.
2410
Geometry, Solid modeling, Text to image, Color, Diffusion models,
Diffusion Model, 3D Human, 3D Generation
BibRef
Dong, Y.[Yuan],
Zuo, Q.[Qi],
Gu, X.D.[Xiao-Dong],
Yuan, W.H.[Wei-Hao],
Zhao, Z.Y.[Zheng-Yi],
Dong, Z.L.[Zi-Long],
Bo, L.F.[Lie-Feng],
Huang, Q.X.[Qi-Xing],
GPLD3D: Latent Diffusion of 3D Shape Generative Models by Enforcing
Geometric and Physical Priors,
CVPR24(56-66)
IEEE DOI
2410
Solid modeling, Codes, Shape, Computational modeling,
Noise reduction, Shape Generative Model, Latent Diffusion, Quality Checker
BibRef
Lee, S.[Seoyoung],
Lee, J.[Joonseok],
PoseDiff: Pose-conditioned Multimodal Diffusion Model for Unbounded
Scene Synthesis from Sparse Inputs,
WACV24(5005-5015)
IEEE DOI
2404
Image color analysis, Computational modeling, Scalability, Cameras,
Tuning, Faces, Algorithms, Generative models for image, video, 3D, etc.,
Vision + language and/or other modalities
BibRef
Wang, H.[Hai],
Xiang, X.Y.[Xiao-Yu],
Fan, Y.C.[Yu-Chen],
Xue, J.H.[Jing-Hao],
Customizing 360-Degree Panoramas through Text-to-Image Diffusion
Models,
WACV24(4921-4931)
IEEE DOI Code:
WWW Link.
2404
Geometry, Codes, Noise reduction, Games, Task analysis, Algorithms,
Generative models for image, video, 3D, etc., Algorithms,
image and video synthesis
BibRef
Chen, M.H.[Ming-Hao],
Laina, I.[Iro],
Vedaldi, A.[Andrea],
Training-Free Layout Control with Cross-Attention Guidance,
WACV24(5331-5341)
IEEE DOI
2404
Training, Visualization, Layout, Semantics, Noise, Benchmark testing,
Algorithms, Generative models for image, video, 3D, etc
BibRef
Tang, J.[Junshu],
Wang, T.F.[Teng-Fei],
Zhang, B.[Bo],
Zhang, T.[Ting],
Yi, R.[Ran],
Ma, L.Z.[Li-Zhuang],
Chen, D.[Dong],
Make-It-3D: High-Fidelity 3D Creation from A Single Image with
Diffusion Prior,
ICCV23(22762-22772)
IEEE DOI
2401
BibRef
Szymanowicz, S.[Stanislaw],
Rupprecht, C.[Christian],
Vedaldi, A.[Andrea],
Viewset Diffusion: (0-)Image-Conditioned 3D Generative Models from 2D
Data,
ICCV23(8829-8839)
IEEE DOI
2401
BibRef
Jiang, Y.[Yutao],
Zhou, Y.[Yang],
Liang, Y.[Yuan],
Liu, W.X.[Wen-Xi],
Jiao, J.B.[Jian-Bo],
Quan, Y.H.[Yu-Hui],
He, S.F.[Sheng-Feng],
Diffuse3D: Wide-Angle 3D Photography via Bilateral Diffusion,
ICCV23(8964-8974)
IEEE DOI Code:
WWW Link.
2401
BibRef
Xiang, J.F.[Jian-Feng],
Yang, J.[Jiaolong],
Huang, B.B.[Bin-Bin],
Tong, X.[Xin],
3D-aware Image Generation using 2D Diffusion Models,
ICCV23(2383-2393)
IEEE DOI
2401
BibRef
Chai, L.[Lucy],
Tucker, R.[Richard],
Li, Z.Q.[Zheng-Qi],
Isola, P.[Phillip],
Snavely, N.[Noah],
Persistent Nature: A Generative Model of Unbounded 3D Worlds,
CVPR23(20863-20874)
IEEE DOI
2309
BibRef
Shim, J.[Jaehyeok],
Joo, K.[Kyungdon],
DITTO: Dual and Integrated Latent Topologies for Implicit 3D
Reconstruction,
CVPR24(5396-5405)
IEEE DOI
2410
Point cloud compression, Shape, Power system stability,
Transformers, Topology, ComputerVision
BibRef
Shim, J.[Jaehyeok],
Kang, C.W.[Chang-Woo],
Joo, K.[Kyungdon],
Diffusion-Based Signed Distance Fields for 3D Shape Generation,
CVPR23(20887-20897)
IEEE DOI
2309
BibRef
Po, R.[Ryan],
Wetzstein, G.[Gordon],
Compositional 3D Scene Generation using Locally Conditioned Diffusion,
3DV24(651-663)
IEEE DOI
2408
Semantics, Pipelines, Manuals, Task analysis
BibRef
Shue, J.R.[J. Ryan],
Chan, E.R.[Eric Ryan],
Po, R.[Ryan],
Ankner, Z.[Zachary],
Wu, J.J.[Jia-Jun],
Wetzstein, G.[Gordon],
3D Neural Field Generation Using Triplane Diffusion,
CVPR23(20875-20886)
IEEE DOI
2309
BibRef
Xu, J.[Jiale],
Wang, X.[Xintao],
Cheng, W.H.[Wei-Hao],
Cao, Y.P.[Yan-Pei],
Shan, Y.[Ying],
Qie, X.[Xiaohu],
Gao, S.H.[Sheng-Hua],
Dream3D: Zero-Shot Text-to-3D Synthesis Using 3D Shape Prior and
Text-to-Image Diffusion Models,
CVPR23(20908-20918)
IEEE DOI
2309
BibRef
Chapter on 3-D Object Description and Computation Techniques, Surfaces, Deformable, View Generation, Video Conferencing continues in
Merging Views, Object Insertion in Image .