3D asset generation: a survey of evolution towards autoregressive and agent-driven paradigms
Hongxing FAN , Haohua CHEN , Zehuan HUANG , Ziwei LIU , Lu SHENG
Front. Comput. Sci. ›› 2026, Vol. 20 ›› Issue (11) : 2011710
3D asset generation: a survey of evolution towards autoregressive and agent-driven paradigms
Generating high-quality 3D assets is a fundamental challenge in computer vision and graphics. While the field has progressed significantly from early VAE/GAN approaches through diffusion models and large reconstruction models, persistent limitations hinder widespread application. Specifically, achieving high geometric and appearance fidelity, intuitive user control, versatile multi-modal conditioning, and directly usable outputs (e.g., structured meshes) remains challenging for established paradigms. This paper surveys the evolution of deep generative models for 3D content creation, with a primary focus on emerging paradigms: autoregressive (AR) generation and Agent-driven approaches, poised to address aforementioned shortcomings. AR models generate assets sequentially (e.g., token-by-token or part-by-part), offering inherent potential for finer control, structured outputs, and integrating user guidance during the step-by-step process. Agent-driven methods, conversely, leverage the reasoning and linguistic capabilities of Large Language Models (LLMs), enabling intuitive and flexible 3D creation by decomposing complex tasks and utilizing external tools through multi-agent systems. We provide a comprehensive overview of these novel techniques, discuss their potential advantages over current methods, and outline key challenges and future directions towards more capable and intelligent 3D generation systems.
3D generation paradigms / autoregressive models / agent-driven 3D generation
| [1] |
Tan Q, Gao L, Lai Y K, Xia S. Variational autoencoders for deforming 3D mesh models. In: Proceedings of 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition. 2018, 5841−5850 |
| [2] |
|
| [3] |
|
| [4] |
|
| [5] |
|
| [6] |
|
| [7] |
|
| [8] |
Gu J, Liu L, Wang P, Theobalt C. StyleNeRF: a style-based 3D aware generator for high-resolution image synthesis. In: Proceedings of the 10th International Conference on Learning Representations. 2022 |
| [9] |
|
| [10] |
|
| [11] |
|
| [12] |
Hong Y, Zhang K, Gu J, Bi S, Zhou Y, Liu D, Liu F, Sunkavalli K, Bui T, Tan H. LRM: large reconstruction model for single image to 3D. In: Proceedings of the 12th International Conference on Learning Representations. 2024 |
| [13] |
Wang P, Tan H, Bi S, Xu Y, Luan F, Sunkavalli K, Wang W, Xu Z, Zhang K. PF-LRM: pose-free large reconstruction model for joint pose and shape prediction. In: Proceedings of the 12th International Conference on Learning Representations. 2024 |
| [14] |
Li J, Tan H, Zhang K, Xu Z, Luan F, Xu Y, Hong Y, Sunkavalli K, Shakhnarovich G, Bi S. Instant3D: fast text-to-3D with sparse-view generation and large reconstruction model. In: Proceedings of the 12th International Conference on Learning Representations. 2024 |
| [15] |
|
| [16] |
|
| [17] |
|
| [18] |
Poole B, Jain A, Barron J T, Mildenhall B. DreamFusion: text-to-3D using 2D diffusion. In: Proceedings of the 11th International Conference on Learning Representations. 2023 |
| [19] |
Liu R, Wu R, Van Hoorick B, Tokmakov P, Zakharov S, Vondrick C. Zero-1-to-3: zero-shot one image to 3D object. In: Proceedings of 2023 IEEE/CVF International Conference on Computer Vision. 2023, 9264−9275 |
| [20] |
Shi Y, Wang P, Ye J, Mai L, Li K, Yang X. MVDream: multi-view diffusion for 3D generation. In: Proceedings of the 12th International Conference on Learning Representations. 2024 |
| [21] |
|
| [22] |
|
| [23] |
|
| [24] |
Liu Y, Lin C, Zeng Z, Long X, Liu L, Komura T, Wang W. SyncDreamer: generating multiview-consistent images from a single-view image. In: Proceedings of the 12th International Conference on Learning Representations. 2024 |
| [25] |
|
| [26] |
|
| [27] |
|
| [28] |
Huang Z, Wen H, Dong J, Wang Y, Li Y, Chen X, Cao Y P, Liang D, Qiao Y, Dai B, Sheng L. EpiDiff: enhancing multi-view synthesis via localized epipolar-constrained diffusion. In: Proceedings of 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition. 2024, 9784−9794 |
| [29] |
|
| [30] |
Wen H, Huang Z, Wang Y, Chen X, Sheng L. Ouroboros3D: image-to-3D generation via 3D-aware recursive diffusion. In: Proceedings of 2025 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR). 2025, 21631−21641 |
| [31] |
Chen Z, Tang J, Dong Y, Cao Z, Hong F, Lan Y, Wang T, Xie H, Wu T, Saito S, Pan L, Lin D, Liu Z. 3DTopia-XL: scaling high-quality 3D asset generation via primitive diffusion. In: Proceedings of 2025 IEEE/CVF Conference on Computer Vision and Pattern Recognition. 2025, 26576−26586 |
| [32] |
|
| [33] |
|
| [34] |
|
| [35] |
|
| [36] |
|
| [37] |
|
| [38] |
Liu Q, Zhang Y, Bai S, Kortylewski A, Yuilie A. DIRECT-3D: learning direct text-to-3D generation on massive noisy 3D data. In: Proceedings of 2024 IEEE/CVF Conference on Computer Vision and Pattern Recognition. 2024, 6881−6891 |
| [39] |
|
| [40] |
|
| [41] |
|
| [42] |
Ren X, Huang J, Zeng X, Museth K, Fidler S, Williams F. XCube: large-scale 3D generative modeling using sparse voxel hierarchies. In: Proceedings of 2024 IEEE/CVF Conference on Computer Vision and Pattern Recognition. 2024, 4209−4219 |
| [43] |
Liu Z, Feng Y, Black M J, Nowrouzezahrai D, Paull L, Liu W. MeshDiffusion: score-based generative 3D mesh modeling. In: Proceedings of the 11th International Conference on Learning Representations. 2023 |
| [44] |
|
| [45] |
|
| [46] |
Chen H, Wang P, Wang F, Tian W, Xiong L, Li H. EPro-PnP: generalized end-to-end probabilistic perspective-n-points for monocular object pose estimation. In: Proceedings of 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition. 2022, 2771−2780 |
| [47] |
|
| [48] |
|
| [49] |
|
| [50] |
Siddiqui Y, Alliegro A, Artemov A, Tommasi T, Sirigatti D, Rosov V, Dai A, Nießner M. MeshGPT: generating triangle meshes with decoder-only transformers. In: Proceedings of 2024 IEEE/CVF Conference on Computer Vision and Pattern Recognition. 2024, 19615−19625 |
| [51] |
|
| [52] |
|
| [53] |
Sun C, Han J, Deng W, Wang X, Qin Z, Gould S. 3D-GPT: procedural 3D modeling with large language models. In: Proceedings of 2025 International Conference on 3D Vision (3DV). 2025, 1253–1263 |
| [54] |
|
| [55] |
Qi C R, Su H, Kaichun M, Guibas L J. PointNet: deep learning on point sets for 3D classification and segmentation. In: Proceedings of 2017 IEEE Conference on Computer Vision and Pattern Recognition. 2017, 77−85 |
| [56] |
|
| [57] |
|
| [58] |
|
| [59] |
Wu Z, Song S, Khosla A, Yu F, Zhang L, Tang X, Xiao J. 3D ShapeNets: a deep representation for volumetric shapes. In: Proceedings of 2015 IEEE Conference on Computer Vision and Pattern Recognition. 2015, 1912−1920 |
| [60] |
Maturana D, Scherer S. VoxNet: a 3D Convolutional Neural Network for real-time object recognition. In: Proceedings of 2015 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS). 2015, 922−928 |
| [61] |
|
| [62] |
|
| [63] |
|
| [64] |
|
| [65] |
Wu G, Yi T, Fang J, Xie L, Zhang X, Wei W, Liu W, Tian Q, Wang X. 4D Gaussian splatting for real-time dynamic scene rendering. In: Proceedings of 2024 IEEE/CVF Conference on Computer Vision and Pattern Recognition. 2024, 20310−20320 |
| [66] |
|
| [67] |
Mescheder L, Oechsle M, Niemeyer M, Nowozin S, Geiger A. Occupancy networks: learning 3D reconstruction in function space. In: Proceedings of 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition. 2019, 4455−4465 |
| [68] |
Park J J, Florence P, Straub J, Newcombe R, Lovegrove S. DeepSDF: learning continuous signed distance functions for shape representation. In: Proceedings of 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition. 2019, 165−174 |
| [69] |
|
| [70] |
|
| [71] |
|
| [72] |
|
| [73] |
|
| [74] |
Su H, Maji S, Kalogerakis E, Learned-Miller E. Multi-view convolutional neural networks for 3D shape recognition. In: Proceedings of 2015 IEEE International Conference on Computer Vision. 2015, 945−953 |
| [75] |
Chen X, Ma H, Wan J, Li B, Xia T. Multi-view 3D object detection network for autonomous driving. In: Proceedings of 2017 IEEE Conference on Computer Vision and Pattern Recognition. 2017, 6526−6534 |
| [76] |
|
| [77] |
|
| [78] |
Newcombe R A, Izadi S, Hilliges O, Molyneaux D, Kim D, Davison A J, Kohi P, Shotton J, Hodges S, Fitzgibbon A. KinectFusion: real-time dense surface mapping and tracking. In: Proceedings of 2011 10th IEEE International Symposium on Mixed and Augmented Reality. 2011, 127−136 |
| [79] |
|
| [80] |
Lin C H, Gao J, Tang L, Takikawa T, Zeng X, Huang X, Kreis K, Fidler S, Liu M Y, Lin T Y. Magic3D: high-resolution text-to-3D content creation. In: Proceedings of 2023 IEEE/CVF Conference on Computer Vision and Pattern Recognition. 2023, 300−309 |
| [81] |
|
| [82] |
|
| [83] |
Shu D, Park S W, Kwon J. 3D point cloud generative adversarial network based on tree structured graph convolutions. In: Proceedings of 2019 IEEE/CVF International Conference on Computer Vision. 2019, 3858−3867 |
| [84] |
|
| [85] |
|
| [86] |
Ramasinghe S, Khan S, Barnes N, Gould S. Spectral-GANs for high-resolution 3D point-cloud generation. In: Proceedings of 2020 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS). 2020, 8169−8176 |
| [87] |
|
| [88] |
|
| [89] |
|
| [90] |
|
| [91] |
Khan S H, Guo Y, Hayat M, Barnes N. Unsupervised primitive discovery for improved 3D generative modeling. In: Proceedings of 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition. 2019, 9731−9740 |
| [92] |
Luo A, Li T, Zhang W H, Lee T S. SurfGen: adversarial 3D shape synthesis with explicit surface discriminators. In: Proceedings of 2021 IEEE/CVF International Conference on Computer Vision. 2021, 16218−16228 |
| [93] |
Wu R, Zhuang Y, Xu K, Zhang H, Chen B. PQ-NET: a generative part Seq2Seq network for 3D shapes. In: Proceedings of 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition. 2020, 826−835 |
| [94] |
|
| [95] |
|
| [96] |
Nguyen-Phuoc T, Li C, Theis L, Richardt C, Yang Y L. HoloGAN: unsupervised learning of 3D representations from natural images. In: Proceedings of 2019 IEEE/CVF International Conference on Computer Vision. 2019, 7587−7596 |
| [97] |
|
| [98] |
Cai S, Obukhov A, Dai D, Van Gool L. Pix2NeRF: unsupervised conditional π-GAN for single image to neural radiance fields translation. In: Proceedings of 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition. 2022, 3971−3980 |
| [99] |
|
| [100] |
Tewari A, R M B, Pan X, Fried O, Agrawala M, Theobalt C. Disentangled3D: learning a 3D generative model with disentangled geometry and appearance from monocular images. In: Proceedings of 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition. 2022, 1506−1515 |
| [101] |
|
| [102] |
Sun J, Wang X, Zhang Y, Li X, Zhang Q, Liu Y, Wang J. FENeRF: face editing in neural radiance fields. In: Proceedings of 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition. 2022, 7662−7672 |
| [103] |
Xu Y, Peng S, Yang C, Shen Y, Zhou B. 3D-aware image synthesis via learning structural and textural representations. In: Proceedings of 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition. 2022, 18409−18418 |
| [104] |
Sun J, Wang X, Wang L, Li X, Zhang Y, Zhang H, Liu Y. Next3D: generative neural texture rasterization for 3D-aware head avatars. In: Proceedings of 2023 IEEE/CVF Conference on Computer Vision and Pattern Recognition. 2023, 20991−21002 |
| [105] |
|
| [106] |
|
| [107] |
Deng Y, Yang J, Xiang J, Tong X. GRAM: generative radiance manifolds for 3D-aware image generation. In: Proceedings of 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition. 2022, 10663−10673 |
| [108] |
|
| [109] |
Yang G, Huang X, Hao Z, Liu M Y, Belongie S, Hariharan B. PointFlow: 3D point cloud generation with continuous normalizing flows. In: Proceedings of 2019 IEEE/CVF International Conference on Computer Vision. 2019, 4540−4549 |
| [110] |
|
| [111] |
|
| [112] |
Mittal P, Cheng Y C, Singh M, Tulsiani S. AutoSDF: shape priors for 3D completion, reconstruction and generation. In: Proceedings of 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition. 2022, 306−315 |
| [113] |
|
| [114] |
|
| [115] |
Chen R, Zhang J, Liang Y, Luo G, Li W, Liu J, Li X, Long X, Feng J, Tan P. Dora: sampling and benchmarking for 3D shape Variational Auto-Encoders. In: Proceedings of 2025 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR). 2025, 16251−16261 |
| [116] |
Wang H, Du X, Li J, Yeh R A, Shakhnarovich G. Score Jacobian chaining: lifting pretrained 2D diffusion models for 3D generation. In: Proceedings of 2023 IEEE/CVF Conference on Computer Vision and Pattern Recognition. 2023, 12619−12629 |
| [117] |
ang P, Xu D, Fan Z, Wang D, Mohan S, Iandola F, Ranjan R, Li Y, Liu Q, Wang Z, Chandra V. Taming mode collapse in score distillation for text-to-3D generation. In: Proceedings of the 2024 IEEE/CVF Conference on Computer Vision and Pattern Recognition. 2024, 9037−9047 |
| [118] |
|
| [119] |
Raj A, Kaza S, Poole B, Niemeyer M, Ruiz N, Mildenhall B, Zada S, Aberman K, Rubinstein M, Barron J, Li Y, Jampani V. DreamBooth3D: subject-driven text-to-3D generation. In: Proceedings of 2023 IEEE/CVF International Conference on Computer Vision. 2023, 2349−2359 |
| [120] |
Zhang J, Li X, Zhang Q, Cao Y, Shan Y, Liao J. HumanRef: single image to 3D human generation via reference-guided diffusion. In: Proceedings of 2024 IEEE/CVF Conference on Computer Vision and Pattern Recognition. 2024, 1844−1854 |
| [121] |
Tsalicoglou C, Manhardt F, Tonioni A, Niemeyer M, Tombari F. TextMesh: generation of realistic 3D meshes from text prompts. In: Proceedings of 2024 International Conference on 3D Vision (3DV). 2024, 1554−1563 |
| [122] |
Metzer G, Richardson E, Patashnik O, Giryes R, Cohen-Or D. Latent-NeRF for shape-guided generation of 3D shapes and textures. In: Proceedings of 2023 IEEE/CVF Conference on Computer Vision and Pattern Recognition. 2023, 12663−12673 |
| [123] |
|
| [124] |
Tang J, Wang T, Zhang B, Zhang T, Yi R, Ma L, Chen D. Make-it-3D: high-fidelity 3D creation from a single image with diffusion prior. In: Proceedings of 2023 IEEE/CVF International Conference on Computer Vision. 2023, 22762−22772 |
| [125] |
Qian G, Mai J, Hamdi A, Ren J, Siarohin A, Li B, Lee H Y, Skorokhodov I, Wonka P, Tulyakov S, Ghanem B. Magic123: one image to high-quality 3D object generation using both 2D and 3D diffusion priors. In: Proceedings of the 12th International Conference on Learning Representations. 2024 |
| [126] |
|
| [127] |
|
| [128] |
|
| [129] |
Luo S, Hu W. Diffusion probabilistic models for 3D point cloud generation. In: Proceedings of 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition. 2021, 2836−2844 |
| [130] |
|
| [131] |
Xiang J, Lv Z, Xu S, Deng Y, Wang R, Zhang B, Chen D, Tong X, Yang J. Structured 3D latents for scalable and versatile 3D generation. In: Proceedings of 2025 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR). 2025, 21469−21480 |
| [132] |
|
| [133] |
|
| [134] |
|
| [135] |
|
| [136] |
Richardson E, Metzer G, Alaluf Y, Giryes R, Cohen-Or D. TEXTure: text-guided texturing of 3D shapes. In: Proceedings of 2023 ACM SIGGRAPH Conference Proceedings. 2023, 54 |
| [137] |
Zeng X, Chen X, Qi Z, Liu W, Zhao Z, Wang Z, Fu B, Liu Y, Yu G. Paint3D: paint anything 3D with lighting-less texture diffusion models. In: Proceedings of 2014 IEEE/CVF Conference on Computer Vision and Pattern Recognition. 2024, 4252−4262 |
| [138] |
|
| [139] |
|
| [140] |
|
| [141] |
Xu D, Yuan Y, Mardani M, Liu S, Song J, Wang Z, Vahdat A. AGG: amortized generative 3D Gaussians for single image to 3D. Transactions on Machine Learning Research, 2024 |
| [142] |
Szymanowicz S, Rupprecht C, Vedaldi A. Splatter image: ultra-fast single-view 3D reconstruction. In: Proceedings of 2024 IEEE/CVF Conference on Computer Vision and Pattern Recognition. 2024, 10208−10217 |
| [143] |
|
| [144] |
|
| [145] |
|
| [146] |
Xiong B, Liu J, Hu J, Wu C, Wu J, Liu X, Zhao C, Ding E, Lian Z. TexGaussian: generating high-quality PBR material via octree-based 3D Gaussian splatting. In: Proceedings of 2025 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR). 2025, 551−561 |
| [147] |
|
| [148] |
Chen Y, He T, Huang D, Ye W, Chen S, Tang J, Cai Z, Yang L, Yu G, Lin G, Zhang C. MeshAnything: artist-created mesh generation with autoregressive transformers. In: Proceedings of the 13th International Conference on Learning Representations. 2025 |
| [149] |
Zhang J, Xiong F, Wang G, Xu M. G3PT: unleash the power of autoregressive modeling in 3D generation via cross-scale querying transformer. In: Proceedings of the 34th International Joint Conference on Artificial Intelligence. 2025, 262 |
| [150] |
Tang J, Li Z, Hao Z, Liu X, Zeng G, Liu M Y, Zhang Q. EdgeRunner: auto-regressive auto-encoder for artistic mesh generation. In: Proceedings of the 13th International Conference on Learning Representations. 2025 |
| [151] |
Weng H, Zhao Z, Lei B, Yang X, Liu J, Lai Z, Chen Z, Liu Y, Jiang J, Guo C, Zhang T, Gao S, Chen C L P. Scaling mesh generation via compressive tokenization. In: Proceedings of 2025 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR). 2025, 11093−11103 |
| [152] |
Chen Y, Lan Y, Zhou S, Wang T, Pan X. SAR3D: autoregressive 3D object generation and understanding via multi-scale 3D VQVAE. In: Proceedings of 2025 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR). 2025, 28371−28382 |
| [153] |
|
| [154] |
|
| [155] |
|
| [156] |
Gao D, Siddiqui Y, Li L, Dai A. MeshArt: generating articulated meshes with structure-guided transformers. In: Proceedings of 2025 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR). 2025, 618−627 |
| [157] |
|
| [158] |
|
| [159] |
|
| [160] |
|
| [161] |
|
| [162] |
|
| [163] |
|
| [164] |
Lionar S, Liang J, Lee G H. TreeMeshGPT: artistic mesh generation with autoregressive tree sequencing. In: Proceedings of 2025 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR). 2025, 26608−26617 |
| [165] |
|
| [166] |
|
| [167] |
|
| [168] |
|
| [169] |
Zou Z X, Yu Z, Guo Y C, Li Y, Liang D, Cao Y P, Zhang S H. Triplane meets Gaussian splatting: fast and generalizable single-view 3D reconstruction with transformers. In: Proceedings of 2024 IEEE/CVF Conference on Computer Vision and Pattern Recognition. 2024, 10324−10335 |
| [170] |
|
| [171] |
|
| [172] |
|
| [173] |
|
| [174] |
|
| [175] |
Peebles W, Xie S. Scalable diffusion models with transformers. In: Proceedings of 2023 IEEE/CVF International Conference on Computer Vision. 2023, 4172−4182 |
| [176] |
|
| [177] |
|
| [178] |
Zhou L, Du Y, Wu J. 3D shape generation and completion through point-voxel diffusion. In: Proceedings of 2021 IEEE/CVF International Conference on Computer Vision. 2021, 5806−5815 |
| [179] |
Huang Z, Guo Y C, An X, Yang Y, Li Y, Zou Z X, Liang D, Liu X, Cao Y P, Sheng L. MIDI: multi-instance diffusion for single image to 3D scene generation. In: Proceedings of 2025 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR). 2025, 23646−23657 |
| [180] |
|
| [181] |
|
| [182] |
|
| [183] |
|
| [184] |
|
| [185] |
Zhang H, Li F, Liu S, Zhang L, Su H, Zhu J, Ni L, Shum H Y. DINO: DETR with improved DeNoising anchor boxes for end-to-end object detection. In: Proceedings of the 11th International Conference on Learning Representations. 2023 |
| [186] |
|
| [187] |
|
| [188] |
|
| [189] |
|
| [190] |
|
| [191] |
|
| [192] |
|
| [193] |
|
| [194] |
|
| [195] |
|
| [196] |
|
| [197] |
|
| [198] |
Pope R, Douglas S, Chowdhery A, Devlin J, Bradbury J, Heek J, Xiao K, Agrawal S, Dean J. Efficiently scaling transformer inference. In: Proceedings of the 6th Conference on Machine Learning and Systems, 2023, 606−624 |
| [199] |
|
| [200] |
Ainslie J, Lee-Thorp J, de Jong M, Zemlyanskiy Y, Lebron F, Sanghai S. GQA: training generalized multi-query transformer models from multi-head checkpoints. In: Proceedings of 2023 Conference on Empirical Methods in Natural Language Processing. 2023, 4895−4901 |
| [201] |
Guo T, Chen X, Wang Y, Chang R, Pei S, Chawla N V, Wiest O, Zhang X. Large language model based multi-agents: a survey of progress and challenges. In: Proceedings of the 33rd International Joint Conference on Artificial Intelligence. 2024, 890 |
| [202] |
|
| [203] |
|
| [204] |
|
| [205] |
Jayaraman P K, Lambourne J G, Desai N, Willis K, Sanghi A, Morris N J W. SolidGen: an autoregressive model for direct B-rep synthesis. Transactions on Machine Learning Research, 2023 |
| [206] |
Zhou S, Tang T, Zhou B. CADParser: a learning approach of sequence modeling for B-rep CAD. In: Proceedings of the 32nd International Joint Conference on Artificial Intelligence. 2023, 1804−1812 |
| [207] |
|
| [208] |
|
| [209] |
|
| [210] |
|
| [211] |
|
Higher Education Press
/
| 〈 |
|
〉 |