Machine Learning Assisted Exploration for Affine Deligne–Lusztig Varieties
Bin Dong , Xuhua He , Pengfei Jin , Felix Schremmer , Qingchao Yu
Peking Mathematical Journal ›› 2026, Vol. 9 ›› Issue (1) : 55 -104.
This paper presents a novel, interdisciplinary study that leverages a Machine Learning (ML) assisted framework to explore the geometry of affine Deligne–Lusztig varieties (ADLV). The primary objective is to investigate the non-emptiness pattern, dimension, and enumeration of irreducible components of ADLV. Our proposed framework demonstrates a recursive pipeline of data generation, model training, pattern analysis, and human examination, presenting an intricate interplay between ML and pure mathematical research. Notably, our data-generation process is nuanced, emphasizing the selection of meaningful subsets and appropriate feature sets. We demonstrate that this framework has a potential to accelerate pure mathematical research, leading to the discovery of new conjectures and promising research directions that could otherwise take significant time to uncover. We rediscover the virtual dimension formula and provide a full mathematical proof of a newly identified problem concerning a certain lower bound of dimension. Furthermore, we extend an open invitation to the readers by providing the source code for computing ADLV and the ML models, promoting further explorations. This paper concludes by sharing valuable experiences and highlighting lessons learned from this collaboration.
Affine Deligne–Lusztig varieties / Affine Weyl groups / Loop groups / AI-assisted math research
| [1] |
|
| [2] |
Bach, F., Jenatton, R., Mairal, J., Obozinski, G.: Convex optimization with sparsity-inducing norms. In: Optimization for Machine Learning, pp. 19–53. MIT Press, Cambridge, MA (2011) |
| [3] |
|
| [4] |
Berglund, P., Campbell, B., Jejjala, V.: Machine learning Kreuzer–Skarke Calabi–Yau threefolds. arXiv:2112.09117 (2021) |
| [5] |
|
| [6] |
|
| [7] |
Bruhat, F., Tits, J.: Groupes réductifs sur un corps local, II. Schémas en groupes. Existence d’une donnée radicielle valuée. Publ. Math. l’Inst. Hautes Études Sci. 60, 5–184 (1984) |
| [8] |
|
| [9] |
|
| [10] |
|
| [11] |
|
| [12] |
|
| [13] |
|
| [14] |
|
| [15] |
|
| [16] |
|
| [17] |
|
| [18] |
|
| [19] |
|
| [20] |
|
| [21] |
|
| [22] |
Han, S., Mao, H., Dally, W.J.: Deep compression: compressing deep neural networks with pruning, trained quantization and Huffman coding. arXiv:1510.00149 (2015) |
| [23] |
|
| [24] |
|
| [25] |
|
| [26] |
He, X.: Hecke algebras and p\documentclass[12pt]{minimal}\usepackage{amsmath}\usepackage{wasysym}\usepackage{amsfonts}\usepackage{amssymb}\usepackage{amsbsy}\usepackage{mathrsfs}\usepackage{upgreek}\setlength{\oddsidemargin}{-69pt}\begin{document}$$p$$\end{document}-adic groups. In: Current Developments in Mathematics 2015, pp. 73–135. International Press, Somerville, MA (2016) |
| [27] |
He, X.: Some results on affine Deligne–Lusztig varieties. In: Proceedings of the International Congress of Mathematicians—Rio de Janeiro 2018. Vol. II. Invited Lectures, pp. 1345–1365. World Scientific, Hackensack, NJ (2018) |
| [28] |
|
| [29] |
|
| [30] |
He, X., Nie, S., Yu, Q.: Affine Deligne–Lusztig varieties with finite Coxeter parts. arXiv:2208.14058 (2022) |
| [31] |
|
| [32] |
Ho, J., Jain, A., Abbeel, P.: Denoising diffusion probabilistic models. In: Advances in Neural Information Processing Systems 33 (NeurIPS 2020), pp. 6840–6851. Curran Associates Inc., Red Hook, NY (2020) |
| [33] |
Jia, W., Wang, H., Chen, M., Lu, D., Lin, L., Car, R., E, W., Zhang, L.: Pushing the limit of molecular dynamics with ab initio accuracy to 100 million atoms with machine learning. In: SC20: International Conference for High Performance Computing, Networking, Storage and Analysis, pp. 1–14. IEEE (2020) |
| [34] |
Jumper, J., Evans, R., Pritzel, A., Green, T., Figurnov, M., Ronneberger, O., Tunyasuvunakool, K., Bates, R., Zidek, A., Potapenko, A., et al.: Highly accurate protein structure prediction with AlphaFold. Nature 596(7873), 583–589 (2021) |
| [35] |
|
| [36] |
Kingma, D.P., Ba, J.: Adam: a method for stochastic optimization. arXiv:1412.6980 (2014) |
| [37] |
|
| [38] |
|
| [39] |
|
| [40] |
|
| [41] |
|
| [42] |
Loshchilov, I., Hutter, F.: Decoupled weight decay regularization. arXiv:1711.05101 (2017) |
| [43] |
|
| [44] |
|
| [45] |
|
| [46] |
|
| [47] |
Nair, V., Hinton, G.E.: Rectified linear units improve restricted Boltzmann machines. In: Proceedings of the 27th International Conference on Machine Learning (ICML’10), pp. 807–814. Omnipress, Madison, WI (2010) |
| [48] |
|
| [49] |
|
| [50] |
|
| [51] |
|
| [52] |
|
| [53] |
Ruder, S.: An overview of gradient descent optimization algorithms. arXiv:1609.04747 (2016) |
| [54] |
|
| [55] |
|
| [56] |
Schremmer, F.: Generic Newton points and cordial elements. arXiv:2205.02039 (2022) |
| [57] |
|
| [58] |
|
| [59] |
|
| [60] |
|
| [61] |
|
| [62] |
Viehmann, E.: On the geometry of the Newton stratification. In: Shimura Varieties, pp. 192–208. Cambridge University Press, Cambridge (2020) |
| [63] |
|
| [64] |
|
| [65] |
|
| [66] |
Williamson, G.: Is deep learning a useful tool for the pure mathematician? Bull. Am. Math. Soc. 61(2), 271–286 (2024) |
| [67] |
|
| [68] |
Zhao, W.X., Zhou, K., Li, J., Tang, T., Wang, X., Hou, Y., Min, Y., Zhang, B., Zhang, J., Dong, Z., et al.: A survey of large language models. arXiv:2303.18223 (2023) |
| [69] |
|
| [70] |
|
The Author(s)
/
| 〈 |
|
〉 |