Physics-Based Active Learning for Design Space Exploration and Surrogate Construction for Multiparametric Optimization

Sergio Torregrosa, Victor Champaney, Amine Ammar, Vincent Herbert, Francisco Chinesta

Communications on Applied Mathematics and Computation ›› 2024, Vol. 6 ›› Issue (3) : 1899-1923. DOI: 10.1007/s42967-023-00329-y
Review Article

Physics-Based Active Learning for Design Space Exploration and Surrogate Construction for Multiparametric Optimization

Author information +
History +

Abstract

The sampling of the training data is a bottleneck in the development of artificial intelligence (AI) models due to the processing of huge amounts of data or to the difficulty of access to the data in industrial practices. Active learning (AL) approaches are useful in such a context since they maximize the performance of the trained model while minimizing the number of training samples. Such smart sampling methodologies iteratively sample the points that should be labeled and added to the training set based on their informativeness and pertinence. To judge the relevance of a data instance, query rules are defined. In this paper, we propose an AL methodology based on a physics-based query rule. Given some industrial objectives from the physical process where the AI model is implied in, the physics-based AL approach iteratively converges to the data instances fulfilling those objectives while sampling training points. Therefore, the trained surrogate model is accurate where the potentially interesting data instances from the industrial point of view are, while coarse everywhere else where the data instances are of no interest in the industrial context studied.

Keywords

Active learning (AL) / Artificial intelligence (AI) / Optimization / Physics based

Cite this article

Download citation ▾
Sergio Torregrosa, Victor Champaney, Amine Ammar, Vincent Herbert, Francisco Chinesta. Physics-Based Active Learning for Design Space Exploration and Surrogate Construction for Multiparametric Optimization. Communications on Applied Mathematics and Computation, 2024, 6(3): 1899‒1923 https://doi.org/10.1007/s42967-023-00329-y

References

[1.]
Angluin D. Queries and concept learning. Mach. Learn., 1988, 2(4): 319-342,
CrossRef Google scholar
[2.]
Ash, J., Zhang, C., Krishnamurthy, A., Langford, J., Agarwal, A.: Deep batch active learning by diverse, uncertain gradient lower bounds. In: Proceedings of the 8th International Conference on Learning Representations, Addis Ababa, Ethiopia (2020)
[3.]
Azimi, J., Fern, A,. Fern, X., Borradaile, G., Heeringa, B.: Batch active learning via coordinated matching. In: Proceedings of the 29th International Conference on Machine Learning, Edinburgh, Scotland, UK (2012)
[4.]
Beluch, W., Genewein, T., Nurnberger, A., Kohler, J.M.: The power of ensembles for active learning in image classification. In: 2018 IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, pp. 9368–9377 (2018)
[5.]
Bilgic, M., Getoor, L.: Link-based active learning. In: NIPS Workshop on Analyzing Networks and Learning with Graphs (2009)
[6.]
Blitzer, J., Dredze, M., Pereira, F.: Biographies, bollywood, boom-boxes and blenders: domain adaptation for sentiment classification. In: Proceedings of the 45th Annual Meeting of the Association for Computational Linguistics, Prague, Czech Republic (2007)
[7.]
Bloodgood, M., Vijay-Shanker, K.: Taking into account the differences between actively and passively acquired data: the case of active learning with support vector machines for imbalanced datasets. In: Human Language Technologies: Conference of the North American Chapter of the Association of Computational Linguistics, Boulder, Colorado, USA, pp. 137–140 (2009)
[8.]
Brunton S, Proctor J, Kutz J. Discovering governing equations from data by sparse identification of nonlinear dynamical systems. Proc. Natl. Acad. Sci., 2016, 113(15): 3932-3937,
CrossRef Google scholar
[9.]
Chinesta F, Cueto E, Abisset-Chavanne E, et al.. Virtual, digital and hybrid twins: a new paradigm in data-based engineering and engineered data. Arch. Comput. Methods Eng., 2020, 27(1): 105-134,
CrossRef Google scholar
[10.]
Chinesta F, Huerta A, Rozza G, et al.. . Encyclopedia of Computational Mechanics, 2015 New York Wiley
[11.]
Dagan, I., Engelson, P.: Committee-based sampling for training probabilistic classifiers. In: Machine Learning, Proceedings of the Twelfth International Conference on Machine Learning, Tahoe City, California, USA, pp. 150–157 (1995)
[12.]
Ducoffe, M., Precioso, F.: Adversarial Active Learning for Deep Networks: a Margin Based Approach. arXiv:1802.09841 (2018)
[13.]
Freytag, A., Rodner, E., Denzler, J.: Selecting influential examples: active learning with expected model output changes. In: Computer Vision-ECCV 2014—13th European Conference, Zurich, Switzerland, pp. 562–577 (2014)
[14.]
Gal, Y., Islam, R., Ghahramani, Z.: Deep bayesian active learning with image data. In: Proceedings of the 34th International Conference on Machine Learning, Sydney, NSW, Australia, vol. 70, pp. 1183–1192 (2017)
[15.]
Geifman, Y., El-Yaniv, R.: Deep Active Learning over the Long Tail. arXiv:1711.00941 (2017)
[16.]
Guo, Y.: Active instance sampling via matrix partition. In: Advances in Neural Information Processing Systems 23: 24th Annual Conference on Neural Information Processing Systems, Vancouver, BC, Canada, pp. 802–810 (2010)
[17.]
Hernandez, Q., Badias, A., Gonzalez, D., Chinesta, F., Cueto, E.: Deep Learning of Thermodynamics-Aware Reduced-Order Models From Data. arXiv:2007.03758 (2020)
[18.]
Hernandez Q, D’Gonzalez A, et al.. Structure-preserving neural networks. J. Comput. Phys., 2021, 426,
CrossRef Google scholar
[19.]
Huang, J., Child, R., Rao, V., Liu, H., Satheesh, S., Coates, A.: Active Learning for Speech Recognition: the Power of Gradients. arXiv:1612.03226 (2016)
[20.]
Ibanez, R., Abisset-Chavanne, E., Ammar, A., González, D., Cueto, E., Huerta, A., Duval, J.L., Chinesta, F.: A multidimensional data-driven sparse identification technique: the sparse proper generalized decomposition. Complexity, 2018, 5608286 (2018). https://doi.org/10.1155/2018/5608286
[21.]
Ibanez R, Abisset-Chavanne E, Cueto E, et al.. Some applications of compressed sensing in computational mechanics: model order reduction, manifold learning, data-driven applications and nonlinear dimensionality reduction. Comput. Mech., 2019, 64(5): 1259-1271,
CrossRef Google scholar
[22.]
Joshi, A., Porikli, F., Papanikolopoulos, N.: Multi-class active learning for image classification. In: 2009 IEEE Conference on Computer Vision and Pattern Recognition, pp. 2372–2379 (2009)
[23.]
King R, Whelan K, Jones F, et al.. Functional genomic hypothesis generation and experimentation by a robot scientist. Nature, 2004, 427(6971): 247-252,
CrossRef Google scholar
[24.]
Krishnamurthy V. Algorithms for optimal scheduling and management of hidden Markov model sensors. IEEE Trans. Signal Process., 2002, 50(6): 1382-1397,
CrossRef Google scholar
[25.]
Laughlin R, Pines D. The theory of everything. Proc. Natl. Acad. Sci. USA, 2000, 97(1): 28,
CrossRef Google scholar
[26.]
Lewis, D., Gale, W.: A Sequential Algorithm for Training Text Classifiers, pp. 3–12 (1994)
[27.]
Loyola D, Pedergnana M, Gimeno S. Smart sampling and incremental function learning for very large high dimensional data. Neural Netw., 2015, 78: 75-87,
CrossRef Google scholar
[28.]
McKay MD, Conover WJ, Beckman R. A comparison of three methods for selecting values of input variables in the analysis of output from a computer code. Technometrics, 1979, 21: 239-245
[29.]
Moya B, Badias A, Alfaro I, Chinesta F. Digital twins that learn and correct themselves. Int. J. Numer. Methods Eng., 2022, 123: 3034-3044,
CrossRef Google scholar
[30.]
Nguyen, T., Smeulders, A.: Active learning using pre-clustering. In: ICML, pp. 79–79 (2004)
[31.]
Nocedal J, Wright S. . Numerical Optimization, 2000 Berlin Springer pp. 529-562
[32.]
Pinillo R, Abisset-Chavanne E, Ammar A, et al.. A multidimensional data-driven sparse identification technique: the sparse proper generalized decomposition. Complexity, 2018, 11: 1-11
[33.]
Ranganathan, H., Venkateswara, H., Chakraborty, S., Panchanathan, S.: Deep active learning for image classification. In: 2017 IEEE International Conference on Image Processing, pp. 3934–3938 (2017)
[34.]
Ren, P., Xiao, Y., Chang, X., Huang, P.-Y., Li, Z., Gupta, B.B., Chen, X., Wang, X.: A Survey of Deep Active Learning. arXiv:2009.00236 (2021)
[35.]
Roy, N., McCallum, A.: Toward optimal active learning through Monte Carlo estimation of error reduction. In: ICML, pp. 441–448 (2001)
[36.]
Sancarlos, A., Cameron, M., Abel, A., Cueto, E., Duval, J.-L., Chinesta, F.: From ROM of electrochemistry to AI-based battery digital and hybrid twin. In: Archives of Computational Methods in Engineering, pp. 1–37 (2020)
[37.]
Sancarlos, A., Champaney, V., Duval, J.L., Cueto, E., Chinesta, F.: PGD-based advanced nonlinear multiparametric regressions for constructing metamodels at the scarce-data limit. arXiv:2103.05358 (2021)
[38.]
Sener, O., Savarese, S.: Active learning for convolutiopnal neural networks: a core-set approach. arXiv:1708.00489 (2018)
[39.]
Settles, B.: Active Learning Literature Survey. Computer Sciences Technical Report 1648, University of Wisconsin-Madison (2010)
[40.]
Settles, B., Craven, M., Ray, S.: Multiple-instance active learning. Adv. Neural Info. Process. Syst. 20, 1289–1296 (2008)
[41.]
Seung, H., Opper, M., Sompolinsky, H.: Query by committee. In: Proceedings of the 5th Annual Workshop on Computational Learning Theory, pp. 287–294 (1992)
[42.]
Shui, C., Zhou, F., Gagne, C., Wang, B.: Deep active learning: unified and principled method for query and training. In: International Conference on Artificial Intelligence and Statistics, PMLR, pp. 1308–1318 (2020)
[43.]
Stein M. Large sample properties of simulations using latin hypercube sampling. Technometrics, 1987, 29(2): 143-151,
CrossRef Google scholar
[44.]
Szegedy, C., Zaremba, W., Sutskever, I., Bruna, J., Erhan, D., Ian Goodfellow, I., Fergus, R.: Intriguing Properties of Neural Networks. arXiv:1312.6199 (2014)
[45.]
Tong S, Koller D. Support vector machine active learning with applications to text classification. J. Mach. Learn. Res., 2002, 2(1): 45-66
[46.]
Torregrosa S, Champaney V, Ammar A, et al.. Surrogate parametric metamodel based on optimal transport. Math. Comput. Simul., 2021, 194: 36-63,
CrossRef Google scholar
[47.]
Udrescu, S., Tan, A., Feng, J., Neto, O., Wu, T., Tegmark, M.: Ai Feynman 2.0: Pareto-Optimal Symbolic Regression Exploiting Graph Modularity. arXiv:2006.10782 (2006)
[48.]
Yang, Y., Loog, M.: A Benchmark and Comparison of Active Learning for Logistic Regression. arXiv:1611.08618 (2018)
[49.]
Yin, C., Qian, B., Cao, S., et al.: Deep similarity-based batch mode active learning with exploration-exploitation. In: IEEE International Conference on Data Mining, pp. 575–584 (2017)
[50.]
Zhdanov, F.: Diverse Mini-Batch Active Learning. arXiv:1901.05954 (2019)

Accesses

Citations

Detail

Sections
Recommended

/