When to use machine learning? And which problems stand to benefit?
M. Z. Naser
Urban Lifeline ›› 2026, Vol. 4 ›› Issue (1) : 3
When to use machine learning? And which problems stand to benefit?
Despite the continued success of machine learning (ML), its indiscriminate use may not always yield the expected benefits, can waste resources, introduce unforeseen complexities, and even lead to failure. From this lens, this paper systematically addresses the following elemental questions: When should we integrate ML, and which problem-solving workflows benefit the most? Thus, this paper critically examines existing decision making frameworks and guidelines to identify their strengths and limitations in defining ML applicability. Then, through a comprehensive literature review and analysis, we outline explicit suitability criteria to assess when ML should be favored over traditional engineering methods. Based on this analysis, we propose a conceptual structured decision making framework that incorporates practical steps and checklists to guide engineers in method selection, model development, validation, and deployment. Additionally, this paper highlights critical challenges and potential future research questions.
Engineering / Artificial intelligence / Prediction
| [1] |
|
| [2] |
|
| [3] |
|
| [4] |
|
| [5] |
|
| [6] |
|
| [7] |
|
| [8] |
|
| [9] |
|
| [10] |
|
| [11] |
|
| [12] |
|
| [13] |
Smith JJ, Amershi S, Barocas S, Wallach H, Wortman Vaughan J (2022) REAL ML: recognizing, exploring, and articulating limitations of machine learning research. In: ACM Int. Conf. Proceeding Ser. https://doi.org/10.1145/3531146.3533122 |
| [14] |
Malik MM (2020) A hierarchy of limitations in machine learning. https://arxiv.org/pdf/2002.05193. Accessed 18 May 2025 |
| [15] |
|
| [16] |
Thompson NC, Greenewald K, Lee K, Manso GF (2020) The computational limits of deep learning. arXiv preprint arXiv:2007.05558. https://doi.org/10.48550/arXiv.2007.05558 |
| [17] |
|
| [18] |
|
| [19] |
|
| [20] |
|
| [21] |
|
| [22] |
Dosilovic FK, Brcic M, Hlupic N (2018) Explainable artificial intelligence: a survey. In: 2018 41st Int. Conv. Inf Commun Technol Electron Microelectron. MIPRO 2018–Proc. https://doi.org/10.23919/MIPRO.2018.8400040 |
| [23] |
|
| [24] |
|
| [25] |
Dangeti P (2017) Statistics for machine learning. https://www.geeksforgeeks.org/machine-learning/statistics-for-machine-learning/. Accessed 11 Sep 2025 |
| [26] |
|
| [27] |
Sugiyama M (2015) Introduction to statistical machine learning. https://doi.org/10.1016/C2014-0-01992-2 |
| [28] |
Karapiperis K, Kochmann DM (2023) Prediction and control of fracture paths in disordered architected materials using graph neural networks. Commun Eng 2(1):32. https://doi.org/10.1038/S44172-023-00085-0 |
| [29] |
|
| [30] |
|
| [31] |
Freedman LP, Cypess RH (2019) Research reproducibility: the case for standards, handbook of biomarkers and precision medicine. pp 579–595. https://doi.org/10.1201/9780429202872-65 |
| [32] |
|
| [33] |
Naser MZ (2021) Demystifying ten big ideas and rules every fire scientist & engineer should know about blackbox, whitebox & causal artificial intelligence. https://arxiv.org/abs/2111.13756v1. Accessed 26 Jan 2022 |
| [34] |
|
| [35] |
|
| [36] |
|
| [37] |
|
| [38] |
|
| [39] |
|
| [40] |
Rebala G, Ravi A, Churiwala S (2019) Designing a machine learning system. An Introd to Mach Learn. https://doi.org/10.1007/978-3-030-15729-6_18 |
| [41] |
|
| [42] |
Marwala T (2021) Rational machines and artificial intelligence. https://doi.org/10.1016/B978-0-12-820676-8.09990-7 |
| [43] |
|
| [44] |
|
| [45] |
|
| [46] |
Udrescu SM, Tan A, Feng J, Neto O, Wu T, Tegmark M (2020) AI Feynman 2.0: Pareto-optimal symbolic regression exploiting graph modularity. Advances in Neural Information Processing Systems. 2020-December. https://github.com/SJ001/AI-Feynman. Accessed 31 Oct 2022 |
| [47] |
Wirth R (2000) CRISP-DM : towards a standard process model for data mining. Proceedings of the Fourth International Conference on the Practical Application of Knowledge Discovery and Data Mining |
| [48] |
Kolyshkina I, Simoff S (2021) The CRISP-ML approach to handling causality and interpretability issues in machine learning. In: Proc. - 2021 IEEE Int. Conf. Big Data, Big Data 2021. https://doi.org/10.1109/BigData52589.2021.9671754 |
| [49] |
|
| [50] |
|
| [51] |
|
| [52] |
Bang H, Selva D (2016) IFEED: interactive feature extraction for engineering design. Proc ASME Des Eng Tech Conf. https://doi.org/10.1115/DETC201660077 |
| [53] |
Liu H (2018) Feature engineering for machine learning and data analytics. https://doi.org/10.1201/9781315181080 |
| [54] |
Kolltveit AB, Li J (2022) Operationalizing machine learning models - a systematic literature review. In: Proc. - Work. Softw. Eng. Responsible AI, SE4RAI 2022. https://doi.org/10.1145/3526073.3527584 |
| [55] |
|
| [56] |
|
| [57] |
|
| [58] |
|
| [59] |
Tamburri DA (2020) Sustainable MLOps: trends and challenges. In: Proc. - 2020 22nd Int. Symp. Symb. Numer. Algorithms Sci. Comput. SYNASC 2020. https://doi.org/10.1109/SYNASC51798.2020.00015 |
| [60] |
|
| [61] |
Lisboa PJG (2013) Interpretability in machine learning – principles and practice, lecture notes in computer science (Including Subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics). 8256 LNAI. pp 15–21. https://doi.org/10.1007/978-3-319-03200-9_2 |
| [62] |
|
| [63] |
|
| [64] |
|
| [65] |
|
| [66] |
DeepLearing.ai, The State of Enterprise AI in 2020, DeepLearing.Ai. (2019) https://www.deeplearning.ai/the-batch/deployment-gap/. Accessed 28 Apr 2025 |
| [67] |
|
| [68] |
Crespo Márquez A (2022) The curse of dimensionality. In: Springer Ser. Reliab. Eng. https://doi.org/10.1007/978-3-030-97660-6_7 |
| [69] |
Lee JA, Verleysen M (eds) (2007) Nonlinear dimensionality reduction. Springer New York, NY |
| [70] |
Li J, Ge W, Wang W, Yang N, Liu X, Wang L, He X, Wang X, Wang J, Kwauk M (2013) From multiscale modeling to Meso-Science. https://doi.org/10.1007/978-3-642-35189-1 |
| [71] |
|
| [72] |
|
| [73] |
|
| [74] |
|
| [75] |
|
| [76] |
Snyder J, Menard A, Spare N (2023) Big data = Big questions for the engineering and construction industry. https://fmicorp.com/insights/industry-insights/big-data-big-questions-for-the-engineering-and-construction-industry. Accessed 11 Jul 2023 |
| [77] |
|
| [78] |
|
| [79] |
|
| [80] |
Vincent P, Larochelle H, Bengio Y, Manzagol PA (2008) Extracting and composing robust features with denoising autoencoders. In: Proc. 25th Int Conf Mach Learn. https://doi.org/10.1145/1390156.1390294 |
| [81] |
|
| [82] |
|
| [83] |
Zhang P (2010) Advanced industrial control technology. https://doi.org/10.1016/C2009-0-20337-0 |
| [84] |
Stuckner J, Piekenbrock M, Arnold SM, Ricks TM (2021) Optimal experimental design with fast neural network surrogate models. Comput Mater Sci. https://doi.org/10.1016/j.commatsci.2021.110747 |
| [85] |
|
| [86] |
Zhao H, Gordon GJ (2022) Inherent tradeoffs in learning fair representations. J Mach Learn Res. 23(57):1–26. arXiv:1906.08386. https://doi.org/10.48550/arXiv.1906.08386 |
| [87] |
|
| [88] |
Simpia LJ (2024) Machine learning adoption: the interplay of cost, time, and systems integration. http://www.theseus.fi/handle/10024/858584. Accessed 19 May 2025 |
| [89] |
|
| [90] |
|
| [91] |
|
| [92] |
|
| [93] |
Jain AK, Chandrasekaran B (1982) 39 dimensionality and sample size considerations in pattern recognition practice. Handbook of Statistics. 2:835–855. https://doi.org/10.1016/S0169-7161(82)02042-2 |
| [94] |
|
| [95] |
|
| [96] |
Widrow B (1987) Adaline and madaline. In: IEEE 1st Int. Conf. Neural Networks. https://scholar.google.com/scholar?hl=en&as_sdt=0%2C23&q=Widrow%2C+B.+1987.+ADALINE+and+MADALINE+-+1963%2C+Plenary+Speech%2C+Vol+I%3A+Proc.+IEEE+Zst+Int.+Conf.+on+Neural+Networks%2C+San+Diego%2C+CA%2C+143-158.+&btnG=. Accessed 18 May 2025 |
| [97] |
|
| [98] |
|
| [99] |
van Smeden M, Moons KG, de Groot JA, Collins GS, Altman DG, Eijkemans MJ, Reitsma JB (2018) Sample size for binary logistic prediction models: beyond events per variable criteria. 28:2455–2474. https://doi.org/10.1177/0962280218784726 |
| [100] |
|
| [101] |
Frank I, Todeschini R (1994) The data analysis handbook. https://books.google.com/books?hl=en&lr=&id=SXEpB0H6L3YC&oi=fnd&pg=PP1&ots=zfmIRO_XO5&sig=dSX6KJdkuav5zRNxaUdcftGSn2k. Accessed 21 Jun 2019 |
| [102] |
|
| [103] |
Prasanna S, Kashinkunti P, Milletari F (2017) TensorRT 3: faster TensorFlow inference and Volta support. https://developer.nvidia.com/blog/tensorrt-3-faster-tensorflow-inference/. Accessed 4 Dec 2017 |
| [104] |
|
| [105] |
|
| [106] |
Lones MA (2021) How to avoid machine learning pitfalls: a guide for academic researchers. http://www.macs.hw.ac.uk/. Accessed 10 Dec 2021 |
| [107] |
Shi Y, Horiguchi M, Lu Y (2024) Suggestions on using machine learning models and cautions for analyzing censored time-to-event outcomes. JCO Precis Oncol. https://doi.org/10.1200/PO.24.00220/ASSET/5179A6CB-F38C-4794-A62B-794AB668439C/ASSETS/PO.24.00220.FP.PNG |
| [108] |
|
| [109] |
Sculley D, Holt G, Golovin D, Davydov E, Phillips T, Ebner D, Chaudhary V, Young M, Crespo J-FF, Dennison D (2015) Hidden technical debt in machine learning systems. In: Proceedings of the 29th International Conference on Neural Information Processing Systems. MIT Press, Cambridge, MA, 2:2503–2511 |
| [110] |
|
| [111] |
Ribeiro MT, Singh S, Guestrin C (2016) “Why should i trust you?” Explaining the predictions of any classifier. In: Proc. ACM SIGKDD Int Conf Knowl Discov Data Min. https://doi.org/10.1145/2939672.2939778 |
| [112] |
Lundberg SM, Lee SI (2017) A unified approach to interpreting model predictions. In: Proceedings of the 31st International Conference on Neural Information Processing Systems. Curran Associates Inc, NY, United States, pp. 4768–4777 |
| [113] |
|
| [114] |
|
| [115] |
|
| [116] |
|
| [117] |
|
| [118] |
Castano J, Martinez-Fernandez S, Franch X, Bogner J (2024) Analyzing the evolution and maintenance of ML models on hugging face. In: Proceedings - 2024 IEEE/ACM 21st International Conference on Mining Software Repositories, MSR 2024. 12:607–618. https://doi.org/10.1145/3643991.3644898 |
| [119] |
Kang D, Raghavan D, Bailis P, Zaharia M (2020) Model assertions for monitoring and improving ML models. In: Proceedings of Machine Learning and Systems 2:481–496. https://doi.org/10.48550/arXiv.2003.01668 |
| [120] |
|
| [121] |
|
| [122] |
Zhou Y, Chen S, Wang Y, Huan W (2020) Review of research on lightweight convolutional neural networks. In: Proc. 2020 IEEE 5th Inf Technol Mechatronics Eng Conf. ITOEC 2020. https://doi.org/10.1109/ITOEC49072.2020.9141847 |
| [123] |
Reuther A, Michaleas P, Jones M, Gadepally V, Samsi S, Kepner J (2020) Survey of machine learning accelerators. In: 2020 IEEE High Perform. Extrem Comput Conf HPEC 2020. https://doi.org/10.1109/HPEC43674.2020.9286149 |
| [124] |
Molnar C (2022) Interpretable machine learning: a guide for making black box models explainable. Independently published, 28 Feb 2022 |
| [125] |
Naser MZ (2023) Machine learning for all! Benchmarking automated, explainable, and coding-free platforms on civil and environmental engineering problems. J Infrastructure Intel Resilience. https://doi.org/10.1016/j.iintel.2023.100028 |
| [126] |
Zhou ZH (2021) Machine learning. Springer Singapore. https://doi.org/10.1007/978-981-15-1967-3 |
| [127] |
Spirtes P, Glymour C, Scheines R (1993) Causation, prediction, and search. In: Lecture Notes in Statistics. Springer, New York, 81. https://doi.org/10.1007/978-1-4612-2748-9 |
| [128] |
|
The Author(s)
/
| 〈 |
|
〉 |