Enhancing job salary prediction with disentangled composition effect modeling: a neural prototyping approach

Yang JI , Ying SUN , Hengshu ZHU

Front. Comput. Sci. ›› 2026, Vol. 20 ›› Issue (5) : 2005345

PDF (3582KB)
Front. Comput. Sci. ›› 2026, Vol. 20 ›› Issue (5) : 2005345 DOI: 10.1007/s11704-025-50421-0
Artificial Intelligence
RESEARCH ARTICLE

Enhancing job salary prediction with disentangled composition effect modeling: a neural prototyping approach

Author information +
History +
PDF (3582KB)

Abstract

In the era of the knowledge economy, understanding how job skills influence salary is crucial for promoting recruitment with competitive salary systems and aligned salary expectations. Despite efforts on salary prediction based on job positions and talent demographics, there still lacks methods to effectively discern the set-structured skills’ intricate composition effect on job salary. While recent advances in neural networks have significantly improved accurate set-based quantitative modeling, their lack of explainability hinders obtaining insights into the skills’ composition effects. Indeed, model explanation for set data is challenging due to the combinatorial nature, rich semantics, and unique format. To this end, in this paper, we propose a novel intrinsically explainable set-based neural prototyping approach, namely LGDESetNet, for explainable salary prediction that can reveal disentangled skill sets that impact salary from both local and global perspectives. Specifically, we propose a skill graph-enhanced disentangled discrete subset selection layer to identify multi-faceted influential input subsets with varied semantics. Furthermore, we propose a set-oriented prototype learning method to extract globally influential prototypical sets. The resulting output is transparently derived from the semantic interplay between these input subsets and global prototypes. Extensive experiments on four real-world datasets demonstrate that our method achieves superior performance than state-of-the-art baselines in salary prediction while providing explainable insights into salary-influencing patterns.

Graphical abstract

Keywords

data mining / job salary prediction / set-based modeling / explainable machine learning

Cite this article

Download citation ▾
Yang JI, Ying SUN, Hengshu ZHU. Enhancing job salary prediction with disentangled composition effect modeling: a neural prototyping approach. Front. Comput. Sci., 2026, 20(5): 2005345 DOI:10.1007/s11704-025-50421-0

登录浏览全文

4963

注册一个新账户 忘记密码

References

[1]

Ng T W H, Feldman D C. A conservation of resources perspective on career hurdles and salary attainment. Journal of Vocational Behavior, 2014, 85( 1): 156–168

[2]

Dix-Carneiro R, Kovak B K. Trade liberalization and the skill premium: a local labor markets approach. American Economic Review, 2015, 105( 5): 551–557

[3]

Burstein A, Vogel J. International trade, technology, and the skill premium. Journal of Political Economy, 2017, 125( 5): 1356–1412

[4]

Sun Y, Zhuang F, Zhu H, Zhang Q, He Q, Xiong H. Market-oriented job skill valuation with cooperative composition neural network. Nature Communications, 2021, 12( 1): 1992

[5]

Ternikov A. Soft and hard skills identification: insights from IT job advertisements in the CIS region. PeerJ Computer Science, 2022, 8: e946

[6]

Lovaglio P G, Cesarini M, Mercorio F, Mezzanzanica M. Skills in demand for ICT and statistical occupations: evidence from web-based job vacancies. Statistical Analysis and Data Mining: The ASA Data Science Journal, 2018, 11( 2): 78–91

[7]

Meng Q, Zhu H, Xiao K, Xiong H. Intelligent salary benchmarking for talent recruitment: a holistic matrix factorization approach. In: Proceedings of 2018 IEEE International Conference on Data Mining (ICDM). 2018, 337−346

[8]

Meng Q, Xiao K, Shen D, Zhu H, Xiong H. Fine-grained job salary benchmarking with a nonparametric Dirichlet process−based latent factor model. INFORMS Journal on Computing, 2022, 34( 5): 2443–2463

[9]

Lazar A. Income prediction via support vector machine. In: Proceedings of 2004 International Conference on Machine Learning and Applications. 2004, 143−149

[10]

Zaheer M, Kottur S, Ravanbhakhsh S, Póczos B, Salakhutdinov R, Smola A J. Deep sets. In: Proceedings of the 31st International Conference on Neural Information Processing Systems. 2017, 3394−3404

[11]

Lee J, Lee Y, Kim J, Kosiorek A R, Choi S, Teh Y W. Set transformer: a framework for attention-based permutation-invariant neural networks. In: Proceedings of the 36th International Conference on Machine Learning. 2019, 3744−3753

[12]

Sundararajan M, Taly A, Yan Q. Axiomatic attribution for deep networks. In: Proceedings of the 34th International Conference on Machine Learning. 2017, 3319−3328

[13]

Nauta M, van Bree R, Seifert C. Neural prototype trees for interpretable fine-grained image recognition. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. 2021, 14928−14938

[14]

Chen C, Li O, Tao C, Barnett A J, Su J, Rudin C. This looks like that: deep learning for interpretable image recognition. In: Proceedings of the 33rd International Conference on Neural Information Processing Systems. 2019, 801

[15]

Li O, Liu H, Chen C, Rudin C. Deep learning for case-based reasoning through prototypes: a neural network that explains its predictions. In: Proceedings of the 32nd AAAI Conference on Artificial Intelligence. 2018, 3530−3537

[16]

Quan T Z, Raheem M. Salary prediction in data science field using specialized skills and job benefits−a literature review. Journal of Applied Technology and Innovation, 2022, 6( 3): 70–74

[17]

Autor D H, Dorn D. The growth of low-skill service jobs and the polarization of the US labor market. American Economic Review, 2013, 103( 5): 1553–1597

[18]

Sun Y, Ji Y, Zhu H, Zhuang F, He Q, Xiong H. Market-aware long-term job skill recommendation with explainable deep reinforcement learning. ACM Transactions on Information Systems, 2025, 43( 2): 46

[19]

Fang C, Qin C, Zhang Q, Yao K, Zhang J, Zhu H, Zhuang F, Xiong H. RecruitPro: a pretrained language model with skill-aware prompt learning for intelligent recruitment. In: Proceedings of the 29th ACM SIGKDD Conference on Knowledge Discovery and Data Mining. 2023, 3991−4002

[20]

Zhang L, Zhang W, Wu L, Zhao H. GCTN: graph competitive transfer network for cross-domain multi-behavior prediction. IEEE Transactions on Knowledge and Data Engineering, 2025, 37( 7): 4075–4088

[21]

Hang J, Dong Z, Zhao H, Song X, Wang P, Zhu H. Outside in: market-aware heterogeneous graph neural network for employee turnover prediction. In: Proceedings of the 15th ACM International Conference on Web Search and Data Mining. 2022, 353−362

[22]

Wu L, Li Z, Zhao H, Liu Q, Wang J, Zhang M, Chen E. Learning the implicit semantic representation on graph-structured data. In: Proceedings of 26th International Conference on Database Systems for Advanced Applications. 2021, 3−19

[23]

Li J, Gan W, Gui Y, Wu Y, Yu P S. Frequent itemset mining with local differential privacy. In: Proceedings of the 31st ACM International Conference on Information & Knowledge Management. 2022, 1146−1155

[24]

Murphy R L, Srinivasan B, Rao V, Ribeiro B. Janossy pooling: learning deep permutation-invariant functions for variable-size inputs. In: Proceedings of the 7th International Conference on Learning Representations (ICLR). 2019

[25]

Maron H, Litany O, Chechik G, Fetaya E. On learning sets of symmetric elements. In: Proceedings of the 37th International Conference on Machine Learning. 2020, 6734−6744

[26]

Hirsch R, Gilad-Bachrach R. Trees with attention for set prediction tasks. In: Proceedings of the 38th International Conference on Machine Learning. 2021, 4250−4261

[27]

Zhang L, Tozzo V, Higgins J, Ranganath R. Set norm and equivariant skip connections: putting the deep in deep sets. In: Proceedings of the 39th International Conference on Machine Learning. 2022, 26559−26574

[28]

Charles R Q, Su H, Kaichun M, Guibas L J. PointNet: deep learning on point sets for 3D classification and segmentation. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. 2017, 77−85

[29]

Xu Y, Fan T, Xu M, Zeng L, Qiao Y. SpiderCNN: deep learning on point sets with parameterized convolutional filters. In: Proceedings of the 15th European Conference on Computer Vision. 2018, 90−105

[30]

Wu X, Jiang L, Wang P S, Liu Z, Liu X, Qiao Y, Ouyang W, He T, Zhao H. Point transformer V3: simpler, faster, stronger. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. 2024, 4840−4851

[31]

Chen G, Wang M, Yang Y, Yu K, Yuan L, Yue Y. PointGPT: auto-regressively generative pre-training from point clouds. In: Proceedings of 37th Conference on Neural Information Processing Systems. 2023, 1291

[32]

Maron H, Ben-Hamu H, Shamir N, Lipman Y. Invariant and equivariant graph networks. In: Proceedings of International Conference on Learning Representations. 2019

[33]

Kim J, Nguyen T D, Min S, Cho S, Lee M, Lee H, Hong S. Pure transformers are powerful graph learners. In: Proceedings of the 36th International Conference on Neural Information Processing Systems. 2022, 1060

[34]

Chien E, Pan C, Peng J, Milenkovic O. You are AllSet: a multiset function framework for hypergraph neural networks. In: Proceedings of the 10h International Conference on Learning Representations. 2022

[35]

Yao H, Hanslovsky P, Huetter J C, Hoeckendorf B, Richmond D. Weakly supervised set-consistency learning improves morphological profiling of single-cell images. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. 2024, 6978−6987

[36]

Gordon J, Bruinsma W P, Foong A Y K, Requeima J, Dubois Y, Turner R E. Convolutional conditional neural processes. In: Proceedings of International Conference on Learning Representations. 2020

[37]

Selvaraju R R, Cogswell M, Das A, Vedantam R, Parikh D, Batra D. Grad-CAM: visual explanations from deep networks via gradient-based localization. In: Proceedings of the IEEE International Conference on Computer Vision. 2017, 618−626

[38]

Ancona M, Ceolini E, Öztireli C, Gross M. Towards better understanding of gradient-based attribution methods for Deep Neural Networks. In: Proceedings of the 6th International Conference on Learning Representations. 2018

[39]

Lundberg S M, Lee S I. A unified approach to interpreting model predictions. In: Proceedings of the 31st International Conference on Neural Information Processing Systems. 2017, 4768−4777

[40]

Rudin C. Stop explaining black box machine learning models for high stakes decisions and use interpretable models instead. Nature Machine Intelligence, 2019, 1( 5): 206–215

[41]

Schröder M, Zamanian A, Ahmidi N. Post-hoc saliency methods fail to capture latent feature importance in time series data. In: Proceedings of the 1st International Workshop on Trustworthy Machine Learning for Healthcare. 2023, 106−121

[42]

Ji Y, Sun Y, Zhang Y, Wang Z, Zhuang Y, Gong Z, Shen D, Qin C, Zhu H, Xiong H. A comprehensive survey on self-interpretable neural networks. 2025, arXiv preprint arXiv: 2501.15638

[43]

Lemhadri I, Ruan F, Tibshirani R. LassoNet: neural networks with feature sparsity. In: Proceedings of the 24th International Conference on Artificial Intelligence and Statistics. 2021, 10−18

[44]

Miao S, Liu M, Li P. Interpretable and generalizable graph learning via stochastic attention mechanism. In: Proceedings of the 39th International Conference on Machine Learning. 2022, 15524−15543

[45]

Kolodner J L. An introduction to case-based reasoning. Artificial Intelligence Review, 1992, 6( 1): 3–34

[46]

van Gog T, Rummel N. Example-based learning: integrating cognitive and social-cognitive research perspectives. Educational Psychology Review, 2010, 22( 2): 155–174

[47]

Ma C, Zhao B, Chen C, Rudin C. This looks like those: illuminating prototypical concepts using multiple visualizations. In: Proceedings of the 37th International Conference on Neural Information Processing Systems. 2023, 1704

[48]

Wang J, Liu H, Wang X, Jing L. Interpretable image recognition by constructing transparent embedding space. In: Proceedings of the IEEE/CVF International Conference on Computer Vision. 2021, 875−884

[49]

Rymarczyk D, Struski Ł Tabor J, Zieliński B. ProtoPShare: prototypical parts sharing for similarity discovery in interpretable image classification. In: Proceedings of the 27th ACM SIGKDD Conference on Knowledge Discovery & Data Mining. 2021, 1420−1430

[50]

Pach M, Lewandowska K, Tabor J, Zieliński B M, Rymarczyk D D. LucidPPN: unambiguous prototypical parts network for user-centric interpretable computer vision. In: Proceedings of the 13th International Conference on Learning Representations. 2025

[51]

Ming Y, Xu P, Qu H, Ren L. Interpretable and steerable sequence learning via prototypes. In: Proceedings of the 25th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining. 2019, 903−913

[52]

Hong D, Wang T, Baek S. ProtoryNet - Interpretable text classification via prototype trajectories. Journal of Machine Learning Research, 2023, 24( 264): 1–39

[53]

Rajagopal D, Balachandran V, Hovy E H, Tsvetkov Y. SELFEXPLAIN: a self-explaining architecture for neural text classifiers. In: Proceedings of 2021 Conference on Empirical Methods in Natural Language Processing. 2021, 836−850

[54]

Zhang Z, Liu Q, Wang H, Lu C, Lee C. ProtGNN: towards self-explaining graph neural networks. In: Proceedings of the 36th AAAI Conference on Artificial Intelligence. 2022, 9127−9135

[55]

Seo S, Kim S, Park C. Interpretable prototype-based graph information bottleneck. In: Proceedings of the 37th International Conference on Neural Information Processing Systems. 2023, 3353

[56]

Jiang Y, Yu W, Song D, Wang L, Cheng W, Chen H. FedSkill: privacy preserved interpretable skill learning via imitation. In: Proceedings of the 29th ACM SIGKDD Conference on Knowledge Discovery and Data Mining. 2023, 1010−1019

[57]

Fauvel K, Chen F, Rossi D. A lightweight, efficient and explainable-by-design convolutional neural network for internet traffic classification. In: Proceedings of the 29th ACM SIGKDD Conference on Knowledge Discovery and Data Mining. 2023, 4013−4023

[58]

Kenny E M, Tucker M, Shah J. Towards interpretable deep reinforcement learning with human-friendly prototypes. In: Proceedings of the 11th International Conference on Learning Representations. 2023

[59]

Vaswani A, Shazeer N, Parmar N, Uszkoreit J, Jones L, Gomez A N, Kaiser Ł Polosukhin I. Attention is all you need. In: Proceedings of the 31st International Conference on Neural Information Processing Systems. 2017, 6000−6010

[60]

Gumbel E J. Statistical Theory of Extreme Values and Some Practical Applications: A Series of Lectures. Washington: U.S. Government Printing Office, 1954

[61]

Geng X, Wang L, Wang X, Qin B, Liu T, Tu Z. How does selective mechanism improve self-attention networks? In: Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics. 2020, 2986−2995

[62]

Gou Q, Dong Y, Wu Y, Ke Q. Semantic similarity-based program retrieval: a multi-relational graph perspective. Frontiers of Computer Science, 2024, 18( 3): 183209

[63]

Chen J, Wu J, Chen J, Xin X, Li Y, He X. How graph convolutions amplify popularity bias for recommendation?. Frontiers of Computer Science, 2024, 18( 5): 185603

[64]

Khuller S, Saha B. On finding dense subgraphs. In: Proceedings of 36th International Colloquium on Automata, Languages and Programming. 2009, 597−608

[65]

Cheng H T, Koc L, Harmsen J, Shaked T, Chandra T, Aradhye H, Anderson G, Corrado G, Chai W, Ispir M, Anil R, Haque Z, Hong L, Jain V, Liu X, Shah H. Wide & deep learning for recommender systems. In: Proceedings of the 1st Workshop on Deep Learning for Recommender Systems. 2016, 7−10

[66]

Guo H, Tang R, Ye Y, Li Z, He X. DeepFM: a factorization-machine based neural network for CTR prediction. In: Proceedings of the 26th International Joint Conference on Artificial Intelligence. 2017, 1725−1731

[67]

Devlin J, Chang M W, Lee K, Toutanova K. BERT: pre-training of deep bidirectional transformers for language understanding. In: Proceedings of 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers). 2019, 4171−4186

[68]

Kingma D P, Ba J. Adam: a method for stochastic optimization. In: Proceedings of the 3rd International Conference on Learning Representations. 2015

[69]

Rong Y, Leemann T, Nguyen T T, Fiedler L, Qian P, Unhelkar V, Seidel T, Kasneci G, Kasneci E. Towards human-centered explainable AI: a survey of user studies for model explanations. IEEE Transactions on Pattern Analysis and Machine Intelligence, 2024, 46( 4): 2104–2122

RIGHTS & PERMISSIONS

Higher Education Press

AI Summary AI Mindmap
PDF (3582KB)

Supplementary files

Highlights

448

Accesses

0

Citation

Detail

Sections
Recommended

AI思维导图

/