Federated Sufficient Dimension Reduction Through High-Dimensional Sparse Sliced Inverse Regression
Wenquan Cui , Yue Zhao , Jianjun Xu , Haoyang Cheng
Communications in Mathematics and Statistics ›› 2023, Vol. 13 ›› Issue (3) : 719 -756.
Federated Sufficient Dimension Reduction Through High-Dimensional Sparse Sliced Inverse Regression
Federated learning has become a popular tool in the big data era nowadays. It trains a centralized model based on data from different clients while keeping data decentralized. In this paper, we propose a federated sparse sliced inverse regression algorithm for the first time. Our method can simultaneously estimate the central dimension reduction subspace and perform variable selection in a federated setting. We transform this federated high-dimensional sparse sliced inverse regression problem into a convex optimization problem by constructing the covariance matrix safely and losslessly. We then use a linearized alternating direction method of multipliers algorithm to estimate the central subspace. We also give approaches of Bayesian information criterion and holdout validation to ascertain the dimension of the central subspace and the hyper-parameter of the algorithm. We establish an upper bound of the statistical error rate of our estimator under the heterogeneous setting. We demonstrate the effectiveness of our method through simulations and real world applications.
Federated learning / Sliced inverse regression / Sufficient dimension reduction / Variable selection / Mathematical Sciences / Statistics / Information and Computing Sciences / Artificial Intelligence and Image Processing
| [1] |
Absil, P.-A., Mahony, R., Sepulchre, R.: Optimization Algorithms on Matrix Manifolds. Princeton University Press, Princeton (2009). https://doi.org/10.1515/9781400830244 |
| [2] |
Agrawal, S., Jia, R.: Optimistic posterior sampling for reinforcement learning: worst-case regret bounds. In: Advances in Neural Information Processing Systems, vol. 30, pp. 1184–1194. Curran Associates, Inc. (2017) |
| [3] |
Boyd, S., Parikh, N., Chu, E., Peleato, B., Eckstein, J., et al.: Distributed optimization and statistical learning via the alternating direction method of multipliers. Found. Trends® Mach. Learn. 3(1), 1–122 (2011). https://doi.org/10.1561/2200000016 |
| [4] |
|
| [5] |
Caldas, S., Duddu, S.M.K., Wu, P., Li, T., Konečný, J., McMahan, H.B., Smith, V., Talwalkar, A.: Leaf: A benchmark for federated settings. arXiv preprint arXiv:1812.01097 (2019) |
| [6] |
Chai, D., Wang, L., Fu, L., Zhang, J., Chen, K., Yang, Q.: Federated singular vector decomposition. arXiv preprint arXiv:2105.08925 (2021) |
| [7] |
|
| [8] |
Cheng, H., Cui, W., Jianjun, X.: Online sparse sliced inverse regression. arXiv preprint arXiv:2009.14615 (2021) |
| [9] |
|
| [10] |
|
| [11] |
|
| [12] |
|
| [13] |
|
| [14] |
|
| [15] |
|
| [16] |
|
| [17] |
Hsieh, K., Phanishayee, A., Mutlu, O., Gibbons, P.: The non-iid data quagmire of decentralized machine learning. In: International Conference on Machine Learning, pp. 4387–4398 (2020). PMLR |
| [18] |
|
| [19] |
Kairouz, P., McMahan, H.B., Avent, B., Bellet, A., Bennis, M., Bhagoji, A.N., Bonawitz, K., Charles, Z., Cormode, G., Cummings, R., et al.: Advances and open problems in federated learning. Found. Trends® in Mach. Learn. 14(1–2), 1–210 (2021). https://doi.org/10.1561/2200000083 |
| [20] |
Karimireddy, S.P., Kale, S., Mohri, M., Reddi, S., Stich, S., Suresh, A.T.: Scaffold: Stochastic controlled averaging for federated learning. In: International Conference on Machine Learning, pp. 5132–5143 (2020). PMLR |
| [21] |
|
| [22] |
Li, B.: Sufficient Dimension Reduction: Methods and Applications with R. Chapman and Hall/CRC Press, Boca Raton (2017). https://doi.org/10.1201/9781315119427 |
| [23] |
|
| [24] |
|
| [25] |
Li, L., Dennis Cook, R., Nachtsheim, C.J.: Model-free variable selection. J. Roy. Stat. Soc. B 67(2), 285–299 (2005). https://doi.org/10.1111/j.1467-9868.2005.00502.x |
| [26] |
|
| [27] |
|
| [28] |
|
| [29] |
McMahan, B., Moore, E., Ramage, D., Hampson, S., y Arcas, B.A.: Communication-efficient learning of deep networks from decentralized data. In: Artificial Intelligence and Statistics, pp. 1273–1282 (2017). PMLR |
| [30] |
|
| [31] |
|
| [32] |
Redmond, M.: Communities and Crime. UCI Machine Learning Repository (2009) |
| [33] |
|
| [34] |
|
| [35] |
Vershynin, R.: High-Dimensional Probability: An Introduction with Applications in Data Science. Cambridge Series in Statistical and Probabilistic Mathematics. Cambridge University Press, Cambridge (2018). https://doi.org/10.1017/9781108231596 |
| [36] |
Vershynin, R.: Introduction to the non-asymptotic analysis of random matrices. arXiv preprint arXiv:1011.3027 (2010). https://doi.org/10.1017/CBO9780511794308.006 |
| [37] |
Vu, V.Q., Cho, J., Lei, J., Rohe, K.: Fantope projection and selection: A near-optimal convex relaxation of sparse pca. Adv. Neural Inf. Process. Syst. 26, vol. 26, pp. 2670–2678 (2013) |
| [38] |
Xia, Y., Tong, H., Li, W.K., Zhu, L.-X.: An adaptive estimation of dimension reduction space. J. Roy. Stat. Soc. B 64(3), 363–410 (2002). https://doi.org/10.1142/9789812836281_0023 |
| [39] |
|
| [40] |
|
| [41] |
Yeh, I.-C., Lien, C.-h: The comparisons of data mining techniques for the predictive accuracy of probability of default of credit card clients. Expert Syst. Appl. 36(2, Part 1), 2473–2480 (2009). https://doi.org/10.1016/j.eswa.2007.12.020 |
| [42] |
|
| [43] |
|
| [44] |
|
| [45] |
|
| [46] |
|
| [47] |
|
| [48] |
|
School of Mathematical Sciences, University of Science and Technology of China and Springer-Verlag GmbH Germany, part of Springer Nature
/
| 〈 |
|
〉 |