TY - GEN
T1 - In-depth Analysis of Low-rank Matrix Factorisation in a Federated Setting
AU - Philippenko, Constantin
AU - Scaman, Kevin
AU - Massoulié, Laurent
N1 - Publisher Copyright:
Copyright © 2025, Association for the Advancement of Artificia Intelligence (www.aaai.org). All rights reserved.
PY - 2025/4/11
Y1 - 2025/4/11
N2 - We analyze a distributed algorithm to compute a low-rank matrix factorization on N clients, each holding a local dataset Si ∈ Rni×d, mathematically, we seek to solve minUi∊Rni×rV∊Rd×r12PNi=1 ∥Si−UiVT∥2F. Considering a power initialization of V, we rewrite the previous smooth non-convex problem into a smooth strongly-convex problem that we solve using a parallel Nesterov gradient descent potentially requiring a single step of communication at the initialization step. For any client i in {1, . . ., N}, we obtain a global V in Rd×r common to all clients and a local variable Ui in Rni×r. We provide a linear rate of convergence of the excess loss which depends on σmax/σr, where σr is the rth singular value of the concatenation S of the matrices (Si)Ni=1. This result improves the rates of convergence given in the literature, which depend on σmax2 /σmin2 . We provide an upper bound on the Frobenius-norm error of reconstruction under the power initialization strategy. We complete our analysis with experiments on both synthetic and real data.
AB - We analyze a distributed algorithm to compute a low-rank matrix factorization on N clients, each holding a local dataset Si ∈ Rni×d, mathematically, we seek to solve minUi∊Rni×rV∊Rd×r12PNi=1 ∥Si−UiVT∥2F. Considering a power initialization of V, we rewrite the previous smooth non-convex problem into a smooth strongly-convex problem that we solve using a parallel Nesterov gradient descent potentially requiring a single step of communication at the initialization step. For any client i in {1, . . ., N}, we obtain a global V in Rd×r common to all clients and a local variable Ui in Rni×r. We provide a linear rate of convergence of the excess loss which depends on σmax/σr, where σr is the rth singular value of the concatenation S of the matrices (Si)Ni=1. This result improves the rates of convergence given in the literature, which depend on σmax2 /σmin2 . We provide an upper bound on the Frobenius-norm error of reconstruction under the power initialization strategy. We complete our analysis with experiments on both synthetic and real data.
UR - https://www.scopus.com/pages/publications/105004286628
U2 - 10.1609/aaai.v39i19.34192
DO - 10.1609/aaai.v39i19.34192
M3 - Conference contribution
AN - SCOPUS:105004286628
T3 - Proceedings of the AAAI Conference on Artificial Intelligence
SP - 19904
EP - 19912
BT - Special Track on AI Alignment
A2 - Walsh, Toby
A2 - Shah, Julie
A2 - Kolter, Zico
PB - Association for the Advancement of Artificial Intelligence
T2 - 39th Annual AAAI Conference on Artificial Intelligence, AAAI 2025
Y2 - 25 February 2025 through 4 March 2025
ER -