Research Interests

My general research interest is in statistical signal processing, machine learning, and nonlinear optimization algorithms. My research topics include:

  • Statistical learning over networks;

  • Large-scale optimization algorithms;

  • Majorization minimization algorithms;

  • Robust covariance matrix estimation;

  • Sparse principal component analysis.

Distributed Learning in Heterogenous Environments

The era of pervasive intelligence features a proliferation of smart devices that continuously sense, learn from, and react to dynamic environments. As the number of smart devices grows, tremendous amounts of data are generated in a distributed fasion and must be processed efficiently to support real-time learning and decision-making. Designing efficient, reliable and scalable data processing algorithms is therefore of paramount importance for engineering applications nowadays. Achieving the goal faces challenges due to the heterogeneity arising from data, computation, and communication resources.

Distributed learning under data heterogeneity

TransFusion: covariate-shift robust transfer learning for high-dimensional regression
Z. He, Y. Sun, J. Liu, and R. Li (AISTATS2024)

Distributed Stochastic bilevel optimization: improved complexity and heterogeneity analysis
Y. Niu, J. Xu, Y. Sun, Y. Huang, and L. Chai (JMLR2025)

Communication, computation, and network structure design

EFSkip: a new error feedback with linear speedup for compressed federated learning with arbitrary data heterogeneity
H. Bao, P. Chen, Y. Sun, and Z. Li (AAAI2025)

The effectiveness of local updates for decentralized learning under data heterogeneity
T. Wu, Z. Li, and Y. Sun (TSP2025)

Tackling data heterogeneity: a new unified framework for decentralized SGD with sample-induced topology
Y. Huang, Y. Sun, Z. Zhu, C. Yan, and J. Xu (ICML2022)

Hybrid local SGD for federated learning with heterogeneous communications
Y. Guo, Y. Sun, R. Hui, and Y. Gong (ICLR2021)

Distributed learning for structured data

Non-convex tensor recovery from local measurements
T. Wu, Y. Sun, and J. Fan (AAAI2025)

Covariance selection over networks
W. Xia, F. Li, Y. Sun, Z. Zhao (AISTATS2025)

Implicit regularization of decentralized gradient descent for sparse regression
T. Wu and Y. Sun (NeurIPS2024)

Distributed sparse regression via penalization
Y. Ji, G. Scutari*, Y. Sun*, and H. Honnappa (JMLR2023) (*equal contribution)

Distributed (ATC) gradient descent for high dimension sparse regression
Y. Ji, G. Scutari, Y. Sun, and H. Honnappa (TIT2023)

Decentralized dictionary learning over time-varying digraphs
A. Daneshmand, Y. Sun, G. Scutari, F. Facchinei, and Brian M. Sadler (JMLR2019)

Efficient Algorithm Design for Statistical Signal Processing

Statistics and optimization demonstrate a close interplay in data analytics. Sophisticated statistical models that produce high quality solutions often lead to complex highly nonconvex optimization problems. However, traditional optimization tools applied to these problems in theory only yield local solutions. Moreover, employing a black-box algorithm can be inefficient due to the ignorance of the problem structure and computational resources at hand. We are interested in developing problem-driven low complexity algorithms for statistical learning with provable guarantees.

Majorization-minimization algorithms

Majorization-minimization algorithms in signal processing, communications, and machine learning
Y. Sun, P. Babu, and D. P. Palomar (TSP2017)

Structured covariance estimation

Low-complexity algorithms for low rank clutter parameters estimation in radar systems
Y. Sun, A. Breloy, P. Babu, D. P. Palomar, F. Pascal, and G. Ginolhac (TSP 2016)

Robust estimation of structured covariance matrix for heavy-tailed elliptical distributions
Y. Sun, P. Babu, and D. P. Palomar (TSP2016)

Regularized robust estimation of mean and covariance matrix under heavy-tailed distributions
Y. Sun, P. Babu, and D. P. Palomar (TSP2015)

Regularized Tyler's scatter estimator: existence, uniqueness, and algorithms
Y. Sun, P. Babu, and D. P. Palomar (TSP2014)

Sparse principal component analysis

Orthogonal Sparse PCA and Covariance Estimation via Procrustes Reformulation
K. Benidis, Y. Sun, P. Babu, D. P. Palomar (TSP2016)

Large-Scale Optimization Algorithms

In the era of "big data", we are witnessing a fast development in data acquisition techniques. New data features such as the massive volume/dimension, heterogeneous structure, and decentralized storage challenge traditional optimization methods, most of which rely on centralized information and computation. We are interested in developing parallel and decentralized algorithms capable of solving large-scale optimization problems leveraging multiple computing units, equipped with the following desirable features:

  • fast and efficient computation

  • flexible to problem type and network architecture

  • robust against delay and asynchrony.

Flexible distributed successive convex approximation

Distributed optimization based on gradient-tracking revisited: enhancing convergence rate via surrogation
Y. Sun, A. Daneshmand, and G. Scutari, (SIOPT2022).

Distributed big-data optimization via block-wise gradient tracking
I. Notarnicola*, Y. Sun*, G. Scutari, G. Notarstefano (TAC2020) (*equal contribution)

Distributed nonconvex constrained optimization over time-varying digraphs
G. Scutari and Y. Sun (Alphabetical order) (Math. Prog. Series B, 2018)

Asynchronous distributed optimization

Achieving linear convergence in distributed asynchronous multi-agent optimization
Y. Tian, Y. Sun, and G. Scutari (TAC2020)