Large Covariance Matrix Estimation by Composite Minimization
The present thesis concerns large covariance matrix estimation via composite minimization under the assumption of low rank plus sparse structure. Existing methods like POET (Principal Orthogonal complEment Thresholding) perform estimation by extracting principal components and then applying a soft...
Main Author: | |
---|---|
Other Authors: | |
Format: | Doctoral Thesis |
Language: | en |
Published: |
Alma Mater Studiorum - Università di Bologna
2016
|
Subjects: | |
Online Access: | http://amsdottorato.unibo.it/7250/ |
id |
ndltd-unibo.it-oai-amsdottorato.cib.unibo.it-7250 |
---|---|
record_format |
oai_dc |
spelling |
ndltd-unibo.it-oai-amsdottorato.cib.unibo.it-72502016-09-01T05:14:36Z Large Covariance Matrix Estimation by Composite Minimization Farne', Matteo <1988> SECS-S/01 Statistica The present thesis concerns large covariance matrix estimation via composite minimization under the assumption of low rank plus sparse structure. Existing methods like POET (Principal Orthogonal complEment Thresholding) perform estimation by extracting principal components and then applying a soft thresholding algorithm. In contrast, our method recovers the low rank plus sparse decomposition of the covariance matrix by least squares minimization under nuclear norm plus $l_1$ norm penalization. This non-smooth convex minimization procedure is based on semidefinite programming and subdifferential methods, resulting in two separable problems solved by a singular value thresholding plus soft thresholding algorithm. The most recent estimator in literature is called LOREC (Low Rank and sparsE Covariance estimator) and provides non-asymptotic error rates as well as identifiability conditions in the context of algebraic geometry. Our work shows that the unshrinkage of the estimated eigenvalues of the low rank component improves the performance of LOREC considerably. The same method also recovers covariance structures with very spiked latent eigenvalues like in the POET setting, thus overcoming the necessary condition $p\leq n$. In addition, it is proved that our method recovers structures with intermediate degrees of spikiness, obtaining a loss which is bounded accordingly. Then, an ad hoc model selection criterion which detects the optimal point in terms of composite penalty is proposed. Empirical results coming from a wide original simulation study where various low rank plus sparse settings are simulated according to different parameter values are described outlining in detail the improvements upon existing methods. Two real data-sets are finally explored highlighting the usefulness of our method in practical applications. Alma Mater Studiorum - Università di Bologna Montanari, Angela 2016-03-04 Doctoral Thesis PeerReviewed application/pdf en http://amsdottorato.unibo.it/7250/ info:eu-repo/semantics/embargoedAccess info:eu-repo/date/embargoEnd/2016-12-20 |
collection |
NDLTD |
language |
en |
format |
Doctoral Thesis |
sources |
NDLTD |
topic |
SECS-S/01 Statistica |
spellingShingle |
SECS-S/01 Statistica Farne', Matteo <1988> Large Covariance Matrix Estimation by Composite Minimization |
description |
The present thesis concerns large covariance matrix estimation via composite minimization under the assumption of low rank plus sparse structure.
Existing methods like POET (Principal Orthogonal complEment Thresholding) perform estimation
by extracting principal components and then applying a soft thresholding algorithm.
In contrast, our method recovers the low rank plus sparse decomposition of the covariance matrix
by least squares minimization under nuclear norm plus $l_1$ norm penalization.
This non-smooth convex minimization procedure is based on semidefinite programming and subdifferential methods,
resulting in two separable problems solved by a singular value thresholding plus soft thresholding algorithm.
The most recent estimator in literature is called LOREC (Low Rank and sparsE Covariance estimator) and provides non-asymptotic error rates as well as identifiability conditions in the context of algebraic geometry.
Our work shows that the unshrinkage of the estimated eigenvalues of the low rank component improves the performance of LOREC considerably.
The same method also recovers covariance structures with very spiked latent eigenvalues like in the POET setting, thus overcoming the necessary condition $p\leq n$.
In addition, it is proved that our method recovers structures with intermediate degrees of spikiness, obtaining a loss which is bounded accordingly.
Then, an ad hoc model selection criterion which detects the optimal point in terms of composite penalty is proposed. Empirical results coming from a wide original simulation study where various low rank plus sparse settings are simulated according to different parameter values are described outlining in detail the improvements upon existing methods. Two real data-sets are finally explored highlighting the usefulness of our method in practical applications. |
author2 |
Montanari, Angela |
author_facet |
Montanari, Angela Farne', Matteo <1988> |
author |
Farne', Matteo <1988> |
author_sort |
Farne', Matteo <1988> |
title |
Large Covariance Matrix Estimation by Composite Minimization |
title_short |
Large Covariance Matrix Estimation by Composite Minimization |
title_full |
Large Covariance Matrix Estimation by Composite Minimization |
title_fullStr |
Large Covariance Matrix Estimation by Composite Minimization |
title_full_unstemmed |
Large Covariance Matrix Estimation by Composite Minimization |
title_sort |
large covariance matrix estimation by composite minimization |
publisher |
Alma Mater Studiorum - Università di Bologna |
publishDate |
2016 |
url |
http://amsdottorato.unibo.it/7250/ |
work_keys_str_mv |
AT farnematteo1988 largecovariancematrixestimationbycompositeminimization |
_version_ |
1718381327740305408 |