Matrix Scaling and Balancing via Box Constrained Newton's Method and Interior Point Methods

© 2017 IEEE. In this paper, we study matrix scaling and balancing, which are fundamental problems in scientific computing, with a long line of work on them that dates back to the 1960s. We provide algorithms for both these problems that, ignoring logarithmic factors involving the dimension of the in...

Full description

Bibliographic Details
Main Authors: Cohen, Michael B. (Author), Madry, Aleksander (Author), Tsipras, Dimitris (Author), Vladu, Adrian (Author)
Format: Article
Language:English
Published: IEEE, 2021-11-08T18:39:01Z.
Subjects:
Online Access:Get fulltext
LEADER 02004 am a22001813u 4500
001 137768
042 |a dc 
100 1 0 |a Cohen, Michael B.  |e author 
700 1 0 |a Madry, Aleksander  |e author 
700 1 0 |a Tsipras, Dimitris  |e author 
700 1 0 |a Vladu, Adrian  |e author 
245 0 0 |a Matrix Scaling and Balancing via Box Constrained Newton's Method and Interior Point Methods 
260 |b IEEE,   |c 2021-11-08T18:39:01Z. 
856 |z Get fulltext  |u https://hdl.handle.net/1721.1/137768 
520 |a © 2017 IEEE. In this paper, we study matrix scaling and balancing, which are fundamental problems in scientific computing, with a long line of work on them that dates back to the 1960s. We provide algorithms for both these problems that, ignoring logarithmic factors involving the dimension of the input matrix and the size of its entries, both run in time \widetilde{O}(m\log \kappa \log^2 (1/∈)) where ∈ is the amount of error we are willing to tolerate. Here, \kappa represents the ratio between the largest and the smallest entries of the optimal scalings. This implies that our algorithms run in nearly-linear time whenever \kappa is quasi-polynomial, which includes, in particular, the case of strictly positive matrices. We complement our results by providing a separate algorithm that uses an interior-point method and runs in time \widetilde{O}(m^{3/2} \log (1/∈)).In order to establish these results, we develop a new second-order optimization framework that enables us to treat both problems in a unified and principled manner. This framework identifies a certain generalization of linear system solving that we can use to efficiently minimize a broad class of functions, which we call second-order robust. We then show that in the context of the specific functions capturing matrix scaling and balancing, we can leverage and generalize the work on Laplacian system solving to make the algorithms obtained via this framework very efficient. 
546 |a en 
655 7 |a Article 
773 |t 10.1109/focs.2017.88