Optimal approximate matrix product in terms of stable rank

07/08/2015
by   Michael B. Cohen, et al.
0

We prove, using the subspace embedding guarantee in a black box way, that one can achieve the spectral norm guarantee for approximate matrix multiplication with a dimensionality-reducing map having m = O(r̃/ε^2) rows. Here r̃ is the maximum stable rank, i.e. squared ratio of Frobenius and operator norms, of the two matrices being multiplied. This is a quantitative improvement over previous work of [MZ11, KVZ14], and is also optimal for any oblivious dimensionality-reducing map. Furthermore, due to the black box reliance on the subspace embedding property in our proofs, our theorem can be applied to a much more general class of sketching matrices than what was known before, in addition to achieving better bounds. For example, one can apply our theorem to efficient subspace embeddings such as the Subsampled Randomized Hadamard Transform or sparse subspace embeddings, or even with subspace embedding constructions that may be developed in the future. Our main theorem, via connections with spectral error matrix multiplication shown in prior work, implies quantitative improvements for approximate least squares regression and low rank approximation. Our main result has also already been applied to improve dimensionality reduction guarantees for k-means clustering [CEMMP14], and implies new results for nonparametric regression [YPW15]. We also separately point out that the proof of the "BSS" deterministic row-sampling result of [BSS12] can be modified to show that for any matrices A, B of stable rank at most r̃, one can achieve the spectral norm guarantee for approximate matrix multiplication of A^T B by deterministically sampling O(r̃/ε^2) rows that can be found in polynomial time. The original result of [BSS12] was for rank instead of stable rank. Our observation leads to a stronger version of a main theorem of [KMST10].

READ FULL TEXT

page 1

page 2

page 3

page 4

research
03/30/2014

Approximate Matrix Multiplication with Application to Linear Embeddings

In this paper, we study the problem of approximately computing the produ...
research
07/06/2018

Leveraging Well-Conditioned Bases: Streaming & Distributed Summaries in Minkowski p-Norms

Work on approximate linear algebra has led to efficient distributed and ...
research
11/09/2020

Reduced-Rank Regression with Operator Norm Error

A common data analysis task is the reduced-rank regression problem: ...
research
07/16/2021

Near-Optimal Algorithms for Linear Algebra in the Current Matrix Multiplication Time

Currently, in the numerical linear algebra community, it is thought that...
research
12/27/2017

Sketching for Kronecker Product Regression and P-splines

TensorSketch is an oblivious linear sketch introduced in Pagh'13 and lat...
research
05/29/2017

Coreset Construction via Randomized Matrix Multiplication

Coresets are small sets of points that approximate the properties of a l...
research
09/05/2020

Revisiting Co-Occurring Directions: Sharper Analysis and Efficient Algorithm for Sparse Matrices

We study the streaming model for approximate matrix multiplication (AMM)...

Please sign up or login with your details

Forgot password? Click here to reset