Solving Attention Kernel Regression Problem via Pre-conditioner

08/28/2023
by   Zhao Song, et al.
0

Large language models have shown impressive performance in many tasks. One of the major features from the computation perspective is computing the attention matrix. Previous works [Zandieh, Han, Daliri, and Karba 2023, Alman and Song 2023] have formally studied the possibility and impossibility of approximating the attention matrix. In this work, we define and study a new problem which is called the attention kernel regression problem. We show how to solve the attention kernel regression in the input sparsity time of the data matrix.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
03/28/2023

Solving Regularized Exp, Cosh and Sinh Regression Problems

In modern machine learning, attention computation is a fundamental task ...
research
05/01/2023

An Iterative Algorithm for Rescaled Hyperbolic Functions Regression

Large language models (LLMs) have numerous real-life applications across...
research
03/04/2014

Matroid Regression

We propose an algebraic combinatorial method for solving large sparse li...
research
06/11/2020

Implicit Kernel Attention

Attention compute the dependency between representations, and it encoura...
research
09/15/2019

A brief TOGAF description using SEMAT Essence Kernel

This work aims to explore the possibility of describing the enterprise a...
research
04/24/2023

Robust, randomized preconditioning for kernel ridge regression

This paper introduces two randomized preconditioning techniques for robu...
research
02/14/2022

Polyhedron Kernel Computation Using a Geometric Approach

The geometric kernel (or simply the kernel) of a polyhedron is the set o...

Please sign up or login with your details

Forgot password? Click here to reset