Two Phases of Scaling Laws for Nearest Neighbor Classifiers

08/16/2023
by   Pengkun Yang, et al.
0

A scaling law refers to the observation that the test performance of a model improves as the number of training data increases. A fast scaling law implies that one can solve machine learning problems by simply boosting the data and the model sizes. Yet, in many cases, the benefit of adding more data can be negligible. In this work, we study the rate of scaling laws of nearest neighbor classifiers. We show that a scaling law can have two phases: in the first phase, the generalization error depends polynomially on the data dimension and decreases fast; whereas in the second phase, the error depends exponentially on the data dimension and decreases slowly. Our analysis highlights the complexity of the data distribution in determining the generalization error. When the data distributes benignly, our result suggests that nearest neighbor classifier can achieve a generalization error that depends polynomially, instead of exponentially, on the data dimension.

READ FULL TEXT

page 3

page 14

page 18

research
10/09/2014

Speculate-Correct Error Bounds for k-Nearest Neighbor Classifiers

We introduce the speculate-correct method to derive error bounds for loc...
research
02/08/2021

Learning Curve Theory

Recently a number of empirical "universal" scaling law papers have been ...
research
08/20/2019

Multi-hypothesis classifier

Accuracy is the most important parameter among few others which defines ...
research
03/24/2023

kNN Prompting: Beyond-Context Learning with Calibration-Free Nearest Neighbor Inference

In-Context Learning (ICL), which formulates target tasks as prompt compl...
research
12/12/2018

Distributed Nearest Neighbor Classification

Nearest neighbor is a popular nonparametric method for classification an...
research
06/11/2013

Efficient Classification for Metric Data

Recent advances in large-margin classification of data residing in gener...
research
02/07/2020

Short sighted deep learning

A theory explaining how deep learning works is yet to be developed. Prev...

Please sign up or login with your details

Forgot password? Click here to reset