Optimization of Decision Tree Evaluation Using SIMD Instructions

05/15/2022
by   Alexey Mironov, et al.
0

Decision forest (decision tree ensemble) is one of the most popular machine learning algorithms. To use large models on big data, like document scoring with learning-to-rank models, we need to evaluate these models efficiently. In this paper, we explore MatrixNet, the ancestor of the popular CatBoost library. Both libraries use the SSE instruction set for scoring on CPU. This paper investigates the opportunities given by the AVX instruction set to evaluate models more efficiently. We achieved 35 (nodes conditions comparison), and 20 ranking model.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
11/01/2022

Optimization of Oblivious Decision Tree Ensembles Evaluation for CPU

CatBoost is a popular machine learning library. CatBoost models are base...
research
06/17/2022

Popular decision tree algorithms are provably noise tolerant

Using the framework of boosting, we prove that all impurity-based decisi...
research
08/28/2015

A Comparison Between Decision Trees and Decision Tree Forest Models for Software Development Effort Estimation

Accurate software effort estimation has been a challenge for many softwa...
research
10/27/2020

A short note on the decision tree based neural turing machine

Turing machine and decision tree have developed independently for a long...
research
10/26/2017

Big Data Classification Using Augmented Decision Trees

We present an algorithm for classification tasks on big data. Experiment...
research
04/25/2019

TreeGrad: Transferring Tree Ensembles to Neural Networks

Gradient Boosting Decision Tree (GBDT) are popular machine learning algo...
research
09/18/2019

The Impact of Operating Environment on Efficiency of Public Libraries

Analysis of technical efficiency is an important tool in management of p...

Please sign up or login with your details

Forgot password? Click here to reset