Connecting MapReduce Computations to Realistic Machine Models

02/18/2020
by   Peter Sanders, et al.
0

We explain how the popular, highly abstract MapReduce model of parallel computation (MRC) can be rooted in reality by explaining how it can be simulated on realistic distributed-memory parallel machine models like BSP. We first refine the model (MRC^+) to include parameters for total work w, bottleneck work ŵ, data volume m, and maximum object sizes m̂. We then show matching upper and lower bounds for executing a MapReduce calculation on the distributed-memory machine –Θ(w/p+ŵ+log p) work and Θ(m/p+m̂+log p) bottleneck communication volume using p processors.

READ FULL TEXT
research
09/30/2020

Communication-Optimal Parallel Standard and Karatsuba Integer Multiplication in the Distributed Memory Model

We present COPSIM a parallel implementation of standard integer multipli...
research
02/21/2022

I/O-Optimal Algorithms for Symmetric Linear Algebra Kernels

In this paper, we consider two fundamental symmetric kernels in linear a...
research
11/20/2018

Parallel Matrix Condensation for Calculating Log-Determinant of Large Matrix

Calculating the log-determinant of a matrix is useful for statistical co...
research
09/29/2020

Communication Lower-Bounds for Distributed-Memory Computations for Mass Spectrometry based Omics Data

Mass spectrometry based omics data analysis require significant time and...
research
04/18/2022

Communication Bounds for Convolutional Neural Networks

Convolutional neural networks (CNNs) are important in a wide variety of ...
research
07/18/2023

Scalable Auction Algorithms for Bipartite Maximum Matching Problems

In this paper, we give new auction algorithms for maximum weighted bipar...
research
10/14/2020

Taurus: Lightweight Parallel Logging for In-Memory Database Management Systems (Extended Version)

Existing single-stream logging schemes are unsuitable for in-memory data...

Please sign up or login with your details

Forgot password? Click here to reset