TonY: An Orchestrator for Distributed Machine Learning Jobs

03/24/2019
by   Anthony Hsu, et al.
0

Training machine learning (ML) models on large datasets requires considerable computing power. To speed up training, it is typical to distribute training across several machines, often with specialized hardware like GPUs or TPUs. Managing a distributed training job is complex and requires dealing with resource contention, distributed configurations, monitoring, and fault tolerance. In this paper, we describe TonY, an open-source orchestrator for distributed ML jobs built at LinkedIn to address these challenges.

READ FULL TEXT

Please sign up or login with your details

Forgot password? Click here to reset