Parallel software codes in high performance computing (HPC) continue to ...
Performance analysis is an important part of the oft-repeated, iterative...
As state-of-the-art neural networks scale to billions of parameters,
des...
Mixture-of-Experts (MoE) is a neural network architecture that adds spar...
Parallel training of neural networks at scale is challenging due to
sign...
Interactive visualization via direct manipulation has inherent design
tr...
The field of deep learning has witnessed a remarkable shift towards extr...
In the last few years, the memory requirements to train state-of-the-art...
In recent years, several HPC facilities have started continuous monitori...
Optimizing the performance of large-scale parallel codes is critical for...