MAGMA: An Optimization Framework for Mapping Multiple DNNs on Multiple Accelerator Cores

Date:

Abstract

As Deep Learning continues to drive a variety of applications in edge and cloud data centers, there is a growing trend towards building large accelerators with several subaccelerator cores/chiplets. This work looks at the problem of supporting multi-tenancy on such accelerators. In particular, we focus on the problem of mapping jobs from several DNNs simultaneously on an accelerator. Given the extremely large search space, we formulate the search as an optimization problem and develop an optimization framework called M3E. In addition, we develop a specialized optimization algorithm called MAGMA with custom operators to enable structured sampleefficient exploration. We quantitatively compare MAGMA with several state-of-the-art methods, black-box optimization, and reinforcement learning methods across different accelerator settings (large/small accelerators) and different subaccelerator configurations (homogeneous/heterogeneous), and observe MAGMA can consistently find better mappings. img_2.png


Code Available

MAGMA Code-base


Youtube

IMAGE ALT TEXT HERE


Paper: MAGMA: An Optimization Framework for Mapping Multiple DNNs on Multiple Accelerator Cores