April 24, 2019
MTL Seminar Series

Energy efficient deep neural network inference hardware for consumer devices

Paul Whatmough, Arm Research
  • Speaker
  • Abstract
  • Exclusive Content

Paul Whatmough received the Doctorate degree from University College London, UK. He was previously with Philips, NXP, Arm, and Harvard, working in the areas of ML, DSP, wireless, accelerators, and circuits. Currently, he leads research at Arm ML Research Lab Boston, and is an Associate at Harvard.

Deep neural networks (DNNs) have quickly become an essential workload across computing form factors, including IoT, mobile, automotive and datacenter. However, DNN inference demands an enormous number of arithmetic operations and a large memory footprint. In this talk, we will explore the co-design of DNN models and hardware to achieve state-of-the-art performance for real-time, energy-constrained inference applications.

This content is restricted to our MIG members and members of the MIT community. Login below, or contact us for more information about our partner programs.