- Видео 14
- Просмотров 298 912
Song Han
Добавлен 2 авг 2012
Видео
TinyML and Efficient Deep Learning
Просмотров 1,7 тыс.2 года назад
once-for-all: ofa.mit.edu mcunet: mcunet.mit.edu tinyTL: tinyml.mit.edu/projects/tinyml/tinyTL/
TinyML and Efficient Deep Learning on IoT Devices
Просмотров 2,1 тыс.2 года назад
Presentation at Duke University, NSF AI Institute for Edge Computing: This video covers three TinyML papers: 1. MCUNet: Tiny Deep Learning on IoT Devices [NeurIPS’20] 2. MCUNetV2: Memory-Efficient Patch-based Inference for Tiny Deep Learning [NeurIPS’21] 3. Network Augmentation for Tiny Deep Learning [ICLR’22] slides: www.dropbox.com/s/n1evxyszq70pftv/Duke Athena talk MCUNet.pdf?dl=0
MIT Prof Song Han on TinyML: Reducing AIs Carbon Footprint. Stanford MLSys Seminar Episode 9
Просмотров 7 тыс.3 года назад
MIT Prof Song Han on TinyML: Reducing AIs Carbon Footprint. Stanford MLSys Seminar Episode 9
Efficient AI: Reducing the Carbon Footprint of AI in the Internet of Things (IoT)
Просмотров 2,1 тыс.3 года назад
MCUNet: computation-efficient inference on device [1] TinyTL: computation-efficient transfer learning on device [2] DiffAugment: data-efficient GAN training [3] [1] MCUNet: Tiny Deep Learning on IoT Devices, arxiv.org/pdf/2007.10319.pdf [2] TinyTL: Reduce Memory, Not Parameters for Efficient On-Device Learning, arxiv.org/pdf/2007.11622.pdf [3] Differentiable Augmentation for Data-Efficient GAN ...
"Once-for-All" DNNs: Simplifying Design of Efficient Models for Diverse Hardware
Просмотров 1 тыс.4 года назад
Presentation at edge ai vision alliance: www.edge-ai-vision.com/2020/08/once-for-all-dnns-simplifying-design-of-efficient-models-for-diverse-hardware-a-presentation-from-mit/
Lite Transformer and Hardware-aware Transformer, Invited Talk @ Microsoft Research:
Просмотров 1,6 тыс.4 года назад
Transformers are essential for NLP applications. However, they are challenging to be deployed on mobile devices due to the intensive computation and memory requirement. We solve the challenge from two aspects: hardware-aware neural architecture search (NAS) and new primitive design. On one hand, to enable low-latency inference on resource-constrained hardware platforms, we propose to design Har...
CVPR'20 tutorial: AutoML for TinyML with Once-for-All Network
Просмотров 1,6 тыс.4 года назад
Once for All: Train One Network and Specialize it for Efficient Deployment, ICLR'2020 #NAS, #TinyML, #EfficientAI Website: ofa.mit.edu Paper: arxiv.org/pdf/1908.09791.pdf Code: github.com/mit-han-lab/once-for-all
Song Han's PhD Defense. June 1, 2017 @Stanford
Просмотров 248 тыс.7 лет назад
Song Han received the Ph.D. degree from Stanford University advised by Prof. Bill Dally. His research focuses on energy-efficient deep learning, at the intersection between machine learning and computer architecture. Song will be joining MIT as assistant professor in July 2018.
Toward Efficient Deep Neural Network Deployment: Deep Compression and EIE, Song Han
Просмотров 17 тыс.8 лет назад
Neural networks are both computationally and memory intensive, making them difficult to deploy on embedded systems with limited hardware resources. Song Han explains how deep compression addresses this limitation by reducing the storage requirement of neural networks by 10x-49x without affecting their accuracy and proposes an energy-efficient inference engine (EIE) that accelerates on the spars...
EIE: Efficient Inference Engine on Compressed Deep Neural Network
Просмотров 3,8 тыс.8 лет назад
Song Han presented on 43rd International Symposium on Computer Architecture (ISCA'16) isca2016.eecs.umich.edu
ICLR 2016 Best Paper Award: Deep Compression by Song Han
Просмотров 7 тыс.8 лет назад
Neural networks are both computationally intensive and memory intensive, making them difficult to deploy on embedded systems with limited hardware resources. To address this limitation, we introduce “deep compression”, a three stage pipeline: pruning, trained quantization and Huffman coding, that work together to reduce the storage requirement of neural networks by 35× to 49× without affecting ...
Techniques for Efficient Implementation of Deep Neural Networks, Song Han @ Embedded vision summit
Просмотров 1,8 тыс.8 лет назад
Deep compression and EIE: Deep learning model compression, design space exploration and hardware acceleration.