Abstract
Deep Neural Networks (DNNs) have achieved great success in a variety of machine learning (ML) applications, delivering high-quality inferencing solutions in computer vision, natural language processing, virtual reality, etc. However, DNN-based ML applications also bring much increased computational and storage requirements, which are particularly challenging for embedded systems with limited compute/storage resources, tight power budgets, and small form factors. Challenges also come from the diverse application-specific requirements, including real-time responses, high-throughput performance, and reliable inference accuracy. To address these challenges, we introduce a series of effective design methodologies, including efficient ML model designs, customized hardware accelerator designs, and hardware/software co-design strategies to enable efficient ML applications on embedded systems.
Original language | English (US) |
---|---|
Title of host publication | Embedded Machine Learning for Cyber-Physical, IoT, and Edge Computing |
Subtitle of host publication | Software Optimizations and Hardware/Software Codesign |
Publisher | Springer |
Pages | 37-74 |
Number of pages | 38 |
ISBN (Electronic) | 9783031399329 |
ISBN (Print) | 9783031399312 |
DOIs | |
State | Published - Jan 1 2023 |
Keywords
- Compilation
- Deep Neural Networks
- Efficient ML model
- Embedded systems
- Hardware accelerator
- Hardware/software co-design
- Machine learning
- Optimization
ASJC Scopus subject areas
- General Computer Science
- General Engineering
- General Social Sciences