计算机科学
互操作性
推论
灵活性(工程)
嵌入式系统
人工智能
分布式计算
深度学习
架空(工程)
计算机体系结构
机器学习
操作系统
统计
数学
作者
Robert David,Jared Duke,Advait Jain,Vijay Janapa Reddi,Nat Jeffries,Jian Li,Nick Kreeger,Ian Nappier,Meghna Natraj,Shlomi Regev,Rocky Rhodes,Tiezhen Wang,Pete Warden
出处
期刊:Cornell University - arXiv
日期:2020-01-01
被引量:274
标识
DOI:10.48550/arxiv.2010.08678
摘要
Deep learning inference on embedded devices is a burgeoning field with myriad applications because tiny embedded devices are omnipresent. But we must overcome major challenges before we can benefit from this opportunity. Embedded processors are severely resource constrained. Their nearest mobile counterparts exhibit at least a 100 -- 1,000x difference in compute capability, memory availability, and power consumption. As a result, the machine-learning (ML) models and associated ML inference framework must not only execute efficiently but also operate in a few kilobytes of memory. Also, the embedded devices' ecosystem is heavily fragmented. To maximize efficiency, system vendors often omit many features that commonly appear in mainstream systems, including dynamic memory allocation and virtual memory, that allow for cross-platform interoperability. The hardware comes in many flavors (e.g., instruction-set architecture and FPU support, or lack thereof). We introduce TensorFlow Lite Micro (TF Micro), an open-source ML inference framework for running deep-learning models on embedded systems. TF Micro tackles the efficiency requirements imposed by embedded-system resource constraints and the fragmentation challenges that make cross-platform interoperability nearly impossible. The framework adopts a unique interpreter-based approach that provides flexibility while overcoming these challenges. This paper explains the design decisions behind TF Micro and describes its implementation details. Also, we present an evaluation to demonstrate its low resource requirement and minimal run-time performance overhead.
科研通智能强力驱动
Strongly Powered by AbleSci AI