MCU AI/ML - Bridging the Gap Between Intelligence and Embedded Systems
Artificial Intelligence (AI) and Machine Learning (ML) are key technologies that enable systems to learn from data, make inferences, and enhance their performance over time. These technologies have often been used in large-scale data centers and powerful GPUs, but there's an increasing demand to deploy them on resource-limited devices, such as microcontrollers (MCUs).
In this blog, we will examine the intersection of MCU technology and AI/ML, and how it affects low-power edge devices. We'll discuss the difficulties, innovations, and practical use cases of running AI on battery-operated MCUs.
AI/ML and MCUs: A Brief Overview
AI creates computer systems that can do human-like tasks, such as understanding language, finding patterns, and deciding. Machine Learning, a subset of AI, involves using algorithms that let computers learn from data and get better over time. ML models can find patterns, sort objects, and predict outcomes from examples.
MCUs play an important role in making AI and ML possible on edge devices.
Some use cases for MCU based AI/ML at the edge include:
Keyword Spotting: Recognize specific words or phrases (e.g., voice commands) without the need for cloud connectivity
Sensor Fusion: Combining data from multiple sensors to make more informed decisions than with single sensor solutions
Anomaly detection: Detecting outliers or abnormal patterns in sensor data that may indicate faults, errors or threats for predictive maintenance or quality control
Object detection: Identifying and locating objects of interest (e.g., faces, pedestrians, vehicles) in images or videos captured by cameras or other sensors.
Gesture recognition: Interpreting human gestures (e.g., hand movements, facial expressions, body poses) in images or videos captured by cameras or other sensors to improve human computer interaction
Challenges of AI/ML on MCUs
Deep learning models, particularly deep neural networks (DNNs), have become indispensable for complex tasks like computer vision and natural language processing. However, their computational demands are substantial. Such resource-intensive models are impractical for everyday devices, especially those powered by low-energy MCUs found in edge devices. The growth of deep learning model complexity is undeniable. As DNNs become more sophisticated, their size balloons, making them incompatible with the limited computing resources available on MCUs.
What is TinyML?
TinyML refers to machine learning models and techniques optimized for deployment on resource-constrained devices. These devices operate at the edge, where data is generated, and inferencing is performed locally. Typically run on low power MCUs, TinyML systems perform inferences on data collected locally to the node. Inferencing is the moment of truth for an AI model, testing how well it can apply knowledge learned during training. Local inferencing enables MCUs to execute AI models directly, making real-time decisions without relying on external servers or cloud services.
Local inferencing in the context of AI/ML is crucial for several reasons:
Resource Constraints: Many embedded devices, especially those running on battery power, have limited resources such as memory, processing capability, and energy efficiency. Traditional general-purpose microcontrollers struggle to perform AI tasks efficiently due to their limited processing power and memory, constrained energy resources, or lack of on-chip acceleration. Local inferencing allows these resource-constrained devices to execute AI workloads without draining excessive power to improve efficiency and performance for things like:
User Experience Enhancement: Consider an example: an AI-enabled electronic cat flap. By training it to distinguish between cats and other objects, it can open the door only for the authorized cat. Here, local inferencing improves the user experience by ensuring safety and convenience without the need for additional hardware like RFID collars.
Efficiency and Performance: GPUs are commonly used for large-scale AI deployments because they can perform many processes in parallel, essential for effective AI training. However, GPUs are costly and exceed power budgets for small-scale embedded applications. AI-optimized MCUs, with specialized architectures, strike a balance by delivering better performance and power efficiency for AI workloads. SILICON LABS includes a matrix vector processor as part of its AI/ML enablement. This specialized peripheral is designed to enhance the performance of AI/ML algorithms or vector math operations to shorten inferencing time and perform these critical tasks at lower power.
In summary, local inferencing at the edge empowers real-time decision-making, reduces latency, enhances security, empowers battery-operated devices with AI capabilities and enhances user experiences making it a critical component of modern computing systems, while respecting resource limitations.
Silicon Labs Pioneering AI/ML Solutions for the Edge:
In the dynamic landscape of technology, Silicon Labs stands out as a trailblazer in bringing Artificial intelligence (AI) and Machine learning (ML) to the edge. Our commitment to innovation has led to groundbreaking solutions that empower resource-constrained devices, such as microcontrollers (MCUs), with intelligent capabilities.
Devices Optimized for TinyML
The EFR32xG24, EFR32xG28, and EFR32xG26 families of MCUs and Wireless MCUs combine a 78 MHz ARM Cortex®-M33 processor, high-performance radios, precision analog performance, and an AI/ML hardware accelerator giving developers a flexible platform for deploying edge intelligence. Supporting a broad range of wireless IoT protocols, these SoCs incorporate the highest security with the best RF performance/energy-efficiency ratio in the market.
Today’s developers are often forced to pay steep performance or energy penalties for deploying AI/ML at the edge. The xG24, xG28, and xG26 families alleviate those penalties as the first ultra-low powered devices with dedicated AI/ML accelerators built in lowering overall design complexity. This specialized hardware is designed to handle complex calculations up to an 8x faster inferencing along with up to a 6x improvement in energy efficiency when compared to a firmware only approach and with even more performance gained when compared to cloud-based solutions. The use of the hardware accelerator offloads the burden of inferencing from the main application MCU leaving more clock cycles available to service your application.
Tools for Simplifying AI/ML Development
The tools to build, test, and deploy the algorithms needed for machine learning are just as important as the MCUs running those algorithms. By partnering with leaders in the TinyML space like TensorFlow, SensiML, and Edge Impulse, Silicon Labs provides options for beginners and experts alike. Using this new AI/ML toolchain with Silicon Labs’s Simplicity Studio, developers can create applications that draw information from various connected devices to make intelligent machine learning-driven decisions.
Silicon Labs provides a variety of tools and resources to support machine learning (ML) applications. Here are some of them:
Machine Learning Applications: The development platform supports embedded machine learning (TinyML) model inference, backed by the TensorFlow Lite for Microcontrollers (TFLM) framework. The repository contains a collection of embedded applications that leverage ML.
Machine Learning Toolkit (MLTK): This is a Python package with command-line utilities and scripts to aid the development of machine-learning models for Silicon Lab's embedded platforms. It includes features for executing ML operations from a command-line interface or a Python script, determining how efficiently an ML model will execute on an embedded platform, and training an ML model using Google TensorFlow.
Silicon Labs provides a TinyML solution as part of the Machine Learning Toolkit (MLTK). The toolkit includes several models that are used by the TinyML benchmark. These models are available on the Silicon Labs GitHub and include anomaly detection, image classification, and keyword spotting.
AI/ML powered edge devices are opening new horizons for how we engage with our surroundings, and they will soon transform our lives in amazing ways. Silicon Labs is at the forefront of TinyML innovation, making it possible to bring these capabilities to low power, connected edge devices like never before.
Learn more about how our EFR and EFM MCU platform is optimized for AI/ML at the Edge in our recent Wireless Compute Tech Talk session, An Optimized Platform for AI/ML at the Edge.
- |
- +1 赞 0
- 收藏
- 评论 0
本文由飞猫警长转载自SILICON LABS Official Website,原文标题为:MCU AI/ML - Bridging the Gap Between Intelligence and Embedded Systems,本站所有转载文章系出于传递更多信息之目的,且明确注明来源,不希望被转载的媒体或个人可与我们联系,我们将立即进行删除处理。
相关研发服务和供应服务
相关推荐
行业前沿的MCU+AI/ML开发工具弥合智能和嵌入式系统之间的差距
在本文中,Silicon Labs产品营销高级经理Gopinath Krishniah先生将带您探究MCU技术和AI/ML的交叉与汇合,以及它如何影响低功耗边缘设备的发展;同时将讨论在电池供电设备的MCU上运行AI的困难、创新和实际用例,并进一步介绍芯科科技专为边缘智能开发所提供全套MCU+AI/ML工具的解决方案。
技术探讨 发布时间 : 2024-07-11
Leveraging Helium and ARM® Cortex®-M85 for Unprecedented DSP and AI Performance on an MCU Core
In conclusion, the Cortex-M85 with Helium can contribute to a significant uplift in AI/ML and DSP performance while outshining the rest of the Cortex-M cores in scalar performance. This makes it an ideal choice for more complex processing tasks.
技术探讨 发布时间 : 2023-04-26
Power Your Edge AI Application with the Industry’s Most Powerful Arm MCUs
As is evident, RA8 MCUs with Helium can significantly improve neural network performance without the need for any additional hardware acceleration, thus providing a low-cost, low-power option for the implementation of simpler AI and machine learning use cases.
技术探讨 发布时间 : 2023-11-03
AI加速边缘计算,聚焦AIOT芯片,NPU SOC,离线语音MCU,高算力智能模组等
世强硬创联合地平线,阿普奇,启英泰伦,美格智能,普林芯驰,唯创知音,九芯电子,芯闻,VINKO,MERRY带来AI新产品,聚焦AIOT芯片,NPU SOC,离线语音MCU,高算力智能模组等,加速边缘计算。
活动 发布时间 : 2023-06-08
Silicon Labs(芯科科技)EFM8™ 8位MCU选型指南
型号- EFM8UB20F32G,EFM8UB20F64G,EFM8,EFM8SB20F64G,SLSTK2000A,SLSTK2022A,SLSTK2010A,SLSTK2020A,SLSTK2030A,EFM8UB10F8G,EFM8SB20F32G,EFM8SB20F16G,SLTB005A,EFM8LB11F32E,EFM8UB11F16G,EFM8SB10F2G,EFM8BB21F16G,EFM8SB10F4G,EFM8LB11F16E,EFM8BB10F2G,EFM8SB10F8G,EFM8 FAMILY,EFM8LB12F64E,EFM8LB10F16E,EFM8LB12F32E,EFM8UB10F16G,EFM8UB31F40G,EFM8BB31F32G,EFM8BB31F64G,SLSTK2001A,SLSTK2011A,SLSTK2021A,EFM8BB10F4G,EFM8BB31F16G,EFM8UB30F40G,EFM8BB10F8G,EFM8BB22F16G
APM32芯得 EP.38 | TinyMaix赋予APM32F411 AI推理能力
TinyMaix是矽速科技开发的轻量级机器学习库,适用于微控制器,能在资源受限的MCU上运行深度学习模型。它支持多种芯片架构和模型转换,具有低内存消耗和用户友好接口。文章介绍了TinyMaix在APM32F411 MCU上的移植过程,包括源码准备、工程配置、编译器设置、解决编译错误和实现计时函数。移植后,TinyMaix能够成功运行手写数字识别、人像检测和图片分类等实例。
设计经验 发布时间 : 2024-10-30
他山(TASHAN)AI触觉传感芯片选型指南
描述- TS3F605 是一款集成了一个 30通道 24 位 电容-数字转换(CDC)、内置 R-SpiNNaker 四核类脑芯片架构的 32位SOC 芯片。内置的 24 位 CDC 可配置成电容测量模式或电压测量模式,这两种模式都可配置成单端模式及差分模式。其中电容测量模式即可实现互电容的测量,又可实现自电容的测量。30 个通道通过内置的模拟路由可配置每个通道不同的功能,可实现 CDC 的片间路由。内置 R-SpiNNaker 四核类脑芯片架构,可实现分布式计算。
型号- TS4M系列,TS3系列,TS3A,TS4M41X,TS4M21X,TS4M43X,TS3M0,TS3F505,TS3F605,TS3F,TS3,TS3F600,TS3M04X,TS3000,TS3M02X,TS3F500,TS3005,TS3A0,TS3A805,TS4M42X,TS4M22X,TS3A700,TS3A800,TS4M23X,TS3A0系列,TS3A705,TS3A系列,TS3M0系列,TS3M03X,TS3F系列,TS3M01X,TS4M
携手无线技术联盟扩展IoT战略-全面布局创新技术和人工智能+安全
物联网(IoT)无线连接技术时刻在发生新变化。近日,物联传媒记者在深圳物联网展会期间特别采访了Silicon Labs亚太区生态高级经理刘俊先生,此次访谈紧密围绕公司在刚结束的深圳物联网展中亮相的前沿科技成果,深入讨论了技术创新、产品优势、应用前景等话题。
原厂动态 发布时间 : 2024-10-23
启英泰伦(Chipintelli)AI语音芯片选型指南
描述- 启英泰伦成立于2015年,是集语音芯片、语音算法、方案、平台于一体的行业领导型智能语音解决方案供应商,公司产品已广泛应用于智能家居家电、照明、玩具、IoT、车载、音响、机器人等领域,是端侧人工智能芯片及解决方案的行业开拓者和领导者。
型号- CI130X,CI130X系列,CI131系列,CI13,CI1311,CI11,CI1312,CI2305,CI231系列,CI2306,CI230系列,CI132X,CI1122,CI1002,CI2311,CI231,CI1102,CI2312,CI1103,CI1301,CI1302,CI1006,CI13系列,CI230,CI1303,CI131,CI11系列,CI1306
芯科科技PG28 32位微控制器配备AI/ML硬件加速器,可在较低能耗的情况下将机器学习推理的性能加倍
本文重点介绍了Silicon Labs(亦称“芯科科技”)的PG28 32位微控制器(MCU)的特性和优势,其领先集成了人工智能和机器学习(AI/ML)硬件加速器,可在边缘位置以更低功耗进行更快推理;同时还兼容EFR32xG28 无线SoC 平台(ZG28、FG28 和 SG28),为开发人员提供实现各种低功耗、高性能嵌入式物联网应用的理想选择。
原厂动态 发布时间 : 2024-06-22
AI融合物联网大势所趋,ML语音识别和手势控制应用分享
芯科科技作为智能、安全物联网无线连接领域的开拓者,正在致力于将AI/ML带到边缘。我们对创新的承诺导致了开创性的解决方案,它赋予资源受限的设备如MCU具备更丰富的智能功能。
原厂动态 发布时间 : 2024-09-30
招聘高算力MCU/SOC工程师,负责智能驾驶、AI、机器人等热门项目
参与汽车电子、人工智能、工业等热门项目的研发。技术大牛带队、接触全品类器件,快速拓宽市场应用领域和技术的广度、深度;服务中国TOP2000+知名硬科技企业,丰厚项目奖、年终奖。
招聘信息 发布时间 : 2024-08-08
AN928.2: EFR32 Series 2 Layout Design Guide
型号- EFR32XG25,EFR32XG23,EFR32XG24,EFR32XG27,EFR32XG28,EFR32XG21-C,EFR32XG21-B,EFR32BG27,EFR32ZG28,EFR32FG23,EFR32FG22,EFR32BG22,EFR32FG28,EFR32,EFR32BG21,EFR32BG24,EFR32FG25,EFR32MG27,EFR32MG24,EFR32MG22,EFR32MG21,EFR32ZG23,EFR32XG21,EFR32 SERIES,EFR32XG22
电子商城
现货市场
登录 | 立即注册
提交评论