Arm's latest IoT CPU and microNPU combine for an unprecedented performance boost. But the real value, says Dipti Vachani, will be in the use cases they'll enable for AI developers
By Dipti Vachani, SVP and GM of the Automotive Line of Business, Arm
After a couple of decades working in the tech sector I’m not easily surprised. But I was astonished to learn that our engineering team had pushed up the artificial intelligence (AI) performance of our newest cores by nearly 500 times.
It was the back end of 2019 and the team had just finished final testing of the new microcontroller and neural network processors we launched today. By pairing the Arm Cortex-M55 CPU and Arm Ethos-U55 micro neural processing unit (microNPU), our engineers had reinvented the performance of Arm-based Internet of Things (IoT) solutions. They’d set the course for the future of what developers might do creatively with AI for IoT.
I realized quickly that our engineers had opened up the last frontier in AI for IoT devices. There were almost no applications I could imagine where it wouldn’t be possible to layer in machine learning (ML) now. And given what I was already seeing from the vast AI developer ecosystem, I knew it wouldn’t be long before we see this new capability turn into real-world products.
Enabling new, advanced use cases
The best technology for me is the kind you wish someone had invented earlier. But the simple truth is that the best ideas must often wait for technology to catch up. Autonomous cars are a great example: they’ll be life-changing, but we have to stretch the realms of what’s possible today to make them real at scale.
Another less top-of-the-mind example of AI for IoT is something that exists already but the Cortex-M55 and Ethos-U55 can take it to a new level. This device isn’t a must-have for everyone, but for those who need it, it may be life-changing. Here are some stats: 10 million people in the US are registered blind or partially-sighted. Globally, that figure is close to 300 million. The device I’m talking about is a connected walking stick. Early versions exist already and here’s one I found recently: WeWalk. These devices use sensing and navigation to intelligently guide people around hazards in the home and outside.
I spoke to the Arm healthcare innovation team about how our new Cortex and Ethos processors might deliver a major upgrade. They told me the biggest challenge for connected stick developers today was a reliance on ultrasound positioning. But ultrasound has limited range and fidelity, and the sensing, processing and power packs needed tend to bloat designs.
Our new processors could enable developers to swap out ultrasound for AI-enabled visual sensing using a 360-degree camera. That, alongside wireless communications and navigation, would be powered by a super-slim battery capable of lasting all day. The device would look like a conventional stick but be the ultimate visual aid. And with the AI compute performed locally, it wouldn’t matter if cell connectivity was lost.
It’s all about team play
As with all solutions though, it’s not just about the star players, whoever they may be, it’s about the whole team. Ultimately, it won’t be Arm or most of our chip partners coming up with the mind-blowing ideas behind the AI for IoT revolution, it will be the massive developer network working with our technology all over the world. So, more than ever, our focus needs to be on the team.
Software development in the AI for IoT space is highly complex, spanning a diverse range of use cases and hardware. And arguably Arm’s biggest team job is to ensure developers’ lives don’t get any harder as we move through technology iterations. We’ve put a lot of thought into that. In particular the evolution of hardware abstraction layers such as CMSIS-Core, as well as CMSIS-DSP and CMSIS-NN APIs and libraries. The work we’ve done means any developer writing ML code for Arm Cortex-M based devices will be able to port that work to use on Cortex-M55 and Ethos-U55 based chips.
AI for IoT will ‘just work’
Another developer concern is ensuring favorite frameworks and toolchains are fully supported by hardware providers. This is at the heart of our CMSIS work, and our collaboration with fellow AI for IoT innovators like Google to ensure TensorFlow Lite Micro will be fully supported by the Cortex-M55 and Ethos-U55 toolchain.
Check out the thoughts of Ian Nappier, Tensor Flow Lite Micro product manager, below. The drivers for these processors will automatically optimize developers’ TensorFlow models to take full advantage of any hardware configuration they want to deploy.
Fast forward to the future
We often talk about the Fifth Wave of Computing, the combination of AI with IoT and 5G, and I already see AI and IoT coming together in huge numbers of devices. Our new Cortex and Ethos processor pairing will take that to a new level of performance. Get the technical detail here.
latest launch maps exactly to Arm’s grander vision for distributed computing: boosting
what’s possible with AI for IoT, efficiently distributing as much compute as we
can at the endpoint, building an AI-driven Network Edge and pushing more
efficiency and cost-saving in the cloud with new server technologies such as Graviton.
But we learned from the recently-published Arm 2020 Global AI Surveythat up to three-quarters of consumers prefer their data to be processed on-device, with only crucial data sent to the cloud. By elevating the AI capability of devices so they don’t have to rely on cloud compute, we enable a richer framework for trust: Sensitive data access can be controlled between the compute layers.
Over to you, developers
I’m delighted that we’ve pushed up the performance of AI for IoT endpoints by almost 500x with our Cortex-M55 and Ethos-U55 processors. But it’s what they will enable that really excites me, inspiring possibilities as developers push AI down into the tiniest of applications and scale it up to grandest of uses. All done more efficiently than ever before.
Finally, coming back to my opening point around how hard it is to surprise me now, I’m already preparing myself for the next joyful moment: when an inventor makes something with our new technology that none of us dreamt of. I’m sure my reaction will be mixture of ‘wow’ and ‘I wish I’d thought of that’. But most of all, I’ll be celebrating the impact that Arm, our ecosystem and its developers will have on the world.
Learn more about how the Arm Cortex-M55 and Arm Ethos-U55 enable the future of AI for IoT endpoint devices.
By Dipti Vachani, SVP and GM of the Automotive Line of Business, Arm
Any re-use permitted for informational and non-commercial or personal use only.
Day one of #MWC24 showed us how sustainable mobile networks can fuel the growth of AI to deliver unparalleled user experiences.
We're pleased to work closely with our partner ecosystem to build this future #onArm Neoverse. Here's how: https://newsroom.arm.com/sustainable-cloud-infrastructure?utm_source=twitter&utm_medium=social-organic&utm_content=blog&utm_campaign=mk19_events_na_2023
We're joining forces with tech leaders and academic institutions as a member of the AI-RAN Alliance, announced today at #MWC24.🙌
Tying research together with testing and implementation, we'll explore how to use AI to advance RAN technology for 5G and 6G: https://ai-ran.org/news/industry-leaders-in-ai-and-wireless-form-ai-ran-alliance/