Register today for upcoming Arm Tech Talks: [ Ссылка ]
Get ready for another one of our Arm Tech Talks! Every fortnight, we discuss and explore some of the latest trends, technologies and best practices in the world of AI, featuring partners from the AI Ecosystem as well as speakers across Arm.
00:00 Intro
02:30 Meet the presenter
03:44 Presentation Start
04:10 Intro to Plumerai
05:47 Binary NN's
07:14 Inference Engine deep dive
21:13 Demo of public benchmarking service
23:28 Further deep dive
43:05 Q&A
Recently Plumerai announced their inference engine for 8-bit deep learning models on Arm Cortex-M microcontrollers. They showed that it is the world’s most efficient on MobileNetV2, beating TensorFlow Lite for Microcontrollers with CMSIS-NN kernels by 40% in terms of latency and 49% in terms of RAM usage with no loss in accuracy. However, that was just on a single network and it might have been cherry-picked. Therefore, they give a live demonstration of Plumerai Benchmark. This is a new service that you can use to test your own models with both TensorFlow Lite for Microcontrollers and Plumerai’s inference engine.
In this talk they explain what they did to get these speedups and memory improvements and they will show benchmarks for the most important publicly available neural network models.
If you enjoyed this video, please subscribe to our channel and make sure to add us on Twitter so you can get a lot more content like this delivered straight to your feed!
[ Ссылка ]
![](https://i.ytimg.com/vi/CV7LiJ_1vAA/maxresdefault.jpg)