www.design-reuse-china.com
搜索,选择,比较,与提供商进行安全高效的联系
Design & Reuse We Chat
D&R中国官方微信公众号,
关注获取最新IP SOC业界资讯

DynapCNN - the World's First 1M Neuron, Event-Driven Neuromorphic AI Processor for Vision Processing

April 12, 2019 -- Today aiCTX AG is announcing its new fully-asynchronous event-driven neuromorphic AI processor for ultra-low power, always-on, real-time applications.

DynapCNN opens brand-new possibilities for dynamic vision processing, bringing event-based vision applications to power-constrained devices for the first time.

DynapCNN is a 12mm2 chip, fabricated in 22nm technology, housing over 1 million spiking neurons and 4 million programmable parameters, with a scalable architecture optimally suited for implementing Convolutional Neural Networks. It is a first of its kind ASIC that brings the power of machine learning and the efficiency of event-driven neuromorphic computation together in one device. DynapCNN is the most direct and power-efficient way of processing data generated by Event-Based and Dynamic Vision Sensors.

As a next-generation vision processing solution, DynapCNN is 100–1000 times more power efficient than the state of the art, and delivers 10 times shorter latencies in real-time vision processing.

Those savings in energy mean that applications based on DynapCNN can be always-on, and crunch data locally on battery powered, portable devices.

Computation in DynapCNN is triggered directly by changes in the visual scene, without using a high-speed clock. Moving objects give rise to sequences of events, which are processed immediately by the processor. Since there is no notion of frames, DynapCNN’s continuous computation enables ultra-low-latency of below 5ms. This represents at least a 10x improvement from the current deep learning solutions available in the market for real-time vision processing.

DynapCNN Development Kits will be available in Q3 2019.

 Back

业务合作

添加产品

供应商免费录入产品信息

点击此处了解更多关于D&R的隐私政策

© 2026 Design And Reuse

版权所有

本网站的任何部分未经Design&Reuse许可,
不得复制,重发, 转载或以其他方式使用。