Find Top SoC Solutions
for AI, Automotive, IoT, Security, Audio & Video...

Edge AI Inferencing Opens Up New World of Opportunities

The ability to do AI inferencing closer to the end user is opening up a whole new world of markets and applications.

www.eetasia.com, May. 28, 2021 – 

While AI originally was targeted for data centers and in the cloud, it has been moving rapidly towards the edge of the network where it is needed to make fast and critical decisions locally and closer to the end user. Sure, training can be still done in the cloud, but in applications such as autonomous driving, it is important that the time-sensitive decision making (spotting a car or pedestrian) is done closer to the end user (the driver). After all, edge systems can make decisions on images coming in at up to 60 frames per second, enabling quick actions.

These systems are made possible through edge inference accelerators that have emerged to replace CPUs, GPUs and FPGAs at much higher throughput/$ and throughput/Watt.

click here to


Partner with us

List your Products

Suppliers, list and add your products for free.

More about D&R Privacy Policy

© 2021 Design And Reuse

All Rights Reserved.

No portion of this site may be copied, retransmitted, reposted, duplicated or otherwise used without the express written permission of Design And Reuse.