|
||
![]() www.design-reuse-embedded.com |
InferX X1 Edge Inference Co-Processor
|
|
Overview High Throughput, Low Cost, Low Power
The InferX X1 Edge Inference Co-Processor is optimized for what the edge needs: large models and large models at batch=1. InferX X1 offers throughput close to data center boards that sell for thousands of dollars but does so at single digit watts and at a fraction of the price. InferX X1 is programmed using TensorFlow Lite and ONNX: a performance modeler is available now. InferX X1 is based on our nnMAX architecture integrating 4 tiles for 4K MACs and 8MB L2 SRAM. InferX X1 connects to a single x32 LPDDR4 DRAM. Four lanes PCIe Gen3 connect to the host processor; a GPIO link is available for hosts without PCIe. Two X1's can work together to double throughput.
Please sign in to view full IP description :
|
Partner with us |
List your ProductsSuppliers, list and add your products for free. |
More about D&R Privacy Policy© 2023 Design And Reuse All Rights Reserved. No portion of this site may be copied, retransmitted, reposted, duplicated or otherwise used without the express written permission of Design And Reuse. |
||||||