|
||
|
InferX X1 Edge Inference Co-Processor
|
|
Overview High Throughput, Low Cost, Low Power
The InferX X1 Edge Inference Co-Processor is optimized for what the edge needs: large models and large models at batch=1. InferX X1 offers throughput close to data center boards that sell for thousands of dollars but does so at single digit watts and at a fraction of the price. InferX X1 is programmed using TensorFlow Lite and ONNX: a performance modeler is available now. InferX X1 is based on our nnMAX architecture integrating 4 tiles for 4K MACs and 8MB L2 SRAM. InferX X1 connects to a single x32 LPDDR4 DRAM. Four lanes PCIe Gen3 connect to the host processor; a GPIO link is available for hosts without PCIe. Two X1's can work together to double throughput.
Please sign in to view full IP description :
|
业务合作 |
广告发布访问我们的广告选项 |
添加产品供应商免费录入产品信息 | ||||||||