February 25, 2019
ICP Electronics Australia is pleased to present iEi Integration’s new Mustang-V100-MX8 computing accelerator card.
The Mustang-V100-MX8 is designed to handle artificial intelligence deep learning inference while still retaining power efficiency and a compact form factor. The card is half-height, half-length, and single-slot, resulting in a small-scale, condensed size perfect for space-conscious applications that require AI acceleration and machine vision.
As a VPU PCIe it can run artificial intelligence programs faster, and is well-suited for low power consumption applications such as surveillance, retail, and transportation. With the advantage of both power efficiency and high performance to dedicate to DNN topologies, it implements perfectly in AI edge computing devices to reduce total power usage and provide longer duty time for rechargeable edge computing equipment.
Finally, it supports Intel’s OpenVINO™ toolkit for the optimisation of pre-trained deep-learning models such as Caffe, MXNET, and Tensorflow; after this optimisation it can execute the inference engine across all kinds of Intel® hardware. As it features eight Intel® Movidius™ Myriad™ X VPUs, the Mustang-V100-MX8 is capable of executing multiple topologies at once.
- Half-Height, Half-Length, Single-Slot Compact Size
- Low Power Consumption; Approx. 2.5W per Intel® Movidius™ Myriad™ X VPU
- Supported Open Visual Inference & Neural Network Optimisation (OpenVINO™) Toolkit, AI Edge Computing Ready Device
- 8x Intel® Movidius™ Myriad™ X VPUs to Execute Multiple Topologies Simultaneously