Many industries are rapidly adopting machine learning (ML) to gain insights on the ever increasing data from billions of connected devices. This, combined with a demand for low latency, drives a growing push to move inferencing hardware closer to the location where the data is created. This white paper will describe why FPGA-based hardware accelerators are required in order to eliminate network dependencies, significantly increase performance, and reduce the latency of the ML application.

