Real-time in situ image analytics impose stringent latency requirements on intelligent neural network inference operations. While conventional software-based implementations on the graphic processing unit (GPU)-accelerated platforms are flexible and have achieved very high inference throughput, they are not suitable for latency-sensitive applications where real-time feedback is needed. Here, we demonstrate that high-performance reconfigurable computing platforms based on field-programmable gate array (FPGA) processing can successfully bridge the gap between low-level hardware processing and high-level intelligent image analytics algorithm deployment within a unified system. The proposed design performs inference operations on a stream of individual images as they are produced and has a deeply pipelined hardware design that allows all layers of a quantized convolutional neural network (QCNN) to compute concurrently with partial image inputs. Using the case of label-free classification of human peripheral blood mononuclear cell (PBMC) subtypes as a proof-of-concept illustration, our system achieves an ultralow classification latency of 34.2 $\mu \text{s}$ with over 95% end-to-end accuracy by using a QCNN, while the cells are imaged at throughput exceeding 29 200 cells/s. Our QCNN design is modular and is readily adaptable to other QCNNs with different latency and resource requirements.
CITATION STYLE
Wang, M., Lee, K. C. M., Chung, B. M. F., Bogaraju, S. V., Ng, H. C., Wong, J. S. J., … So, H. K. H. (2022). Low-Latency In Situ Image Analytics With FPGA-Based Quantized Convolutional Neural Network. IEEE Transactions on Neural Networks and Learning Systems, 33(7), 2853–2866. https://doi.org/10.1109/TNNLS.2020.3046452
Mendeley helps you to discover research relevant for your work.