https://hgpu.org/?p=25799
Optimizing a Hardware Network Stack to Realize an In-Network ML Inference Application