https://hgpu.org/?p=29113
QUICK: Quantization-aware Interleaving and Conflict-free Kernel for efficient LLM inference