I'm interested in knowing if TensorRT currently supports quantization for 2-bit and 4-bit models. Could you confirm the same for mixed precision as well? Additionally, I'd appreciate insights on accuracy metrics and any potential challenges associated with achieving accurate inference at such low bit-widths.