INT4 LoRA high-quality-tuning vs QLoRA: A user inquired about the differences in between INT4 LoRA fine-tuning and QLoRA in terms of precision and speed. Another member explained that QLoRA with HQQ requires frozen quantized weights, will not use tinnygemm, and utilizes dequantizing alongside torch.matmulTweet from Robert Graham (@ErrataRob): nVidi