QLoRA memory for common model sizes:
- B model: ~-GB VRAM (fits RTX 3080/4080)
- B model: ~-GB VRAM (fits RTX 3090/4090)
- B model: ~-GB VRAM (fits A100 GB)
These estimates include adapter training overhead. Actual usage depends on batch size, sequence length, and gradient checkpointing settings.