- SFT: Use bf16 or fp32 for training; avoid 8bit. For evaluation, fp16, bf16, or fp32 is fine. Follow established scripts for reliability.
- Unsloth: Train LoRA with fp16, bf16, or fp32. Avoid 8bit or lower unless validated through replication of original experiments. No QLoRA unless core setups are stable and everything before this has worked.
Created
November 29, 2024 21:59
-
-
Save brando90/2cfbbd9548d96d60bde9c35d08598698 to your computer and use it in GitHub Desktop.
nothing bellow 16 bits for training
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment