It was tough to get this working, but I think I’ve figured it out enough to share. Here’s a quick guide on how to set up LLaMA-Factory with support for Flash Attention 2 and Unsloth training on Windows. This is using a RTX3060 12GB GPU, Windows 10, and CUDA 12.1. Unsloth is an optimization library […]
Read More… from LLaMA-Factory with Flash Attention 2 and Unsloth