Skip to content

Qwen QLoRA trainFine-tune Qwen3 and Qwen3.5 on limited GPU memory

Practical QLoRA workflow for smaller setups. Focused on clear dataset handling, controlled truncation, and reproducible training/inference steps.

Part of a two-repo ecosystem

RepoPurpose
qwen35-toolkitModel prep — BNB quantization, visual tower strip, verify, upload
qwen-qlora-train (this repo)LoRA training, adapter inference, CPU merge

⚠️ Validated training on RTX 3070 8 GB currently covers Qwen3 1.7B and 4B (see Quickstart). Qwen3 8B OOMs on unsloth 2026.3.4+, and sizes above 4B should be treated as experimental on this hardware class.

Released under the Apache 2.0 License.