File size: 1,473 Bytes
797ce4a
 
 
 
 
6c57636
 
 
7e670a7
6c57636
 
 
 
 
 
 
 
7e670a7
6c57636
12202a8
 
 
 
6c57636
 
797ce4a
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
---
tags:
- accuracy recovery adapter
- AI Toolkit
---
# Accuracy Recovery Adapters

This repo contains various accuracy recovery adapters (ARAs) that I have trained, primarialy for use with [AI Toolkit](https://github.com/ostris/ai-toolkit).
An ARA is a LoRA that is trained via student teacher training with the student being quantized down to a low precision and the teacher having a high precision.
The goal is to have a side chain LoRA, at bfloat16, that runs parallel to highly quantized layers in a network to compensate for the loss in precision that happens
when weights are quantized. The training is done on a per layer basis in order to match the parent output as much as possible. 

While this can be used on inference, my primary goal is to make large models finetunable on consumer grade hardware. With the 3bit Qwen Image adapter, it
is now possible to train a LoRA on top of it, with 1 MP images, on a 24 GB GPU, such as a 3090/4090. 

I have found the sweet spot, at least for [Qwen-Image](https://huggingface.co/Qwen/Qwen-Image), is 3 bit quantization with a rank 16 adapter.

More info, examples, links, training scripts, AI Toolkit example configs, and adapters to some soon.

## License

All adapters inherit the parent model license. Apache 2.0 for Apache 2.0, BFL License for BFL License, etc.

## Qwen-Image 3 bit quantization

![image/jpeg](https://cdn-uploads.huggingface.co/production/uploads/643cb43e6eeb746f5ad81c26/omdVOGwi3H8P83o8d6nKm.jpeg)