Commit History
support to disable exllama for gptq (#604)
		faecff9
	
		
		unverified
	Delete duplicate lines (#606)
		aa656e0
	
		
		unverified
	btlm and falcon monkey patches for flash attn (#566)
		6b9b229
	
		
		unverified
	make phi training work with Loras (#588)
		62eaee7
	
		
		unverified
	don't resize embeddings if it's already large enough (#577)
		3607882
	
		
		unverified
	Support Sample packing for phi arch (#586)
		12a2dbb
	
		
		unverified
	Add training callback to send predictions to WandB table (#521)
		5b67ea9
	
		
		unverified
	fix for quant config from model (#540)
		a94f9cb
	
		
		unverified
	Add support for GPTQ using native transformers/peft (#468)
		3355706
	
		
		unverified
	fix: bad dtype for full finetune (#504)
		1991946
	
		
		unverified
	Refactor train cfg cli (#499)
		125cccb
	
		
		unverified
	simplify linear layer locator
		267b7b2
	
		
		
	fsdp requires params be the same type too (#493)
		98bf76e
	
		
		unverified
	Fix(tokenizer): Make sure to add pad for CodeLlamaTokenizer (#489)
		4c37bd0
	
		
		unverified
	fix condition and add logging
		3a011ea
	
		
		
	rename var and reformat
		f319b0b
	
		
		
	Update src/axolotl/utils/models.py
		7fd662d
	
		
		unverified
	Update src/axolotl/utils/models.py
		9e69968
	
		
		unverified
	ignore: address pr review
		d03887f
	
		
		unverified
	
		Maxime
		
	commited on
		
		
ignore: linter
		a184549
	
		
		unverified
	
		Maxime
		
	commited on
		
		
fix: finetune model inference needs the dtype fix to work with flash-attn
		f311df9
	
		
		unverified
	
		Maxime
		
	commited on
		
		
 
		 
		