Commit History
add lion-pytorch optimizer (#1299) [skip ci]
		1648279
	
		
		unverified
	make mlflow optional (#1317)
		5894f0e
	
		
		unverified
	Allow load_best_model_at_end to be configured for early stopping on custom evaluation datasets (#1291)
		3c00f40
	
		
		unverified
	
		David Meikle
		
	commited on
		
		
Add seq2seq eval benchmark callback (#1274)
		5a5d474
	
		
		unverified
	Scheduler implementation of Continual Pre-Training of Large Language Models: How to (re)warm your model?  (#1273)
		8430db2
	
		
		unverified
	allow the optimizer prune ratio for ReLoRA to be configurable (#1287)
		4b997c3
	
		
		unverified
	simplify haldning for newer multipack patches so they can be added in a single place (#1270)
		5698943
	
		
		unverified
	Add more save strategies for DPO training. (#1255)
		13eea21
	
		
		unverified
	
		Philip May
		
	commited on
		
		
relora: magnitude pruning of the optimizer (#1245)
		8c2e05a
	
		
		unverified
	support for true batches with multipack (#1230)
		00568c1
	
		
		unverified
	Fix and document test_datasets (#1228)
		5787e1a
	
		
		unverified
	FEAT: add tagging support to axolotl for DPOTrainer (#1209)
		18f8119
	
		
		unverified
	precompute dpo logprobs setting and fixes (#1199) [skip ci]
		33e1170
	
		
		unverified
	more dpo fixes for dataset loading and docs (#1185) [skip ci]
		5bce45f
	
		
		unverified
	DPO fixes v2 (#1174)
		59a31fe
	
		
		unverified
	Phi2 multipack (#1173)
		814aee6
	
		
		unverified
	Add mlflow callback for pushing config to mlflow artifacts (#1125)
		b8e5603
	
		
		unverified
	
		JohanWork
		
	commited on
		
		
jupyter lab fixes  (#1139) [skip ci]
		eaaeefc
	
		
		unverified
	Qwen2 (#1166)
		f5a828a
	
		
		unverified
	Multipack simplify for Mixtral (#1142)
		6910e6a
	
		
		unverified
	swap the data collator for evals if not using sample packing (#1076)
		ead34c5
	
		
		unverified
	paired kto support (#1069)
		d7057cc
	
		
		unverified
	Add: mlflow for experiment tracking (#1059) [skip ci]
		090c24d
	
		
		unverified
	Phi2 rewrite (#1058)
		732851f
	
		
		unverified
	streaming multipack for pretraining dataset (#959)
		553c80f
	
		
		unverified
	feat: always push checkpoint to hub if set (#1049) [skip ci]
		cbdbf9e
	
		
		unverified
	RL/DPO (#935)
		f243c21
	
		
		
	use recommended setting for use_reentrant w gradient checkpointing (#1021)
		4d2e842
	
		
		unverified
	remove landmark attn and xpos rope implementations (#1010)
		70b46ca
	
		
		unverified
	FEAT: add tagging support to axolotl (#1004)
		db9094d
	
		
		unverified
	fix: add lr scheduler kwargs to Trainer (#972)
		13e9381
	
		
		unverified
	fix: switch to using the HuggingFace Transformers NEFT implementation (#941)
		ef24342
	
		
		unverified
	
		kallewoof
		
	commited on
		
		
 
		 
		