Commit History
Fix Lora config error for Llama3 (#1659)
		230e0ac
	
		
		unverified
	Generalizing the chat_template prompt strategy (#1660) [skip ci]
		cc11c6b
	
		
		unverified
	
		Keith Stevens
		
	commited on
		
		
Fix Google Colab notebook 2024-05 (#1662) [skip ci]
		5f91064
	
		
		unverified
	
		Maciek
		
	commited on
		
		
Update tiny-llama qlora.yml addressing eval packing error (#1638)
		84bb806
	
		
		unverified
	
		Jaydeep Thik
		
	commited on
		
		
update outputs path so that we can mount workspace to /workspace/data (#1623)
		4fde300
	
		
		unverified
	Add ORPO example and e2e test (#1572)
		98c25e1
	
		
		unverified
	fix(yml): update llama-3 config (#1543) [skip ci]
		0e8f340
	
		
		unverified
	fix broken linting (#1541)
		c10563c
	
		
		unverified
	Adding Llama-3 qlora (#1536)
		37c037c
	
		
		unverified
	llama-3 examples (#1537)
		15f7910
	
		
		unverified
	Create mixtral_22.yml (#1514) [skip ci]
		0eadfc8
	
		
		unverified
	
		Atlas
		
	commited on
		
		
DBRX Model Support (#1462)
		132eb74
	
		
		unverified
	Fix the wrong adapter in qwen2-moe-qlora example (#1501) [skip ci]
		7f17eff
	
		
		unverified
	qwen2_moe support w multipack (#1455)
		6086be8
	
		
		unverified
	fix some of the edge cases for Jamba (#1452)
		05b398a
	
		
		unverified
	Jamba (#1451)
		02af082
	
		
		unverified
	turn sample_packing on for training (#1438) [skip ci]
		c19d060
	
		
		unverified
	chore(config): refactor old mistral config (#1435)
		f1ebaa0
	
		
		unverified
	strip out hacky qlora-fsdp workarounds now that qlora-fsdp fixes are upstreamed (#1428)
		2a1589f
	
		
		unverified
	Fix Gemma 7b qlora.yml (#1405)
		6366b0c
	
		
		unverified
	Train parameters exclusively in specific ranges (#1390)
		05bcc9e
	
		
		unverified
	FDSP + QLoRA (#1378)
		9b6ee83
	
		
		unverified
	Update tinyllama lora.yml to fix eval packing issue (#1362)
		8984bf1
	
		
		unverified
	chore: enable sample_packing for Gemma (#1351)
		170d4d7
	
		
		unverified
	Add StableLM 2 Example Scripts (#1327) [skip ci]
		f30d062
	
		
		unverified
	multipack for gemma (#1313)
		2752d5f
	
		
		unverified
	Adding Google's gemma Model (#1312)
		9e300ac
	
		
		unverified
	Add instructions for playing with qlora model to colab example (#1290)
		6ab69ec
	
		
		unverified
	fix(examples): remove is_*_derived as it's parsed automatically (#1297)
		a7a9a14
	
		
		unverified
	Add seq2seq eval benchmark callback (#1274)
		5a5d474
	
		
		unverified
	Add MPS support (#1264)
		fac2d98
	
		
		unverified
	lock pytorch (#1247) [skip ci]
		1c7ed26
	
		
		unverified
	
		JohanWork
		
	commited on
		
		
Pretrain transforms (#1261)
		c7cf381
	
		
		unverified
	Peft lotfq (#1222)
		4cb7900
	
		
		unverified
	Update qlora.yml - remove `max_packed_sequence_len` (#1210) [skip ci]
		5407ddd
	
		
		unverified
	add colab example (#1196) [skip ci]
		ee0b5f6
	
		
		unverified
	
		JohanWork
		
	commited on