Commit History
Create phi3-ft-fsdp.yml (#1580)
a82a711
unverified
Phi-3 conversation format, example training script and perplexity metric (#1582)
cf64284
unverified
fix lint issue that snuck through (#1665)
f5febc7
unverified
Fix Lora config error for Llama3 (#1659)
230e0ac
unverified
Generalizing the chat_template prompt strategy (#1660) [skip ci]
cc11c6b
unverified
Keith Stevens
commited on
Fix Google Colab notebook 2024-05 (#1662) [skip ci]
5f91064
unverified
Maciek
commited on
Update tiny-llama qlora.yml addressing eval packing error (#1638)
84bb806
unverified
Jaydeep Thik
commited on
update outputs path so that we can mount workspace to /workspace/data (#1623)
4fde300
unverified
Add ORPO example and e2e test (#1572)
98c25e1
unverified
fix(yml): update llama-3 config (#1543) [skip ci]
0e8f340
unverified
fix broken linting (#1541)
c10563c
unverified
Adding Llama-3 qlora (#1536)
37c037c
unverified
llama-3 examples (#1537)
15f7910
unverified
Create mixtral_22.yml (#1514) [skip ci]
0eadfc8
unverified
Atlas
commited on
DBRX Model Support (#1462)
132eb74
unverified
Fix the wrong adapter in qwen2-moe-qlora example (#1501) [skip ci]
7f17eff
unverified
qwen2_moe support w multipack (#1455)
6086be8
unverified
fix some of the edge cases for Jamba (#1452)
05b398a
unverified
Jamba (#1451)
02af082
unverified
turn sample_packing on for training (#1438) [skip ci]
c19d060
unverified
chore(config): refactor old mistral config (#1435)
f1ebaa0
unverified
strip out hacky qlora-fsdp workarounds now that qlora-fsdp fixes are upstreamed (#1428)
2a1589f
unverified
Fix Gemma 7b qlora.yml (#1405)
6366b0c
unverified
Train parameters exclusively in specific ranges (#1390)
05bcc9e
unverified
FDSP + QLoRA (#1378)
9b6ee83
unverified
Update tinyllama lora.yml to fix eval packing issue (#1362)
8984bf1
unverified
chore: enable sample_packing for Gemma (#1351)
170d4d7
unverified
Add StableLM 2 Example Scripts (#1327) [skip ci]
f30d062
unverified
multipack for gemma (#1313)
2752d5f
unverified
Adding Google's gemma Model (#1312)
9e300ac
unverified
Add instructions for playing with qlora model to colab example (#1290)
6ab69ec
unverified
fix(examples): remove is_*_derived as it's parsed automatically (#1297)
a7a9a14
unverified
Add seq2seq eval benchmark callback (#1274)
5a5d474
unverified
Add MPS support (#1264)
fac2d98
unverified
lock pytorch (#1247) [skip ci]
1c7ed26
unverified
JohanWork
commited on
Pretrain transforms (#1261)
c7cf381
unverified
Peft lotfq (#1222)
4cb7900
unverified
Update qlora.yml - remove `max_packed_sequence_len` (#1210) [skip ci]
5407ddd
unverified
add colab example (#1196) [skip ci]
ee0b5f6
unverified
JohanWork
commited on