Commit History
always drop samples that are too long (#452)
50682a3
unverified
winglian
commited on
set env var for FSDP layer to wrap (#453)
5a1985b
unverified
winglian
commited on
add missing positional arg (#450)
58cf7e7
unverified
winglian
commited on
fix evals (#447)
ee26281
unverified
winglian
commited on
disable eval using multipack for now (#437)
f733d0f
unverified
winglian
commited on
fix comma, not a tuple (#436)
008505c
unverified
winglian
commited on
use save_strategy from config if available (#434)
b3f5e00
unverified
winglian
commited on
set env for FSDP offload params (#433)
5247c50
unverified
winglian
commited on
Fix(config): Update handling of deepspeed config (#404)
c01015f
unverified
Nanobit
commited on
fix eval steps and strategy (#403)
da10af0
unverified
winglian
commited on
Feat(config): add max steps (#387)
3c2ad00
unverified
ittailup
commited on
Added "epoch" evaluation_strategy (#388)
5d48a10
unverified
flotos
commited on
Feat(config): Add hub_strategy (#386)
73a0b6e
unverified
Nanobit
commited on
improve GPU logging to break out pytorch cache and system mem
7b55fe6
tmm1
commited on
Attention mask and position id fixes for packing (#285)
2bb0b78
unverified
winglian
commited on
log GPU memory usage
e303d64
tmm1
commited on
fix axolotl training args dataclass annotation
ebaec3c
winglian
commited on
Merge branch 'OpenAccess-AI-Collective:main' into logging_enhancement
83237b8
unverified
The Objective Dad
commited on
Merge pull request #274 from OpenAccess-AI-Collective/NanoCode012-patch-2
168a7a0
unverified
Nanobit
commited on
Adding logging enhancement
553a86b
theobjectivedad
commited on
Feat: Add save_safetensors
5491278
Nanobit
commited on
Set push to hub as private by default
1514739
unverified
Nanobit
commited on
Merge branch 'main' into quadratic-warmup
c4cf567
unverified
winglian
commited on
better configuration for quadratic warmup
c49729d
winglian
commited on
Fix future deprecation push_to_hub_model_id
e79c8e6
Nanobit
commited on
push intermediate model checkpoints to hub
612aabd
winglian
commited on
support adamw and grad norm hyperparams
6d0ee4b
winglian
commited on
add axolotl trainer and quadratic warmup
7dc580b
winglian
commited on
Merge branch 'main' into flash-optimum
fd2c981
unverified
winglian
commited on
Fix set mem_id for inference and refactor
974dc00
Nanobit
commited on
fix formatting
958da70
winglian
commited on
address PR feedback
0c6f928
winglian
commited on
fix bettertransformers save, force it to skip after saving correctly in callback
1a82082
winglian
commited on
more tweaks to do pre-training with bettertransformers
1210dc8
winglian
commited on
Feat: Add landmark attention
55b8542
Nanobit
commited on
Refactor out unmodified save_steps and eval_steps
2ef4634
Nanobit
commited on
Set to use cfg.seed or 42 for backward compat
2cfe9e9
Nanobit
commited on
fix relative path for fixtures
cfcc549
winglian
commited on
Apply isort then black
37293dc
Nanobit
commited on
Fix mypy typing
e9650d3
Nanobit
commited on
Lint trainer.py
ddb86ea
Nanobit
commited on
fix relative path for fixtures
e65aeed
winglian
commited on
refactor(param): rename load_4bit config param by gptq
dd00657
Thytu
commited on
fixes to make qlora actually work
34c99f9
winglian
commited on
apply black formatting
ce34d64
winglian
commited on
fix missing fp16 kwarg
2ae936f
winglian
commited on
Add qa style data for alpaca instructions, fix one_cycle scheduler
3a50377
winglian
commited on
don't need to set here
de6da13
winglian
commited on
be able to use adam bnb 8bit and one cycle scheduler w fsdp
9493b1b
winglian
commited on