Commit History
Attention mask and position id fixes for packing (#285)
2bb0b78
unverified
log GPU memory usage
e303d64
fix axolotl training args dataclass annotation
ebaec3c
Merge branch 'OpenAccess-AI-Collective:main' into logging_enhancement
83237b8
unverified
The Objective Dad
commited on
Merge pull request #274 from OpenAccess-AI-Collective/NanoCode012-patch-2
168a7a0
unverified
Adding logging enhancement
553a86b
Feat: Add save_safetensors
5491278
Set push to hub as private by default
1514739
unverified
Merge branch 'main' into quadratic-warmup
c4cf567
unverified
better configuration for quadratic warmup
c49729d
Fix future deprecation push_to_hub_model_id
e79c8e6
push intermediate model checkpoints to hub
612aabd
support adamw and grad norm hyperparams
6d0ee4b
add axolotl trainer and quadratic warmup
7dc580b
Merge branch 'main' into flash-optimum
fd2c981
unverified
Fix set mem_id for inference and refactor
974dc00
fix formatting
958da70
address PR feedback
0c6f928
fix bettertransformers save, force it to skip after saving correctly in callback
1a82082
more tweaks to do pre-training with bettertransformers
1210dc8
Feat: Add landmark attention
55b8542
Refactor out unmodified save_steps and eval_steps
2ef4634
Set to use cfg.seed or 42 for backward compat
2cfe9e9
fix relative path for fixtures
cfcc549
Apply isort then black
37293dc
Fix mypy typing
e9650d3
Lint trainer.py
ddb86ea
fix relative path for fixtures
e65aeed
refactor(param): rename load_4bit config param by gptq
dd00657
Thytu
commited on