Commit History
Merge pull request #189 from OpenAccess-AI-Collective/fixes-20230711
f620706
unverified
get rid of some configs, formalize pythioa lora config
77762a5
new validation for mpt w grad checkpoints
14668fa
Fix strict and Lint
b565ecf
match up gradient checkpointing when using lora w config
fe0b768
Merge pull request #186 from akj2018/main
e944311
unverified
Update FAQS.md
e3e7b52
unverified
Fix set mem_id for inference and refactor
974dc00
Set mem cache args on inference
572d114
Clean up landmark patching
a6190c8
Fix undefined LlamaForCausalLM and del try except
563b6d8
peft no longer needs device_map
cd0a6f6
Update FAQS.md
dd7d16d
unverified
Refactor landmark attention patch
919727b
Update FAQS.md
5ffefee
unverified
Merge pull request #183 from OpenAccess-AI-Collective/inference-from-stdin
d9f713e
unverified
fix formatting
958da70
pass a prompt in from stdin for inference
c4e4f81
Fix missing cfg.
a808bf9
unverified
Angainor Development
commited on
Merge pull request #182 from OpenAccess-AI-Collective/fix-llama-ref
0124825
unverified
address PR feedback
0c6f928
add streaming dataset support for pretraining datasets
eea2731
linting fix
1db46a9
more gpt-neox long ctx fixes
ab5cd28
fix bettertransformers save, force it to skip after saving correctly in callback
1a82082
more tweaks to do pre-training with bettertransformers
1210dc8
experimental expansion of ctx len
488a67d
add validation/warning for bettertransformers and torch version
71a43f8
use pythia-12b, neox-20b is flaky
3961902
add flash attn context for efficient training and attempt setting model to train mode:
8792199
add support for opimum bettertransformers
1edc30c
fix for local variable 'LlamaForCausalLM' referenced before assignment
14163c1
Merge pull request #181 from OpenAccess-AI-Collective/xpos-rope
41e4f6c
unverified
Merge branch 'main' into patch-1
79e2a6f
unverified
Angainor Development
commited on
Remove explicit definition of cfg.inference
c250898
unverified
Angainor Development
commited on