Commit History
support for loading a model by git revision
69a2350
skip explicit model type too if using trust_remote_code
d69da99
don't use llama if trust_remote_code is set since that needs to use AutoModel path
66afb76
optionally define whether to use_fast tokenizer
47d601f
add float16 docs and tweak typehints
88e17ff
style correction
136522f
maciej.karasek
commited on
issue #205 bugfix
556fe40
maciej.karasek
commited on
Merge branch 'main' into flash-optimum
fd2c981
unverified
Merge pull request #187 from OpenAccess-AI-Collective/strip-peft-device-map
93dacba
unverified
Merge pull request #177 from NanoCode012/fix/landmark-patch
8002ffb
unverified
Merge branch 'main' into strip-peft-device-map
5e616d9
unverified
Merge pull request #159 from AngainorDev/patch-1
8e568bb
unverified
add check for attr
c9a149f
Fix strict and Lint
b565ecf
match up gradient checkpointing when using lora w config
fe0b768
Fix undefined LlamaForCausalLM and del try except
563b6d8
peft no longer needs device_map
cd0a6f6
Refactor landmark attention patch
919727b
Fix missing cfg.
a808bf9
unverified
Angainor Development
commited on
Merge pull request #182 from OpenAccess-AI-Collective/fix-llama-ref
0124825
unverified
more gpt-neox long ctx fixes
ab5cd28
more tweaks to do pre-training with bettertransformers
1210dc8
add support for opimum bettertransformers
1edc30c
fix for local variable 'LlamaForCausalLM' referenced before assignment
14163c1
Merge branch 'main' into patch-1
79e2a6f
unverified
Angainor Development
commited on
add support to extend context with xpos rope
a03a7d7
fix for max sequence len across different model types
7f09106
Fix backward compat for peft
aefb2fc
WIP: Rely on cfg.inference
813cfa4
unverified
Angainor Development
commited on
Fix patching via import instead of hijacking
e44c9e0
Feat: Add landmark attention
55b8542
Fix future deprecate prepare_model_for_int8_training
df9528f
Fix training over existing lora
193c73b
unverified
Angainor Development
commited on