Wing Lian
d0d7eaa4f3
update openllama and clean up paths
2023-06-11 11:03:31 -04:00
Wing Lian
a6ebf57e82
fix table formatting
2023-06-11 10:55:32 -04:00
Wing Lian
280832cec2
more matrix updates
2023-06-11 10:52:36 -04:00
Wing Lian
a43bae9ff0
update the support matrix
2023-06-11 10:44:03 -04:00
Wing Lian
effbbf6dd1
more pruning
2023-06-11 10:38:24 -04:00
Wing Lian
c9a149f9e8
add check for attr
2023-06-11 10:11:17 -04:00
Wing Lian
c530e4b9c8
more config pruning and migrating
2023-06-11 10:09:05 -04:00
Wing Lian
f620706776
Merge pull request #189 from OpenAccess-AI-Collective/fixes-20230711
...
various fixes
2023-06-11 09:49:23 -04:00
Wing Lian
77762a5d6b
get rid of some configs, formalize pythioa lora config
2023-06-11 09:41:41 -04:00
Wing Lian
14668fa54e
new validation for mpt w grad checkpoints
2023-06-11 09:26:10 -04:00
AngainorDev
b565ecf0a1
Fix strict and Lint
2023-06-11 15:23:38 +02:00
Wing Lian
fe0b76854e
match up gradient checkpointing when using lora w config
2023-06-11 09:20:40 -04:00
NanoCode012
e944311442
Merge pull request #186 from akj2018/main
...
Update FAQS.md
2023-06-11 19:45:06 +09:00
Akshay Jain
e3e7b52a5b
Update FAQS.md
...
Converted (```) to single backtick (') uniformly.
2023-06-10 23:36:14 -07:00
NanoCode012
974dc00a7d
Fix set mem_id for inference and refactor
2023-06-11 14:00:54 +09:00
NanoCode012
572d1141e6
Set mem cache args on inference
2023-06-11 12:05:37 +09:00
NanoCode012
a6190c8094
Clean up landmark patching
2023-06-11 11:59:03 +09:00
NanoCode012
563b6d89e6
Fix undefined LlamaForCausalLM and del try except
2023-06-11 11:58:31 +09:00
Wing Lian
cd0a6f6027
peft no longer needs device_map
2023-06-10 22:50:09 -04:00
Akshay Jain
0e664a5ebc
Update FAQS.md
...
Co-authored-by: NanoCode012 <kevinvong@rocketmail.com >
2023-06-10 19:26:12 -07:00
Akshay Jain
dd7d16d2eb
Update FAQS.md
...
Updated FAQS.md with backticks around error message
2023-06-10 19:15:50 -07:00
NanoCode012
e285e24f7f
Address PR suggestion
...
Co-authored-by: Wing Lian <wing.lian@gmail.com >
2023-06-11 10:52:12 +09:00
NanoCode012
919727b4d7
Refactor landmark attention patch
2023-06-11 10:51:05 +09:00
Akshay Jain
5ffefee37f
Update FAQS.md
...
Update FAQS.md with the following statement
Error invalid argument at line 359 in file /workspace/bitsandbytes/csrc/pythonInterface.c
/arrow/cpp/src/arrow/filesystem/s3fs.cc:2598: arrow::fs::FinalizeS3 was not called even though S3 was initialized. This could lead to a segmentation fault at exit
try reinstalling bitsandbytes and transformers from source
2023-06-10 18:34:54 -07:00
Wing Lian
d9f713e4e3
Merge pull request #183 from OpenAccess-AI-Collective/inference-from-stdin
...
pass a prompt in from stdin for inference
2023-06-10 17:06:55 -04:00
Wing Lian
958da70376
fix formatting
2023-06-10 15:28:08 -04:00
Wing Lian
c4e4f8115c
pass a prompt in from stdin for inference
2023-06-10 15:07:40 -04:00
Angainor Development
a808bf913f
Fix missing cfg.
2023-06-10 20:28:49 +02:00
Wing Lian
01248253a3
Merge pull request #182 from OpenAccess-AI-Collective/fix-llama-ref
...
fix for local variable 'LlamaForCausalLM' referenced before assignment
2023-06-10 14:25:51 -04:00
Wing Lian
759e8673ce
Update scripts/finetune.py
...
Co-authored-by: NanoCode012 <kevinvong@rocketmail.com >
2023-06-10 14:25:21 -04:00
Wing Lian
0c6f928601
address PR feedback
2023-06-10 14:23:56 -04:00
Wing Lian
eea2731a5e
add streaming dataset support for pretraining datasets
2023-06-10 14:23:56 -04:00
Wing Lian
1db46a9c72
linting fix
2023-06-10 14:23:56 -04:00
Wing Lian
ab5cd28acf
more gpt-neox long ctx fixes
2023-06-10 14:23:55 -04:00
Wing Lian
1a82082e91
fix bettertransformers save, force it to skip after saving correctly in callback
2023-06-10 14:23:55 -04:00
Wing Lian
1210dc8fd5
more tweaks to do pre-training with bettertransformers
2023-06-10 14:23:55 -04:00
Wing Lian
488a67d75a
experimental expansion of ctx len
2023-06-10 14:23:53 -04:00
Wing Lian
71a43f8479
add validation/warning for bettertransformers and torch version
2023-06-10 14:22:31 -04:00
Wing Lian
39619028a3
use pythia-12b, neox-20b is flaky
2023-06-10 14:22:30 -04:00
Wing Lian
8792199799
add flash attn context for efficient training and attempt setting model to train mode:
2023-06-10 14:22:30 -04:00
Wing Lian
1edc30c786
add support for opimum bettertransformers
2023-06-10 14:22:30 -04:00
Wing Lian
14163c15d9
fix for local variable 'LlamaForCausalLM' referenced before assignment
2023-06-10 14:11:13 -04:00
Wing Lian
41e4f6ca31
Merge pull request #181 from OpenAccess-AI-Collective/xpos-rope
...
add support to extend context with xpos rope
2023-06-10 14:04:03 -04:00
Angainor Development
79e2a6f140
Merge branch 'main' into patch-1
2023-06-10 19:07:54 +02:00
Angainor Development
c2508987a6
Remove explicit definition of cfg.inference
2023-06-10 19:06:10 +02:00
Wing Lian
215d775147
Merge pull request #180 from Glavin001/feat/stream-inference
...
Add streaming inference & fix stopping at EOS
2023-06-10 12:04:34 -04:00
Wing Lian
f36e227eaf
formatting for linter
2023-06-10 12:00:52 -04:00
Wing Lian
5878bb1f3a
add option to readme
2023-06-10 11:57:41 -04:00
Wing Lian
a03a7d7d8b
add support to extend context with xpos rope
2023-06-10 10:29:46 -04:00
Glavin Wiechert
fec6bcc3e6
Add streaming inference & fix stopping at EOS
2023-06-10 08:14:47 +00:00