NanoCode012
52cde69288
Fix config path after config moved
2023-06-12 17:06:15 +09:00
Wing Lian
9a58e99e81
config fixes
2023-06-12 01:52:58 -04:00
Wing Lian
c7dee56b87
add typehints
2023-06-11 19:52:34 -04:00
Wing Lian
aac4b7691e
add new sharegpt, refactor prompt so it can be customized later, add exception if no data is processed
2023-06-11 19:42:25 -04:00
NanoCode012
f31a338cbb
Merge pull request #191 from OpenAccess-AI-Collective/NanoCode012-patch-1
...
Add save_steps and eval_steps to Readme
2023-06-12 02:55:37 +09:00
NanoCode012
4cd1deeef2
Add save_steps and eval_steps to Readme
2023-06-12 02:44:46 +09:00
Wing Lian
9ac16ed8d1
Merge pull request #190 from OpenAccess-AI-Collective/fixes-20230711-v2
...
more config pruning and migrating
2023-06-11 13:27:08 -04:00
Wing Lian
6b3f509d9e
forgot to add this file
2023-06-11 11:50:12 -04:00
Wing Lian
336aa3fd48
gptq lora llama is obviously good
2023-06-11 11:05:29 -04:00
Wing Lian
d0d7eaa4f3
update openllama and clean up paths
2023-06-11 11:03:31 -04:00
Wing Lian
a6ebf57e82
fix table formatting
2023-06-11 10:55:32 -04:00
Wing Lian
280832cec2
more matrix updates
2023-06-11 10:52:36 -04:00
Wing Lian
a43bae9ff0
update the support matrix
2023-06-11 10:44:03 -04:00
Wing Lian
effbbf6dd1
more pruning
2023-06-11 10:38:24 -04:00
Wing Lian
c9a149f9e8
add check for attr
2023-06-11 10:11:17 -04:00
Wing Lian
c530e4b9c8
more config pruning and migrating
2023-06-11 10:09:05 -04:00
Wing Lian
f620706776
Merge pull request #189 from OpenAccess-AI-Collective/fixes-20230711
...
various fixes
2023-06-11 09:49:23 -04:00
Wing Lian
77762a5d6b
get rid of some configs, formalize pythioa lora config
2023-06-11 09:41:41 -04:00
Wing Lian
14668fa54e
new validation for mpt w grad checkpoints
2023-06-11 09:26:10 -04:00
AngainorDev
b565ecf0a1
Fix strict and Lint
2023-06-11 15:23:38 +02:00
Wing Lian
fe0b76854e
match up gradient checkpointing when using lora w config
2023-06-11 09:20:40 -04:00
NanoCode012
e944311442
Merge pull request #186 from akj2018/main
...
Update FAQS.md
2023-06-11 19:45:06 +09:00
Akshay Jain
e3e7b52a5b
Update FAQS.md
...
Converted (```) to single backtick (') uniformly.
2023-06-10 23:36:14 -07:00
NanoCode012
974dc00a7d
Fix set mem_id for inference and refactor
2023-06-11 14:00:54 +09:00
NanoCode012
572d1141e6
Set mem cache args on inference
2023-06-11 12:05:37 +09:00
NanoCode012
a6190c8094
Clean up landmark patching
2023-06-11 11:59:03 +09:00
NanoCode012
563b6d89e6
Fix undefined LlamaForCausalLM and del try except
2023-06-11 11:58:31 +09:00
Wing Lian
cd0a6f6027
peft no longer needs device_map
2023-06-10 22:50:09 -04:00
Akshay Jain
0e664a5ebc
Update FAQS.md
...
Co-authored-by: NanoCode012 <kevinvong@rocketmail.com >
2023-06-10 19:26:12 -07:00
Akshay Jain
dd7d16d2eb
Update FAQS.md
...
Updated FAQS.md with backticks around error message
2023-06-10 19:15:50 -07:00
NanoCode012
e285e24f7f
Address PR suggestion
...
Co-authored-by: Wing Lian <wing.lian@gmail.com >
2023-06-11 10:52:12 +09:00
NanoCode012
919727b4d7
Refactor landmark attention patch
2023-06-11 10:51:05 +09:00
Akshay Jain
5ffefee37f
Update FAQS.md
...
Update FAQS.md with the following statement
Error invalid argument at line 359 in file /workspace/bitsandbytes/csrc/pythonInterface.c
/arrow/cpp/src/arrow/filesystem/s3fs.cc:2598: arrow::fs::FinalizeS3 was not called even though S3 was initialized. This could lead to a segmentation fault at exit
try reinstalling bitsandbytes and transformers from source
2023-06-10 18:34:54 -07:00
Wing Lian
d9f713e4e3
Merge pull request #183 from OpenAccess-AI-Collective/inference-from-stdin
...
pass a prompt in from stdin for inference
2023-06-10 17:06:55 -04:00
Wing Lian
958da70376
fix formatting
2023-06-10 15:28:08 -04:00
Wing Lian
c4e4f8115c
pass a prompt in from stdin for inference
2023-06-10 15:07:40 -04:00
Angainor Development
a808bf913f
Fix missing cfg.
2023-06-10 20:28:49 +02:00
Wing Lian
01248253a3
Merge pull request #182 from OpenAccess-AI-Collective/fix-llama-ref
...
fix for local variable 'LlamaForCausalLM' referenced before assignment
2023-06-10 14:25:51 -04:00
Wing Lian
759e8673ce
Update scripts/finetune.py
...
Co-authored-by: NanoCode012 <kevinvong@rocketmail.com >
2023-06-10 14:25:21 -04:00
Wing Lian
0c6f928601
address PR feedback
2023-06-10 14:23:56 -04:00
Wing Lian
eea2731a5e
add streaming dataset support for pretraining datasets
2023-06-10 14:23:56 -04:00
Wing Lian
1db46a9c72
linting fix
2023-06-10 14:23:56 -04:00
Wing Lian
ab5cd28acf
more gpt-neox long ctx fixes
2023-06-10 14:23:55 -04:00
Wing Lian
1a82082e91
fix bettertransformers save, force it to skip after saving correctly in callback
2023-06-10 14:23:55 -04:00
Wing Lian
1210dc8fd5
more tweaks to do pre-training with bettertransformers
2023-06-10 14:23:55 -04:00
Wing Lian
488a67d75a
experimental expansion of ctx len
2023-06-10 14:23:53 -04:00
Wing Lian
71a43f8479
add validation/warning for bettertransformers and torch version
2023-06-10 14:22:31 -04:00
Wing Lian
39619028a3
use pythia-12b, neox-20b is flaky
2023-06-10 14:22:30 -04:00
Wing Lian
8792199799
add flash attn context for efficient training and attempt setting model to train mode:
2023-06-10 14:22:30 -04:00
Wing Lian
1edc30c786
add support for opimum bettertransformers
2023-06-10 14:22:30 -04:00