try #2: pin hf transformers and accelerate to latest release, don't reinstall pytorch (#867)
Browse files* isolate torch from the requirements.txt
* fix typo for removed line ending
* pin transformers and accelerate to latest releases
* try w auto-gptq==0.5.1
* update README to remove manual peft install
* pin xformers to 0.0.22
* bump flash-attn to 2.3.3
* pin flash attn to exact version
- .github/workflows/tests.yml +1 -0
- README.md +0 -1
- requirements.txt +5 -7
.github/workflows/tests.yml
CHANGED
|
@@ -71,6 +71,7 @@ jobs:
|
|
| 71 |
|
| 72 |
- name: Install dependencies
|
| 73 |
run: |
|
|
|
|
| 74 |
pip3 uninstall -y transformers accelerate
|
| 75 |
pip3 install -U -e .[flash-attn]
|
| 76 |
pip3 install -r requirements-tests.txt
|
|
|
|
| 71 |
|
| 72 |
- name: Install dependencies
|
| 73 |
run: |
|
| 74 |
+
pip3 install --extra-index-url https://download.pytorch.org/whl/cu118 -U torch==2.0.1
|
| 75 |
pip3 uninstall -y transformers accelerate
|
| 76 |
pip3 install -U -e .[flash-attn]
|
| 77 |
pip3 install -r requirements-tests.txt
|
README.md
CHANGED
|
@@ -91,7 +91,6 @@ cd axolotl
|
|
| 91 |
|
| 92 |
pip3 install packaging
|
| 93 |
pip3 install -e '.[flash-attn,deepspeed]'
|
| 94 |
-
pip3 install -U git+https://github.com/huggingface/peft.git
|
| 95 |
|
| 96 |
# finetune lora
|
| 97 |
accelerate launch -m axolotl.cli.train examples/openllama-3b/lora.yml
|
|
|
|
| 91 |
|
| 92 |
pip3 install packaging
|
| 93 |
pip3 install -e '.[flash-attn,deepspeed]'
|
|
|
|
| 94 |
|
| 95 |
# finetune lora
|
| 96 |
accelerate launch -m axolotl.cli.train examples/openllama-3b/lora.yml
|
requirements.txt
CHANGED
|
@@ -1,22 +1,20 @@
|
|
| 1 |
-
--extra-index-url https://download.pytorch.org/whl/cu118
|
| 2 |
--extra-index-url https://huggingface.github.io/autogptq-index/whl/cu118/
|
| 3 |
-
|
| 4 |
-
auto-gptq==0.4.2
|
| 5 |
packaging
|
| 6 |
peft==0.6.0
|
| 7 |
-
transformers
|
| 8 |
bitsandbytes>=0.41.1
|
| 9 |
-
accelerate
|
| 10 |
deepspeed
|
| 11 |
addict
|
| 12 |
fire
|
| 13 |
PyYAML>=6.0
|
| 14 |
datasets>=2.14.0
|
| 15 |
-
flash-attn
|
| 16 |
sentencepiece
|
| 17 |
wandb
|
| 18 |
einops
|
| 19 |
-
xformers
|
| 20 |
optimum==1.13.2
|
| 21 |
hf_transfer
|
| 22 |
colorama
|
|
|
|
|
|
|
| 1 |
--extra-index-url https://huggingface.github.io/autogptq-index/whl/cu118/
|
| 2 |
+
auto-gptq==0.5.1
|
|
|
|
| 3 |
packaging
|
| 4 |
peft==0.6.0
|
| 5 |
+
transformers==4.35.1
|
| 6 |
bitsandbytes>=0.41.1
|
| 7 |
+
accelerate==0.24.1
|
| 8 |
deepspeed
|
| 9 |
addict
|
| 10 |
fire
|
| 11 |
PyYAML>=6.0
|
| 12 |
datasets>=2.14.0
|
| 13 |
+
flash-attn==2.3.3
|
| 14 |
sentencepiece
|
| 15 |
wandb
|
| 16 |
einops
|
| 17 |
+
xformers==0.0.22
|
| 18 |
optimum==1.13.2
|
| 19 |
hf_transfer
|
| 20 |
colorama
|