qgallouedec HF Staff commited on
Commit
cd37a38
·
1 Parent(s): f5a4c3a

rm flash attn

Browse files
Dockerfile CHANGED
@@ -15,8 +15,7 @@ COPY . .
15
 
16
  # Install Python dependencies
17
  RUN pip install --upgrade pip && \
18
- pip install --no-cache-dir -r requirements.txt && \
19
- MAX_JOBS=4 pip install flash-attn --use-pep517 --no-build-isolation
20
 
21
  # Make script executable and move it to PATH
22
  RUN install -m 755 sft.py /usr/local/bin/sft
 
15
 
16
  # Install Python dependencies
17
  RUN pip install --upgrade pip && \
18
+ pip install --no-cache-dir -r requirements.txt
 
19
 
20
  # Make script executable and move it to PATH
21
  RUN install -m 755 sft.py /usr/local/bin/sft
configs/Qwen3-4B-Base_a10g-small.yaml CHANGED
@@ -2,7 +2,6 @@
2
  model_name_or_path: Qwen/Qwen2.5-1.5B
3
  model_revision: main
4
  torch_dtype: bfloat16
5
- attn_implementation: flash_attention_2
6
 
7
  # Data training arguments
8
  dataset_name: trl-lib/tldr
 
2
  model_name_or_path: Qwen/Qwen2.5-1.5B
3
  model_revision: main
4
  torch_dtype: bfloat16
 
5
 
6
  # Data training arguments
7
  dataset_name: trl-lib/tldr