Browse Source

update notes for int8 lack of support in FSDP

Hamid Shojanazeri 1 year atrás
parent
commit
79028b14b3
2 changed files with 3 additions and 1 deletions
  1. 1 1
      README.md
  2. 2 0
      docs/mutli_gpu.md

+ 1 - 1
README.md

@@ -81,7 +81,7 @@ Here we make use of Parameter Efficient Methods (PEFT) as described in the next
 
 ### Multiple GPUs One Node:
 
-**NOTE** please make sure to use PyTorch Nightlies for using PEFT+FSDP .
+**NOTE** please make sure to use PyTorch Nightlies for using PEFT+FSDP. Also, note that int8 quantization from bit&bytes currently is not supported in FSDP.
 
 ```bash
 

+ 2 - 0
docs/mutli_gpu.md

@@ -26,6 +26,8 @@ This runs with the `samsum_dataset` for summarization application by default.
 
 **Multiple GPUs one node**:
 
+**NOTE** please make sure to use PyTorch Nightlies for using PEFT+FSDP. Also, note that int8 quantization from bit&bytes currently is not supported in FSDP.
+
 ```bash
 
 torchrun --nnodes 1 --nproc_per_node 4  ../llama_finetuning.py --enable_fsdp --model_name /patht_of_model_folder/7B --use_peft --peft_method lora --output_dir Path/to/save/PEFT/model