Jelajahi Sumber

Merge pull request #377 from askmyteapot/Fix-Multi-gpu-GPTQ-Llama-no-tokens

Update GPTQ_Loader.py
oobabooga 2 tahun lalu
induk
melakukan
4c130679c7
1 mengubah file dengan 1 tambahan dan 1 penghapusan
  1. 1 1
      modules/GPTQ_loader.py

+ 1 - 1
modules/GPTQ_loader.py

@@ -61,7 +61,7 @@ def load_quantized(model_name):
             max_memory[i] = f"{shared.args.gpu_memory[i]}GiB"
             max_memory[i] = f"{shared.args.gpu_memory[i]}GiB"
         max_memory['cpu'] = f"{shared.args.cpu_memory or '99'}GiB"
         max_memory['cpu'] = f"{shared.args.cpu_memory or '99'}GiB"
 
 
-        device_map = accelerate.infer_auto_device_map(model, max_memory=max_memory, no_split_module_classes=["LLaMADecoderLayer"])
+        device_map = accelerate.infer_auto_device_map(model, max_memory=max_memory, no_split_module_classes=["LlamaDecoderLayer"])
         model = accelerate.dispatch_model(model, device_map=device_map)
         model = accelerate.dispatch_model(model, device_map=device_map)
 
 
     # Single GPU
     # Single GPU