浏览代码

Merge pull request #377 from askmyteapot/Fix-Multi-gpu-GPTQ-Llama-no-tokens

Update GPTQ_Loader.py
oobabooga 2 年之前
父节点
当前提交
4c130679c7
共有 1 个文件被更改,包括 1 次插入1 次删除
  1. 1 1
      modules/GPTQ_loader.py

+ 1 - 1
modules/GPTQ_loader.py

@@ -61,7 +61,7 @@ def load_quantized(model_name):
             max_memory[i] = f"{shared.args.gpu_memory[i]}GiB"
         max_memory['cpu'] = f"{shared.args.cpu_memory or '99'}GiB"
 
-        device_map = accelerate.infer_auto_device_map(model, max_memory=max_memory, no_split_module_classes=["LLaMADecoderLayer"])
+        device_map = accelerate.infer_auto_device_map(model, max_memory=max_memory, no_split_module_classes=["LlamaDecoderLayer"])
         model = accelerate.dispatch_model(model, device_map=device_map)
 
     # Single GPU