Detect "vicuna" as llama model type (#772)
This commit is contained in:
@@ -52,7 +52,7 @@ def load_quantized(model_name):
|
|||||||
if not shared.args.model_type:
|
if not shared.args.model_type:
|
||||||
# Try to determine model type from model name
|
# Try to determine model type from model name
|
||||||
name = model_name.lower()
|
name = model_name.lower()
|
||||||
if any((k in name for k in ['llama', 'alpaca'])):
|
if any((k in name for k in ['llama', 'alpaca', 'vicuna'])):
|
||||||
model_type = 'llama'
|
model_type = 'llama'
|
||||||
elif any((k in name for k in ['opt-', 'galactica'])):
|
elif any((k in name for k in ['opt-', 'galactica'])):
|
||||||
model_type = 'opt'
|
model_type = 'opt'
|
||||||
|
|||||||
Reference in New Issue
Block a user