|
@@ -179,9 +179,14 @@ def load_model(model_name):
|
|
|
tokenizer = AutoTokenizer.from_pretrained(Path(f"{shared.args.model_dir}/gpt-j-6B/"))
|
|
tokenizer = AutoTokenizer.from_pretrained(Path(f"{shared.args.model_dir}/gpt-j-6B/"))
|
|
|
elif type(model) is transformers.LlamaForCausalLM:
|
|
elif type(model) is transformers.LlamaForCausalLM:
|
|
|
tokenizer = LlamaTokenizer.from_pretrained(Path(f"{shared.args.model_dir}/{shared.model_name}/"), clean_up_tokenization_spaces=True)
|
|
tokenizer = LlamaTokenizer.from_pretrained(Path(f"{shared.args.model_dir}/{shared.model_name}/"), clean_up_tokenization_spaces=True)
|
|
|
- tokenizer.eos_token_id = 2
|
|
|
|
|
- tokenizer.bos_token_id = 1
|
|
|
|
|
- tokenizer.pad_token_id = 0
|
|
|
|
|
|
|
+ # Leaving this here until the LLaMA tokenizer gets figured out.
|
|
|
|
|
+ # For some people this fixes things, for others it causes an error.
|
|
|
|
|
+ try:
|
|
|
|
|
+ tokenizer.eos_token_id = 2
|
|
|
|
|
+ tokenizer.bos_token_id = 1
|
|
|
|
|
+ tokenizer.pad_token_id = 0
|
|
|
|
|
+ except:
|
|
|
|
|
+ continue
|
|
|
else:
|
|
else:
|
|
|
tokenizer = AutoTokenizer.from_pretrained(Path(f"{shared.args.model_dir}/{shared.model_name}/"))
|
|
tokenizer = AutoTokenizer.from_pretrained(Path(f"{shared.args.model_dir}/{shared.model_name}/"))
|
|
|
tokenizer.truncation_side = 'left'
|
|
tokenizer.truncation_side = 'left'
|