Commit History

Author SHA1 Message Date
  oobabooga 265ba384b7 Rename a file, add deprecation warning for --load-in-4bit 3 years ago
  Ayanami Rei 8778b756e6 use updated load_quantized 3 years ago
  Ayanami Rei e1c952c41c make argument non case-sensitive 3 years ago
  Ayanami Rei 3c9afd5ca3 rename method 3 years ago
  Ayanami Rei edbc61139f use new quant loader 3 years ago
  oobabooga 65dda28c9d Rename --llama-bits to --gptq-bits 3 years ago
  oobabooga fed3617f07 Move LLaMA 4-bit into a separate file 3 years ago
  draff 001e638b47 Make it actually work 3 years ago
  draff 804486214b Re-implement --load-in-4bit and update --llama-bits arg description 3 years ago
  ItsLogic 9ba8156a70 remove unnecessary Path() 3 years ago
  draff e6c631aea4 Replace --load-in-4bit with --llama-bits 3 years ago
  oobabooga e9dbdafb14 Merge branch 'main' into pt-path-changes 3 years ago
  oobabooga 706a03b2cb Minor changes 3 years ago
  oobabooga de7dd8b6aa Add comments 3 years ago
  oobabooga e461c0b7a0 Move the import to the top 3 years ago
  deepdiffuser 9fbd60bf22 add no_split_module_classes to prevent tensor split error 3 years ago
  deepdiffuser ab47044459 add multi-gpu support for 4bit gptq LLaMA 3 years ago
  rohvani 2ac2913747 fix reference issue 3 years ago
  rohvani 826e297b0e add llama-65b-4bit support & multiple pt paths 3 years ago
  oobabooga 9849aac0f1 Don't show .pt models in the list 3 years ago
  oobabooga 74102d5ee4 Insert to the path instead of appending 3 years ago
  oobabooga 2965aa1625 Check if the .pt file exists 3 years ago
  oobabooga 828a524f9a Add LLaMA 4-bit support 3 years ago
  oobabooga e91f4bc25a Add RWKV tokenizer 3 years ago
  oobabooga c33715ad5b Move towards HF LLaMA implementation 3 years ago
  oobabooga bd8aac8fa4 Add LLaMA 8-bit support 3 years ago
  oobabooga ed8b35efd2 Add --pin-weight parameter for FlexGen 3 years ago
  oobabooga ea5c5eb3da Add LLaMA support 3 years ago
  oobabooga 659bb76722 Add RWKVModel class 3 years ago
  oobabooga 6837d4d72a Load the model by name 3 years ago