Commit History

Autor SHA1 Mensaxe Data
  Ayanami Rei 8778b756e6 use updated load_quantized %!s(int64=2) %!d(string=hai) anos
  Ayanami Rei e1c952c41c make argument non case-sensitive %!s(int64=2) %!d(string=hai) anos
  Ayanami Rei 3c9afd5ca3 rename method %!s(int64=2) %!d(string=hai) anos
  Ayanami Rei edbc61139f use new quant loader %!s(int64=2) %!d(string=hai) anos
  oobabooga 65dda28c9d Rename --llama-bits to --gptq-bits %!s(int64=2) %!d(string=hai) anos
  oobabooga fed3617f07 Move LLaMA 4-bit into a separate file %!s(int64=2) %!d(string=hai) anos
  draff 001e638b47 Make it actually work %!s(int64=2) %!d(string=hai) anos
  draff 804486214b Re-implement --load-in-4bit and update --llama-bits arg description %!s(int64=2) %!d(string=hai) anos
  ItsLogic 9ba8156a70 remove unnecessary Path() %!s(int64=2) %!d(string=hai) anos
  draff e6c631aea4 Replace --load-in-4bit with --llama-bits %!s(int64=2) %!d(string=hai) anos
  oobabooga e9dbdafb14 Merge branch 'main' into pt-path-changes %!s(int64=2) %!d(string=hai) anos
  oobabooga 706a03b2cb Minor changes %!s(int64=2) %!d(string=hai) anos
  oobabooga de7dd8b6aa Add comments %!s(int64=2) %!d(string=hai) anos
  oobabooga e461c0b7a0 Move the import to the top %!s(int64=2) %!d(string=hai) anos
  deepdiffuser 9fbd60bf22 add no_split_module_classes to prevent tensor split error %!s(int64=2) %!d(string=hai) anos
  deepdiffuser ab47044459 add multi-gpu support for 4bit gptq LLaMA %!s(int64=2) %!d(string=hai) anos
  rohvani 2ac2913747 fix reference issue %!s(int64=2) %!d(string=hai) anos
  rohvani 826e297b0e add llama-65b-4bit support & multiple pt paths %!s(int64=2) %!d(string=hai) anos
  oobabooga 9849aac0f1 Don't show .pt models in the list %!s(int64=2) %!d(string=hai) anos
  oobabooga 74102d5ee4 Insert to the path instead of appending %!s(int64=2) %!d(string=hai) anos
  oobabooga 2965aa1625 Check if the .pt file exists %!s(int64=2) %!d(string=hai) anos
  oobabooga 828a524f9a Add LLaMA 4-bit support %!s(int64=2) %!d(string=hai) anos
  oobabooga e91f4bc25a Add RWKV tokenizer %!s(int64=2) %!d(string=hai) anos
  oobabooga c33715ad5b Move towards HF LLaMA implementation %!s(int64=2) %!d(string=hai) anos
  oobabooga bd8aac8fa4 Add LLaMA 8-bit support %!s(int64=2) %!d(string=hai) anos
  oobabooga ed8b35efd2 Add --pin-weight parameter for FlexGen %!s(int64=2) %!d(string=hai) anos
  oobabooga ea5c5eb3da Add LLaMA support %!s(int64=2) %!d(string=hai) anos
  oobabooga 659bb76722 Add RWKVModel class %!s(int64=2) %!d(string=hai) anos
  oobabooga 6837d4d72a Load the model by name %!s(int64=2) %!d(string=hai) anos
  oobabooga 70e522732c Move RWKV loader into a separate file %!s(int64=2) %!d(string=hai) anos