Commit History

Autor SHA1 Mensaxe Data
  EyeDeck 39f3fec913 Broaden GPTQ-for-LLaMA branch support (#820) %!s(int64=2) %!d(string=hai) anos
  oobabooga 3d6cb5ed63 Minor rewrite %!s(int64=2) %!d(string=hai) anos
  oobabooga f3a2e0b8a9 Disable pre_layer when the model type is not llama %!s(int64=2) %!d(string=hai) anos
  catalpaaa 4ab679480e allow quantized model to be loaded from model dir (#760) %!s(int64=2) %!d(string=hai) anos
  OWKenobi ee4547cd34 Detect "vicuna" as llama model type (#772) %!s(int64=2) %!d(string=hai) anos
  oobabooga 1cb9246160 Adapt to the new model names %!s(int64=2) %!d(string=hai) anos
  oobabooga 010b259dde Update documentation %!s(int64=2) %!d(string=hai) anos
  oobabooga 0bec15ebcd Reorder imports %!s(int64=2) %!d(string=hai) anos
  Maya Eary 41ec682834 Disable kernel threshold for gpt-j %!s(int64=2) %!d(string=hai) anos
  Maya Eary 1c075d8d21 Fix typo %!s(int64=2) %!d(string=hai) anos
  Maya Eary c8207d474f Generalized load_quantized %!s(int64=2) %!d(string=hai) anos
  oobabooga 49c10c5570 Add support for the latest GPTQ models with group-size (#530) %!s(int64=2) %!d(string=hai) anos
  EyeDeck dcfd866402 Allow loading of .safetensors through GPTQ-for-LLaMa %!s(int64=2) %!d(string=hai) anos
  oobabooga db4219a340 Update comments %!s(int64=2) %!d(string=hai) anos
  oobabooga 7618f3fe8c Add -gptq-preload for 4-bit offloading (#460) %!s(int64=2) %!d(string=hai) anos
  oobabooga 9a3bed50c3 Attempt at fixing 4-bit with CPU offload %!s(int64=2) %!d(string=hai) anos
  askmyteapot 53b6a66beb Update GPTQ_Loader.py %!s(int64=2) %!d(string=hai) anos
  oobabooga 265ba384b7 Rename a file, add deprecation warning for --load-in-4bit %!s(int64=2) %!d(string=hai) anos