oobabooga
|
1c378965e1
Remove unused imports
|
2 éve |
oobabooga
|
66256ac1dd
Make the "no GPU has been detected" message more descriptive
|
2 éve |
oobabooga
|
265ba384b7
Rename a file, add deprecation warning for --load-in-4bit
|
2 éve |
Ayanami Rei
|
8778b756e6
use updated load_quantized
|
2 éve |
Ayanami Rei
|
e1c952c41c
make argument non case-sensitive
|
2 éve |
Ayanami Rei
|
3c9afd5ca3
rename method
|
2 éve |
Ayanami Rei
|
edbc61139f
use new quant loader
|
2 éve |
oobabooga
|
65dda28c9d
Rename --llama-bits to --gptq-bits
|
2 éve |
oobabooga
|
fed3617f07
Move LLaMA 4-bit into a separate file
|
2 éve |
draff
|
001e638b47
Make it actually work
|
2 éve |
draff
|
804486214b
Re-implement --load-in-4bit and update --llama-bits arg description
|
2 éve |
ItsLogic
|
9ba8156a70
remove unnecessary Path()
|
2 éve |
draff
|
e6c631aea4
Replace --load-in-4bit with --llama-bits
|
2 éve |
oobabooga
|
e9dbdafb14
Merge branch 'main' into pt-path-changes
|
2 éve |
oobabooga
|
706a03b2cb
Minor changes
|
2 éve |
oobabooga
|
de7dd8b6aa
Add comments
|
2 éve |
oobabooga
|
e461c0b7a0
Move the import to the top
|
2 éve |
deepdiffuser
|
9fbd60bf22
add no_split_module_classes to prevent tensor split error
|
2 éve |
deepdiffuser
|
ab47044459
add multi-gpu support for 4bit gptq LLaMA
|
2 éve |
rohvani
|
2ac2913747
fix reference issue
|
2 éve |
rohvani
|
826e297b0e
add llama-65b-4bit support & multiple pt paths
|
2 éve |
oobabooga
|
9849aac0f1
Don't show .pt models in the list
|
2 éve |
oobabooga
|
74102d5ee4
Insert to the path instead of appending
|
2 éve |
oobabooga
|
2965aa1625
Check if the .pt file exists
|
2 éve |
oobabooga
|
828a524f9a
Add LLaMA 4-bit support
|
2 éve |
oobabooga
|
e91f4bc25a
Add RWKV tokenizer
|
2 éve |
oobabooga
|
c33715ad5b
Move towards HF LLaMA implementation
|
2 éve |
oobabooga
|
bd8aac8fa4
Add LLaMA 8-bit support
|
2 éve |
oobabooga
|
ed8b35efd2
Add --pin-weight parameter for FlexGen
|
2 éve |
oobabooga
|
ea5c5eb3da
Add LLaMA support
|
2 éve |