Cronologia Commit

Autore SHA1 Messaggio Data
  oobabooga 8e89bc596b Fix encode() for RWKV 2 anni fa
  oobabooga 19a34941ed Add proper streaming to RWKV 2 anni fa
  oobabooga 8660227e1b Add top_k to RWKV 2 anni fa
  oobabooga 20bd645f6a Fix bug in multigpu setups (attempt 3) 2 anni fa
  oobabooga 09a7c36e1b Minor improvement while running custom models 2 anni fa
  oobabooga 24c4c20391 Fix bug in multigpu setups (attempt #2) 2 anni fa
  oobabooga d88b7836c6 Fix bug in multigpu setups 2 anni fa
  oobabooga e91f4bc25a Add RWKV tokenizer 2 anni fa
  oobabooga a54b91af77 Improve readability 2 anni fa
  oobabooga 8e706df20e Fix a memory leak when text streaming is on 2 anni fa
  oobabooga c33715ad5b Move towards HF LLaMA implementation 2 anni fa
  oobabooga c93f1fa99b Count the tokens more conservatively 2 anni fa
  oobabooga 05e703b4a4 Print the performance information more reliably 2 anni fa
  oobabooga a345a2acd2 Add a tokenizer placeholder 2 anni fa
  oobabooga 5b354817f6 Make chat minimally work with LLaMA 2 anni fa
  oobabooga ea5c5eb3da Add LLaMA support 2 anni fa
  oobabooga 7bbe32f618 Don't return a value in an iterator function 2 anni fa
  oobabooga ff9f649c0c Remove some unused imports 2 anni fa
  oobabooga 955cf431e8 Minor consistency fix 2 anni fa
  oobabooga 831ac7ed3f Add top_p 2 anni fa
  oobabooga 7c4d5ca8cc Improve the text generation call a bit 2 anni fa
  oobabooga 0f6708c471 Sort the imports 2 anni fa
  oobabooga e735806c51 Add a generate() function for RWKV 2 anni fa
  oobabooga f871971de1 Trying to get the chat to work 2 anni fa
  oobabooga ebd698905c Add streaming to RWKV 2 anni fa
  oobabooga 70e522732c Move RWKV loader into a separate file 2 anni fa
  oobabooga ebc64a408c RWKV support prototype 2 anni fa
  oobabooga 6e843a11d6 Fix FlexGen in chat mode 2 anni fa
  oobabooga fa58fd5559 Proper way to free the cuda cache 2 anni fa
  oobabooga 700311ce40 Empty the cuda cache at model.generate() 2 anni fa