You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Below is an instruction that describes a task, paired with an input that provides further context. Write a response that appropriately completes the request.
### Instruction:
Write a python function that adds two numbers
### Response:
Screenshot
Logs
Output generated in 4.61 seconds (4.12 tokens/s, 19 tokens)
System Info
Windows 11, RTX3090
The text was updated successfully, but these errors were encountered:
After further testing, it appears that issue occurs when max_new_tokens is >= 1992. This issue does not occur when max_new_tokens is less than 1992, but it does duplicate the last token of the input.
For generating long text (like a long novel) in several iterations (clicking on Continue multiple times), it is convenient to truncate the beginning of the text to make room for more tokens. This is related to #498 btw
Describe the bug
Ive tested several alpaca models and iv'e noticed that start of the response is almost always missing.
Might be similar to this issue: #300
Is there an existing issue for this?
Reproduction
Load Alpaca model https://huggingface.co/elinas/alpaca-30b-lora-int4/tree/main and generate using this prompt:
Screenshot
Logs
Output generated in 4.61 seconds (4.12 tokens/s, 19 tokens)
System Info
The text was updated successfully, but these errors were encountered: