Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[Question]: Content Length Limit Reached Unexpectedly in RAGFlow #5215

Closed
deulofeu1 opened this issue Feb 21, 2025 · 2 comments
Closed

[Question]: Content Length Limit Reached Unexpectedly in RAGFlow #5215

deulofeu1 opened this issue Feb 21, 2025 · 2 comments
Labels
question Further information is requested

Comments

@deulofeu1
Copy link

Describe your problem

Hello,

I am experiencing an issue with my dialogue setup involving a RAGFlow dialogue assistant and a local dialogue model. Despite configuring both models to allow for maximum token generation (max_tokens set to 128,000 for RAGFlow and 16,384 for the local model), the models terminate early after generating approximately 185 words or about 1,100 characters.

Here are some additional details of my setup:

RAGFlow Model: max_tokens is configured at 128,000.
Local Dialogue Model: max_tokens is set to 16,384.
Embedding Model: Deployed locally as text-embedding-bge-m3.
Context Length for Embedding Model: Set at 2048 tokens.
The generation stops with the message: "For content length reason, it stopped, continue?" This occurs despite the relatively short output. Could this issue be related to my embedding model configuration or other parameters? I am looking for guidance on how to resolve this premature termination due to supposed length constraints.

Thank you in advance for your support and advice.

@deulofeu1 deulofeu1 added the question Further information is requested label Feb 21, 2025
@KevinHuSh
Copy link
Collaborator

You could disable max tokens in dialog setting.

@deulofeu1
Copy link
Author

You could disable max tokens in dialog setting.您可以在对话框设置中禁用 max tokens

Yep this resolves the issue. I think the MAX_Token parameter seems to have some problems here.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
question Further information is requested
Projects
None yet
Development

No branches or pull requests

2 participants