Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

LLM Output is truncated #163

Open
fka-at opened this issue Jan 31, 2025 · 0 comments
Open

LLM Output is truncated #163

fka-at opened this issue Jan 31, 2025 · 0 comments
Labels
bug Something isn't working

Comments

@fka-at
Copy link

fka-at commented Jan 31, 2025

Describe the bug

When running LLMs (Deepseek R1) from inside the AI Playground, the output seems to stop / get truncated after approximately 25 sec..
This is noticable when running Deepseek R1 which ouputs a lot of inner monologue. The output suddenly stops.
(According to the GPU-Utilization in Taskmanager, it appears, the model might keep running in the background, but its ouput is not displayed.)

To Reproduce

Steps to reproduce the behavior:

  1. Go to the Answer tab
  2. Load Deepseek-R1-Distill-Qwen-7B
  3. Ask it literally anything that makes it think more than a few seconds
  4. wait and observe

Expected behavior

A complete Output

Environment (please complete the following information):

  • OS: Windows 11 24H2
  • GPU: Intel Arc A770 16GB
  • CPU: Ryzen 5800X
  • RAM: 32GB
  • Version: 2.0.0 alpha
  • Backend: IPEX-LLM
@fka-at fka-at added the bug Something isn't working label Jan 31, 2025
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
bug Something isn't working
Projects
None yet
Development

No branches or pull requests

1 participant