-
Notifications
You must be signed in to change notification settings - Fork 16.2k
Eval bug: Model output directed to Reasoning block instead of standard response. #20265
Copy link
Copy link
Open
Labels
bugSomething isn't workingSomething isn't workingchat parserIssues related to the chat parser and chat templatesIssues related to the chat parser and chat templatesregressionA regression introduced in a new build (something that was previously working correctly)A regression introduced in a new build (something that was previously working correctly)
Description
Name and Version
command is : llama-server -m Qwen3.5-9B-Claude-4.6-HighIQ-INSTRUCT-HERETIC-UNCENSORED.i1-Q4_K_M.gguf --host 0.0.0.0
version: llama-b8244-bin-win-cuda-12.4-x64

Operating systems
Windows
GGML backends
CUDA
Hardware
RTX 5070
Models
Qwen3.5-9B-Claude-4.6-HighIQ-INSTRUCT-HERETIC-UNCENSORED.i1-Q4_K_M.gguf
Problem description & steps to reproduce
Model output directed to Reasoning block instead of standard response.
First Bad Commit
No response
Relevant log output
Logs
Reactions are currently unavailable
Metadata
Metadata
Assignees
Labels
bugSomething isn't workingSomething isn't workingchat parserIssues related to the chat parser and chat templatesIssues related to the chat parser and chat templatesregressionA regression introduced in a new build (something that was previously working correctly)A regression introduced in a new build (something that was previously working correctly)