-
Notifications
You must be signed in to change notification settings - Fork 14.2k
Closed
Labels
bugSomething isn't workingSomething isn't workingcoreAnything pertaining to core functionality of the application (opencode server stuff)Anything pertaining to core functionality of the application (opencode server stuff)
Description
Description
There are already many patterns for detecting context overflow, for different providers. However, there are no dedicated patterns for vLLM yet. This means that when using a vLLM instance as a provider, auto compaction is not trigger when running into an error due to context overflow.
Plugins
No response
OpenCode version
1.2.26
Steps to reproduce
- Serve some model with vLLM
- configure that vLLM instance as a provider (OpenAI compatible)
- send a message to vLLM that is too large for the model's context window
- Observe that there is an error, but no auto compaction
Screenshot and/or share link
No response
Operating System
Arch Linux (irrelevant)
Terminal
Alacritty (irrelevant)
Reactions are currently unavailable
Metadata
Metadata
Assignees
Labels
bugSomething isn't workingSomething isn't workingcoreAnything pertaining to core functionality of the application (opencode server stuff)Anything pertaining to core functionality of the application (opencode server stuff)