-
Notifications
You must be signed in to change notification settings - Fork 14.9k
NVIDIA connector is experiencing CRITICAL issues #10885
Description
Description
The connector doesn't cover the entire list of available models, but only a small subset, which is displayed in the open code.
I wrote a small script to query the available models:
import os
import sys
import requests
# Load environment variables if dotenv is available
try:
from dotenv import load_dotenv
load_dotenv()
except ImportError:
pass
NGC_KEY = os.getenv('NGC_KEY')
if not NGC_KEY:
print("Error: NGC_KEY environment variable is not set.", file=sys.stderr)
sys.exit(1)
def fetch_models():
url = 'https://integrate.api.nvidia.com/v1/models'
headers = {'X-Api-Key': NGC_KEY}
try:
response = requests.get(url, headers=headers)
response.raise_for_status()
data = response.json()
for model in data.get('data', []):
model_id = model.get('id')
if model_id:
print(model_id)
except requests.exceptions.RequestException as e:
print(f"Error fetching models: {e}", file=sys.stderr)
sys.exit(1)
if __name__ == "__main__":
fetch_models()As a result, I got the following list (spoiler)
- 01-ai/yi-large
- abacusai/dracarys-llama-3.1-70b-instruct
- adept/fuyu-8b
- ai21labs/jamba-1.5-large-instruct
- ai21labs/jamba-1.5-mini-instruct
- aisingapore/sea-lion-7b-instruct
- baai/bge-m3
- baichuan-inc/baichuan2-13b-chat
- bigcode/starcoder2-15b
- bigcode/starcoder2-7b
- bytedance/seed-oss-36b-instruct
- databricks/dbrx-instruct
- deepseek-ai/deepseek-coder-6.7b-instruct
- deepseek-ai/deepseek-r1-distill-llama-8b
- deepseek-ai/deepseek-r1-distill-qwen-14b
- deepseek-ai/deepseek-r1-distill-qwen-32b
- deepseek-ai/deepseek-r1-distill-qwen-7b
- deepseek-ai/deepseek-v3.1
- deepseek-ai/deepseek-v3.1-terminus
- deepseek-ai/deepseek-v3.2
- google/codegemma-1.1-7b
- google/codegemma-7b
- google/deplot
- google/gemma-2-27b-it
- google/gemma-2-2b-it
- google/gemma-2-9b-it
- google/gemma-2b
- google/gemma-3-12b-it
- google/gemma-3-1b-it
- google/gemma-3-27b-it
- google/gemma-3-4b-it
- google/gemma-3n-e2b-it
- google/gemma-3n-e4b-it
- google/gemma-7b
- google/paligemma
- google/recurrentgemma-2b
- google/shieldgemma-9b
- gotocompany/gemma-2-9b-cpt-sahabatai-instruct
- ibm/granite-3.0-3b-a800m-instruct
- ibm/granite-3.0-8b-instruct
- ibm/granite-3.3-8b-instruct
- ibm/granite-34b-code-instruct
- ibm/granite-8b-code-instruct
- ibm/granite-guardian-3.0-8b
- igenius/colosseum_355b_instruct_16k
- igenius/italia_10b_instruct_16k
- institute-of-science-tokyo/llama-3.1-swallow-70b-instruct-v0.1
- institute-of-science-tokyo/llama-3.1-swallow-8b-instruct-v0.1
- marin/marin-8b-instruct
- mediatek/breeze-7b-instruct
- meta/codellama-70b
- meta/llama-3.1-405b-instruct
- meta/llama-3.1-70b-instruct
- meta/llama-3.1-8b-instruct
- meta/llama-3.2-11b-vision-instruct
- meta/llama-3.2-1b-instruct
- meta/llama-3.2-3b-instruct
- meta/llama-3.2-90b-vision-instruct
- meta/llama-3.3-70b-instruct
- meta/llama-4-maverick-17b-128e-instruct
- meta/llama-4-scout-17b-16e-instruct
- meta/llama-guard-4-12b
- meta/llama2-70b
- meta/llama3-70b-instruct
- meta/llama3-8b-instruct
- microsoft/kosmos-2
- microsoft/phi-3-medium-128k-instruct
- microsoft/phi-3-medium-4k-instruct
- microsoft/phi-3-mini-128k-instruct
- microsoft/phi-3-mini-4k-instruct
- microsoft/phi-3-small-128k-instruct
- microsoft/phi-3-small-8k-instruct
- microsoft/phi-3-vision-128k-instruct
- microsoft/phi-3.5-mini-instruct
- microsoft/phi-3.5-moe-instruct
- microsoft/phi-3.5-vision-instruct
- microsoft/phi-4-mini-flash-reasoning
- microsoft/phi-4-mini-instruct
- microsoft/phi-4-multimodal-instruct
- minimaxai/minimax-m2
- minimaxai/minimax-m2.1
- mistralai/codestral-22b-instruct-v0.1
- mistralai/devstral-2-123b-instruct-2512
- mistralai/magistral-small-2506
- mistralai/mamba-codestral-7b-v0.1
- mistralai/mathstral-7b-v0.1
- mistralai/ministral-14b-instruct-2512
- mistralai/mistral-7b-instruct-v0.2
- mistralai/mistral-7b-instruct-v0.3
- mistralai/mistral-large
- mistralai/mistral-large-2-instruct
- mistralai/mistral-large-3-675b-instruct-2512
- mistralai/mistral-medium-3-instruct
- mistralai/mistral-nemotron
- mistralai/mistral-small-24b-instruct
- mistralai/mistral-small-3.1-24b-instruct-2503
- mistralai/mixtral-8x22b-instruct-v0.1
- mistralai/mixtral-8x22b-v0.1
- mistralai/mixtral-8x7b-instruct-v0.1
- moonshotai/kimi-k2-instruct
- moonshotai/kimi-k2-instruct-0905
- moonshotai/kimi-k2-thinking
- moonshotai/kimi-k2.5
- nv-mistralai/mistral-nemo-12b-instruct
- nvidia/cosmos-reason2-8b
- nvidia/embed-qa-4
- nvidia/llama-3.1-nemoguard-8b-content-safety
- nvidia/llama-3.1-nemoguard-8b-topic-control
- nvidia/llama-3.1-nemotron-51b-instruct
- nvidia/llama-3.1-nemotron-70b-instruct
- nvidia/llama-3.1-nemotron-70b-reward
- nvidia/llama-3.1-nemotron-nano-4b-v1.1
- nvidia/llama-3.1-nemotron-nano-8b-v1
- nvidia/llama-3.1-nemotron-nano-vl-8b-v1
- nvidia/llama-3.1-nemotron-safety-guard-8b-v3
- nvidia/llama-3.1-nemotron-ultra-253b-v1
- nvidia/llama-3.2-nemoretriever-1b-vlm-embed-v1
- nvidia/llama-3.2-nemoretriever-300m-embed-v1
- nvidia/llama-3.2-nemoretriever-300m-embed-v2
- nvidia/llama-3.2-nv-embedqa-1b-v1
- nvidia/llama-3.2-nv-embedqa-1b-v2
- nvidia/llama-3.3-nemotron-super-49b-v1
- nvidia/llama-3.3-nemotron-super-49b-v1.5
- nvidia/llama3-chatqa-1.5-70b
- nvidia/llama3-chatqa-1.5-8b
- nvidia/mistral-nemo-minitron-8b-8k-instruct
- nvidia/mistral-nemo-minitron-8b-base
- nvidia/nemoretriever-parse
- nvidia/nemotron-3-nano-30b-a3b
- nvidia/nemotron-4-340b-instruct
- nvidia/nemotron-4-340b-reward
- nvidia/nemotron-4-mini-hindi-4b-instruct
- nvidia/nemotron-content-safety-reasoning-4b
- nvidia/nemotron-mini-4b-instruct
- nvidia/nemotron-nano-12b-v2-vl
- nvidia/nemotron-nano-3-30b-a3b
- nvidia/nemotron-parse
- nvidia/neva-22b
- nvidia/nv-embed-v1
- nvidia/nv-embedcode-7b-v1
- nvidia/nv-embedqa-e5-v5
- nvidia/nv-embedqa-mistral-7b-v2
- nvidia/nvclip
- nvidia/nvidia-nemotron-nano-9b-v2
- nvidia/riva-translate-4b-instruct
- nvidia/riva-translate-4b-instruct-v1.1
- nvidia/streampetr
- nvidia/usdcode-llama-3.1-70b-instruct
- nvidia/vila
- openai/gpt-oss-120b
- openai/gpt-oss-120b
- openai/gpt-oss-20b
- openai/gpt-oss-20b
- opengpt-x/teuken-7b-instruct-commercial-v0.4
- qwen/qwen2-7b-instruct
- qwen/qwen2.5-7b-instruct
- qwen/qwen2.5-coder-32b-instruct
- qwen/qwen2.5-coder-7b-instruct
- qwen/qwen3-235b-a22b
- qwen/qwen3-coder-480b-a35b-instruct
- qwen/qwen3-next-80b-a3b-instruct
- qwen/qwen3-next-80b-a3b-thinking
- qwen/qwq-32b
- rakuten/rakutenai-7b-chat
- rakuten/rakutenai-7b-instruct
- sarvamai/sarvam-m
- snowflake/arctic-embed-l
- speakleash/bielik-11b-v2.3-instruct
- speakleash/bielik-11b-v2.6-instruct
- stockmark/stockmark-2-100b-instruct
- thudm/chatglm3-6b
- tiiuae/falcon3-7b-instruct
- tokyotech-llm/llama-3-swallow-70b-instruct-v0.1
- upstage/solar-10.7b-instruct
- utter-project/eurollm-9b-instruct
- writer/palmyra-creative-122b
- writer/palmyra-fin-70b-32k
- writer/palmyra-med-70b
- writer/palmyra-med-70b-32k
- yentinglin/llama-3-taiwan-70b-instruct
- z-ai/glm4.7
- zyphra/zamba2-7b-instruct
Yes, not all of them support text input, for example nvidia/nemotron-parse, but nevertheless I surveyed each of the models and based on the survey results
received successful responses from 116 models. (spoiler)
abacusai/dracarys-llama-3.1-70b-instruct: 1.24s
ai21labs/jamba-1.5-mini-instruct: 1.24s
deepseek-ai/deepseek-r1-distill-qwen-7b: 1.42s
deepseek-ai/deepseek-r1-distill-qwen-14b: 1.70s
google/gemma-3-1b-it: 0.85s
bytedance/seed-oss-36b-instruct: 2.11s
google/gemma-2-2b-it: 1.42s
google/paligemma: Error: 500 (0.97s)
google/shieldgemma-9b: 0.85s
google/gemma-2-27b-it: 1.85s
google/gemma-2-9b-it: 1.89s
google/gemma-3-27b-it: 1.64s
baichuan-inc/baichuan2-13b-chat: 2.99s
gotocompany/gemma-2-9b-cpt-sahabatai-instruct: 1.17s
google/gemma-7b: 2.04s
ibm/granite-guardian-3.0-8b: 0.99s
igenius/italia_10b_instruct_16k: 0.92s
institute-of-science-tokyo/llama-3.1-swallow-70b-instruct-v0.1: 1.07s
institute-of-science-tokyo/llama-3.1-swallow-8b-instruct-v0.1: 1.05s
google/gemma-3n-e2b-it: 2.89s
deepseek-ai/deepseek-v3.1: 4.17s
marin/marin-8b-instruct: 1.31s
mediatek/breeze-7b-instruct: 1.39s
meta/llama-3.2-1b-instruct: 0.77s
meta/llama-3.2-3b-instruct: 0.66s
meta/llama-3.1-8b-instruct: 1.06s
ibm/granite-3.3-8b-instruct: 2.18s
google/gemma-3n-e4b-it: 3.48s
meta/llama-3.1-70b-instruct: 1.35s
meta/llama-3.1-405b-instruct: 1.79s
igenius/colosseum_355b_instruct_16k: 2.45s
meta/llama-4-scout-17b-16e-instruct: 0.67s
meta/llama-guard-4-12b: 0.67s
meta/llama-4-maverick-17b-128e-instruct: 1.13s
meta/llama-3.2-11b-vision-instruct: 1.75s
meta/llama-3.2-90b-vision-instruct: 1.52s
meta/llama-3.3-70b-instruct: 1.47s
meta/llama3-8b-instruct: 1.20s
google/gemma-3-4b-it: 4.75s
meta/llama3-70b-instruct: 1.79s
microsoft/phi-3-small-128k-instruct: 1.33s
microsoft/phi-3-medium-4k-instruct: 1.61s
microsoft/phi-3-mini-4k-instruct: 1.51s
microsoft/phi-3.5-vision-instruct: 1.03s
microsoft/phi-3-small-8k-instruct: 1.59s
microsoft/phi-4-mini-flash-reasoning: 1.28s
microsoft/phi-3-mini-128k-instruct: 2.04s
microsoft/phi-3.5-mini-instruct: 1.93s
minimaxai/minimax-m2: 1.58s
google/gemma-3-12b-it: 6.65s
microsoft/phi-3-medium-128k-instruct: 2.96s
mistralai/ministral-14b-instruct-2512: 1.33s
minimaxai/minimax-m2.1: 2.06s
mistralai/mathstral-7b-v0.1: 1.54s
mistralai/mistral-7b-instruct-v0.2: 1.50s
mistralai/mistral-7b-instruct-v0.3: 1.72s
mistralai/mamba-codestral-7b-v0.1: 2.03s
mistralai/devstral-2-123b-instruct-2512: 2.30s
mistralai/magistral-small-2506: 2.33s
moonshotai/kimi-k2-instruct: 1.25s
mistralai/mistral-nemotron: 1.96s
mistralai/mixtral-8x7b-instruct-v0.1: 1.51s
moonshotai/kimi-k2-instruct-0905: 1.38s
mistralai/mistral-medium-3-instruct: 2.22s
mistralai/mixtral-8x22b-instruct-v0.1: 1.91s
moonshotai/kimi-k2-thinking: 1.54s
nvidia/llama-3.1-nemoguard-8b-topic-control: 0.64s
nvidia/llama-3.1-nemoguard-8b-content-safety: 0.81s
mistralai/mistral-small-24b-instruct: 2.54s
mistralai/mistral-small-3.1-24b-instruct-2503: 2.58s
nvidia/llama-3.1-nemotron-safety-guard-8b-v3: 0.70s
nvidia/llama-3.1-nemotron-nano-4b-v1.1: 1.25s
nvidia/llama-3.1-nemotron-nano-vl-8b-v1: 1.34s
nvidia/llama-3.1-nemotron-nano-8b-v1: 1.43s
nvidia/llama3-chatqa-1.5-8b: 0.89s
nvidia/llama-3.1-nemotron-70b-reward: 1.98s
nvidia/nemoretriever-parse: Error: 400 (0.80s)
nvidia/llama-3.1-nemotron-ultra-253b-v1: 1.87s
nvidia/nemotron-3-nano-30b-a3b: 1.12s
nvidia/nemotron-content-safety-reasoning-4b: 0.87s
nvidia/nemotron-4-mini-hindi-4b-instruct: 1.12s
nvidia/nemotron-mini-4b-instruct: 1.10s
nvidia/nemotron-parse: Error: 400 (0.87s)
nvidia/llama-3.3-nemotron-super-49b-v1.5: 2.20s
nvidia/llama-3.3-nemotron-super-49b-v1: 2.29s
nvidia/nemotron-nano-12b-v2-vl: 1.50s
openai/gpt-oss-20b: 0.80s
openai/gpt-oss-120b: 1.03s
openai/gpt-oss-120b: 1.22s
nvidia/riva-translate-4b-instruct-v1.1: 1.30s
openai/gpt-oss-20b: 1.22s
qwen/qwen2-7b-instruct: 1.20s
qwen/qwen2.5-7b-instruct: 1.23s
qwen/qwen2.5-coder-7b-instruct: 0.97s
opengpt-x/teuken-7b-instruct-commercial-v0.4: 1.80s
qwen/qwen3-235b-a22b: 1.84s
qwen/qwen2.5-coder-32b-instruct: 2.22s
rakuten/rakutenai-7b-instruct: 1.67s
rakuten/rakutenai-7b-chat: 2.02s
nvidia/usdcode-llama-3.1-70b-instruct: 4.11s
sarvamai/sarvam-m: 2.27s
speakleash/bielik-11b-v2.3-instruct: 1.62s
moonshotai/kimi-k2.5: 8.52s
deepseek-ai/deepseek-v3.1-terminus: 17.10s
speakleash/bielik-11b-v2.6-instruct: 2.17s
qwen/qwen3-coder-480b-a35b-instruct: 3.95s
thudm/chatglm3-6b: 1.57s
tiiuae/falcon3-7b-instruct: 1.33s
stockmark/stockmark-2-100b-instruct: 2.66s
utter-project/eurollm-9b-instruct: 1.44s
upstage/solar-10.7b-instruct: 2.36s
tokyotech-llm/llama-3-swallow-70b-instruct-v0.1: 3.24s
yentinglin/llama-3-taiwan-70b-instruct: 2.37s
nvidia/nvidia-nemotron-nano-9b-v2: 9.11s
z-ai/glm4.7: 26.84s
microsoft/phi-4-multimodal-instruct: 58.71s
qwen/qwen3-next-80b-a3b-thinking: 183.53s
qwen/qwen3-next-80b-a3b-instruct: 266.68s
mistralai/mistral-large-3-675b-instruct-2512: 293.33s
deepseek-ai/deepseek-r1-distill-llama-8b: Error: Timeout (300.00s)
deepseek-ai/deepseek-v3.2: Error: Timeout (300.00s)
deepseek-ai/deepseek-r1-distill-qwen-32b: Error: Timeout (300.00s)
microsoft/phi-4-mini-instruct: Error: Timeout (300.00s)
qwen/qwq-32b: Error: Timeout (300.00s)
Yes, large models like the Qwen3 and DeepSeek R1 didn't respond and were timed out, but the very high-quality GLM model did respond. If you select the GLM model in the open code, it simply doesn't respond, although I personally verified that it works fine using the API.
I urge you to look into this issue, as this connector is a gem in vibe coding.
I also found a problem with some llama models in opencode, instead of calling tools they send a <tool_call> message and this is not handled in any way:
Check file VERSION.md
11:33 PM
Thinking: Okay, the user wants me to check the contents of the VERSION.md file. Let me start by recalling the available tools. The 'read' tool is used to read files from the local filesystem. The parameters require an absolute file path. Since the user mentioned VERSION.md, I need to confirm the absolute path.
Looking at the environment details, the working directory is /home/albert/Desktop/mytest. So the absolute path would be /home/albert/Desktop/mytest/VERSION.md. I should use the read tool with this file path. There's no mention of needing offset or limit, so I'll omit those parameters. Let me make sure that the file exists. Wait, the user didn't specify, but the read tool can handle non-existent files by returning an error. But since the user is asking to check it, it's likely that the file exists. I'll proceed to call the read tool with the specified path.
<tool_call>
{"name": "read", "arguments": {"filePath": "/home/user/mytest/VERSION.md"}}
</tool_call>
Due to this critical issue, the model becomes completely unsuitable for any interaction, since it can only effectively respond in text.
Plugins
none
OpenCode version
1.1.36
Steps to reproduce
No response
Screenshot and/or share link
No response
Operating System
Linux Mint 22.1 x86_64
Terminal
gnome-terminal / bash 5.2.21