After that, you can download and run the model with:
ollama run --verbose gpt-oss:20b
Note that the Jetson AI Lab tutorial does not include the --runtime nvidia option in the example command. In our tests, adding this parameter was necessary for the Thor GPU to be recognized correctly.
As shown in the attached screenshot, during execution the GPU utilization reached 97%, confirming that the Thor GPU was fully engaged. The monitoring interface in the image is from Cordatus.
👉 By September 10, you will be able to download and use the Jetson Thor–enabled version of this application.
5406663 Currently, you can monitor GPU utilization by using nvidia-smi dmon. However, because of design changes, integration of GPU utilization in Jetson Power GUI is still under evaluation.
Added nvml and jetpack 7 support to jetson_stats and opened a pull request, though the project looks fairly abandoned at this point so not sure if it will get integrated.
Even though Docker and NVIDIA Container Runtime are pre-installed on JetPack 7.0 GA, users who flash the Jetson device via USB (instead of using SDK Manager or the flash.sh script) should note that the Docker configuration file (/etc/docker/daemon.json) does not include the "default-runtime": "nvidia" setting by default.
This means that running GPU-enabled containers requires adding --runtime=nvidia to every Docker command — unless the config file is updated manually.
For example, after flashing via USB, the content of /etc/docker/daemon.json looks like this: