Hi developers, I got a brand new NVIDIA Jeston Orin board,and I tried to deploy Qwen2.5-VL-7B via TensorRT-LLM,but I couldn’t find a complete guide for my purpose.Any suggestions or guides would be good for me.
Thanks!
Hi developers, I got a brand new NVIDIA Jeston Orin board,and I tried to deploy Qwen2.5-VL-7B via TensorRT-LLM,but I couldn’t find a complete guide for my purpose.Any suggestions or guides would be good for me.
Thanks!
*** Please note that this reply is generated by LLM automatically ***
I’d be happy to help you with deploying Qwen25VL7B via TensorRT on your NVIDIA Jetson Orin board.
Firstly, I’d like to clarify that the Jetson Orin board is a powerful AI computing platform, and TensorRT is a software development kit (SDK) for optimizing and deploying deep learning models on NVIDIA GPUs, including the Jetson Orin.
To deploy Qwen25VL7B via TensorRT on your Jetson Orin board, you’ll need to follow these general steps:
Here are some specific guides and resources that may be helpful for deploying Qwen25VL7B via TensorRT on your Jetson Orin board:
In terms of specific guides for deploying Qwen25VL7B via TensorRT on the Jetson Orin board, I couldn’t find any publicly available guides that are specific to this model and board combination. However, the general steps and resources I provided above should be helpful in getting you started.
If you have any specific questions or need further guidance, feel free to ask!
Comparison with other boards:
The Jetson Orin board is a powerful AI computing platform that is designed for edge AI applications. Compared to other boards, such as the Jetson Nano or Jetson Xavier, the Jetson Orin board offers a number of advantages, including:
Overall, the Jetson Orin board is a powerful and flexible platform that is well-suited for a wide range of AI applications.
*** Please give the thumbs up if you get answers you like. Or provide feedback to help us improve the answer. ***
Hi,
It’s recommended to use the vLLM container.
Please find the tutorial below:
Thanks.