-
-
Notifications
You must be signed in to change notification settings - Fork 369
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Yolov8s no bounding box on default settings #597
Comments
Did you solve this error, the default yolov8s after running it with deepstream-app -c deep* no detection is shown no bounding boxes |
@PaoXi No, I didn't have time to dig into it. And I didn't find any clue yet. |
@PaoXi is your board also Orin Nano? |
ok for rtp/h264) - [How to configure h265 stream?](marcoslucianops/DeepStream-Yolo#600) - [Yolov8s no bounding box on default settings #597](marcoslucianops/DeepStream-Yolo#597)
Yes it's 16 Gb |
Can someone send me the exported onnx file (from Orin Nano)? |
There're two version, one exported using ultralytics API. other one directly with pytorch script https://www.mediafire.com/file/edzascweikxrup9/yolov8s.pt.onnx/file |
They're two version, one with ultralytics's API, and the other using directly pytorch https://www.mediafire.com/file/edzascweikxrup9/yolov8s.pt.onnx/file |
@marcoslucianops I don't know exact model version, but I have downloaded here. export from pt to onnx command: $ yolo export model=yolov8s.pt format=onnx
$ yolo version
8.3.33 Attached below: |
@lida2003 you need to export with the |
@marcoslucianops The result is the same (no bounding boxes), but I got the right way to generate onnx file, thanks. PS: delete |
@marcoslucianops We are trying to reproduce the performance mentioned in the following link, which claims to achieve 181 FPS with INT8 precision on Jetson Orin NX. However, we are currently stuck on the bounding box selection issue. Any good suggestions? EDIT: BTW, I did try |
This PyTorch is from nvidia binary release
EDIT: This is the latest(maybe the last) binary release for Jetpack 5.1.4 (ubuntu20.04). |
Can you try |
No, on this board runing jetpack 5.1.3/5.1.4 there is only one release version torch-2.1.0a0+41361538.nv23.06-cp38-cp38-linux_aarch64.whl, which is in developer.download.nvidia.cn/compute/redist/jp/v512/pytorch/. Old versions(links) might be removed from their web server. I will try to export in x86 env and to see if it helps, and get back to you soon. Also I have made a request here: [REQUEST] build script for pytorch or up to date pytorh binary release supporting jetson boards running L4T35.6(ubuntu20.04) |
I think the versions I said also works on 5.1.3/5.1.4. Can you try? |
@marcoslucianops Do you mean it might be related with PyTorch version?
We don't have other Pytorch GPU version for L4T35.6 except 2.1. So only possible version is CPU versions which might be v2.0.0 or v1.14.0, which you are refering. So I'm going to use x86(ubuntu 22.04) latest version on laptop to export pt file if possible, which will not mess up the current jetson env(I did mess up the env just couple of weeks before). EDIT: x86 donwloading torch-2.5.1-cp310-cp310-manylinux1_x86_64.whl, well, it's time consuming... :( |
Well, don't have much time today. But ... ... It's great! It works using below exported onnx file to my jetson orin board.
$ python3 ./utils/export_yoloV8.py -w yolov8s.pt --dynamic
Note1: PS: The above test takes quite a lot of time for downloading python components. EDIT: Should not use
|
Thanks, @lida2003 for your comment and for sharing this approach! This version works pretty fine for me as well. I'm grateful for your insights—it's been super helpful. By the way, could you share more details about how you exported this ONNX version? I'm curious if there are specific steps or tweaks you used that made it work so well. Also, just to share my setup, this ONNX file worked for me with the following PyTorch version:
Here’s my Jetson setup:
Looking forward to your thoughts! |
No special steps, just use 2.5.1 on x86 to export the onnx as guide said. But I have found some issue related with BYTETrack here: #605. Not sure if it's related with onnx file.
@marcoslucianops we have the above results, what root cause might be? I'm trying build jetson orin pytorch 2.5.1, still some issues now. |
@marcoslucianops Hi I could not get this to show the detection boxes at all. I donno where i am going wrong. As suggested by you, I have tried torch 2.0 and 1.4 still could not get it to work. I am on Jetson orin nano 8 gb, deepstream-app version 6.2.0, DeepStreamSDK 6.2.0, |
@sdhamodaran It might be something to do with onnxruntime. Check onnxruntime version into account also. |
Thanks for your reply. I have '1.19.2' version of onnxruntime. Which one you have? |
@sdhamodaran I use onnxruntime_gpu-1.17.0-cp38-cp38-linux_aarch64.whl, which comes from nvidia, on jetson orin. I have upgraded pytorch to pytorch-v2.5.1+l4t35.6-cp38-cp38-aarch64 on jetson orin L4T 35.6 Jetpack 5.1.4. And it seems still got trouble. Right now, I'm compiling onnxruntime 1.19.2, still got issue on jetson orin L4T 35.6 Jetpack 5.1.4. BTW, you can export onnx file using x86 or try @PaoXi 's combination torch: 1.14.0a0+44dac51c.nv23.2 + torchvision: 0.14. And we have uploaded those onnx file, which is OK for us, you can try. |
Have you tried the last exported file by @lida2003 ? Try exporting the ONNX file on an x86 architecture workstation using the latest versions of CUDA, PyTorch, and ONNX Runtime. Then, use the generated ONNX file to export the engine file. |
Also a general question. When you guys try different combos of torch, torchvision and onnxruntime, from which step do you repeat the process from. I am starting from creation of onnx file and then running deepstream-app command. Am i doing it right? |
YES, but I didn't know the rootcause. Probably onnxruntime, Need to confirm. When x86 onnx file works, i thought it was pytorch version issue. After ugraded to 2.5.1 on jetson, I found it will still fail exporting on jetson, so I think it might be onnxruntime now. Anyway, x86's onnx file works. We need more tests or more experienced person to take time and effort to find out rootcause.
YES, you are absolutely right. |
@marcoslucianops As you can see in my first post of this thread, I didn't noticed onnxruntime. But I checked my system, there are two onnxruntime w/o gpu. How can I distinguish whether the application is actually using
|
The issue is quite similar to #390, but I need to use latest up to date versions.
Here is the video when I test yolov8s: https://drive.google.com/file/d/1I5MGC9_91h0drNASEM2z9VQUDptLNW_4/view?usp=drive_link
yolov4 is OK, https://drive.google.com/file/d/1bIdyqcfNa6JbuOyBR6NYOjnPqPTp-o-m/view?usp=sharing
The text was updated successfully, but these errors were encountered: