Hi,
I’m working on a project that involves deploying a custom-trained YOLOv5s model on the Jetson Nano 4GB Developer Kit for real-time object detection.
Here’s a breakdown of the current setup:
-
Device: Jetson Nano (4GB)
-
JetPack Version: 4.6.1
-
Camera: Logitech C270 (USB)
-
Model Format: Converted YOLOv5s to ONNX
-
Software: PyTorch 1.11, ONNX Runtime 1.10, Python 3.6
What I’ve Done
-
Trained and exported YOLOv5s model to ONNX.
-
Confirmed ONNX model runs correctly on desktop.
-
Installed ONNX Runtime and ran inference on Jetson Nano.
-
Attempted optimization with TensorRT but didn’t see expected improvements.
What I Need Help With
-
Is there a recommended pipeline or script to run YOLOv5 ONNX models with TensorRT on Jetson Nano?
-
What are the best practices for converting ONNX models to TensorRT (especially layer compatibility)?
-
How can I enable FP16 or INT8 inference on the Nano?
-
What input image resolution and batch size are optimal for achieving real-time performance?
-
Are there memory-saving techniques to deal with Jetson Nano’s limited 4GB RAM?
I’m trying to get at least ~15 FPS for basic object detection with one camera input. I’ve seen mentions of DeepStream and other TensorRT wrappers — if those are better suited for this, I’d appreciate pointers to sample projects or official documentation.
Additional Info:
I’m using a standard Jetson Nano dev kit sourced locally from this page, which includes GPIO headers and accessories. Mentioning in case the hardware variant affects optimization.
Thanks in advance for your support!