Hi all,
I have successfully set up my pipeline using customized data and tested it across different model sizes: YOLOv8n, YOLOv8s, and YOLOv8m.
When testing the performance from a single-channel stream to multi-channel streams, I observed the following behavior:
- The inference FPS displayed in hailortcli monitor is fixed at 4 FPS for all models.
- However, the total FPS (as displayed in tools like
fpsdisplaysink
) remains at around 19-20 FPS. - The CPU utilization varies depending on the model size, but the inference FPS does not change.
It seems the inference frame rate is being locked at a certain value, regardless of the model size or input configuration.
I’m wondering if this is due to:
- Internal HailoRT limitations or specific settings.
- Queue or buffer configurations in my GStreamer pipeline.
- Resource constraints or an implicit frame rate cap.
Has anyone encountered similar behavior? Any suggestions or insights on where to look to resolve this would be greatly appreciated.