I’m currently working with YOLO-based models, like yolov11n for detection and yolov8s-pose for keypoints, on the Hailo-8L accelerator.
I have a couple of questions:
Is it feasible to load and run inference on two different .hef models (on the same input)—say, one for detection and another for pose estimation—sequentially on the same device?
Also, could you share any code examples or documentation on how to run compiled .hef models directly on Hailo hardware using HailoRT?
I’m really trying to get a better grasp on executing multiple models and managing the performance trade-offs involved.