Hey @hyperwolf,
To run multiple models on the Hailo-8, we use the scheduling interface. For each HEF file, we create a separate HailoAsyncInference
class and run them using a scheduler.
Here’s an example in C++: Hailo Scheduler Example
And here’s the async object detection class in Python: Async Object Detection Class
By creating instances of this class for each HEF and running them with the HailoSchedulingAlgorithm
of your choice, you can effectively run two or more models on the chip simultaneously.
Regards