Hello,
I noticed a clear difference in performance between my original PyTorch model (.pt) and the compiled Hailo Executable Format (.hef) version. While the .pt model works fine, the .hef model sometimes fails to detect objects or behaves differently in terms of inference.
For example, if you take a screenshot of the same image, the .hef model in production may detect nothing, whereas the .pt model detects all objects correctly.
I initially thought that the production model wasn’t receiving images of the same input size, but even after adjusting this, it didn’t improve. (It’s possible I did something incorrectly during this adjustment.)
One possible cause could be the optimization level used during compilation. I’m currently using level 0, and my local machine has no GPU, which might affect how certain operations are executed.
I’m trying to compile on Google Colab, which has NVIDIA GPUs, but the Hailo SDK doesn’t seem to detect the GPU, likely because it expects specific Hailo hardware (Hailo-8 or Hailo-15), not generic CUDA GPUs.
Additionally, I don’t fully understand the calibration dataset requirements: how different the images should be, how many images are needed, and whether I should only include images from my production cameras.
Thank you in advance.