Hi Hailo community,
I’m seeing what looks like non-deterministic inference output on Hailo8 when running the exact same input multiple times.
Environment
-
Device: Hailo-8
-
Platform: Raspberry Pi CM4 / Linux
-
API: HailoRT C++ InferModel + ConfiguredInferModel
-
Batch size: 1
-
Inference mode: synchronous (
configured_model->run())
What I Already Tried
Switched from async (run_async) to sync (configured_model->run())
Ensured batch=1 and same binding is used
Removed preprocessing completely, forcing the same static input buffer
Increased timeout
Verified model works normally on other pipelines / sensors (issue appears only in this specific case)
Questions
-
Is deterministic output guaranteed for Hailo8 for identical input (same bytes)?
-
Could this be related to buffer allocation method (
mmap + MemoryView) or cache/DMA coherency on ARM (CM4)? -
Should I always use
DmaMappedBuffer/ VStreams instead of rawmmapbuffers? -
Any known issues if requesting output as
FLOAT32when HEF output isUINT8(format conversion path)?