Run Clip application on Raspberrypi5 and Hailo8

CLIP(Contrastive Language-Image Pre-Training) is a neural network trained on a variety of (image, text) pairs. It can be instructed in natural language to predict the most relevant text snippet, given an image, without directly optimizing for the task, similarly to the zero-shot capabilities of GPT-2 and 3. We found CLIP matches the performance of the original ResNet50 on ImageNet “zero-shot” without using any of the original 1.28M labeled examples, overcoming several major challenges in computer vision.

Here is the link
If you like it please give a star.

In the video shown below, you can see that when I input “banana,” the CLIP model recognizes a banana, and when I input “apple,” the model recognizes an apple. You only need to input different words, and the CLIP model will recognize different objects.