How does this project perform in real life? I used YOLO for an ESP32-CAM-based object detection project. I saw that the performance was heavily dependent on light.
Jaryd; Thank You for the superb video/guide. Spot on!!!
Have you got YOLO11 converted to NCNN? Same tool?
Two years ago, these FPS were achievable with projects from the Coral project. What the heck can i do with that device!? Have to switch to the AIHAT?
What other rpi CV stuff is around? Your favourties, anywayâŚ
Please and thanks.
Hey @ahsrab292840 welcome to the forums!
This is going to depend a lot on your set-up. The default settings on the Pi 5 can give about 5 FPS and it isnât very dependent on light. In the written guide we go through a few ways to increase the processing speed of it, and we get a good 30 fps out of it, but it can also be pushed more.
This is also going to be a higher resolution and possibly more powerful model than what is run on your ESP32-CAM-based project.
Hey @Aeon293088, thank you for the kind words!
We were able to do all the conversion stuff with YOLO11 just fine. As of now, only models made by Ultralytics can be converted with their method (so YOLO5, YOLO8, and now YOLO11).
The Coral could still give this a run for its money! But there is a lot of optimisation and a few corners cut to get this FPS on the Pi 5âs silicon. The AI HAT is even more incredible as we were able to run some much larger and more powerful models at 30 fps and with only 10 Watts of power! However, both the AI HAT and a Coral TPU would have different workflows than what we cover in this guide, its not as simple as enabling the processing of this script to be done on it.
As for more guides, we have a guide on using the AI HAT if you wish to check out the workflow for it. It is a little more involved that this guide but still straight forward.
We have a pose estimation guide as well where we make space invaders controlled with your nose. If you have done this object recognition guide, you have already set up 90% of this guide.
We also have a face recognition guide. Its a bit low-tech compared to the YOLO models but can be useful none the less.
YOLO is really the creme of the crop here and I would also check out their YOLO11 page to see all the model types that are available. This guide set up a lot of the needed libraries to use them and it should be as easy as using their name (the pose guide is a good guide on how to do this). There are a few other tasks that it can do like describing whats in an image and segmenting it with vectors.
If you have any more questions let us know!
Thanks Man.
iâll tell ya, Jaryd: The Goal is actually emotion detection on the pi (to instantiate a kinetic process, like your solenoid).
Iâve found some yolo8 based emotion detection, but havenât âhit paydirtâ yet.
I tried using the World yolo, and prompting âhappy faceâ âemotion:happyâ etc etc⌠no luck there.
What do you think? Emotion recognition is an uncanny place. . .
I think your best route would be to find someone who has pre-trained a Yolo-based model for emotion detection. If you can get your hands on a model with a .pt model format, you should be able to just download it and put it into the project folder with all the other files. Then change the line in the main script to:
# Load YOLOv8
model = YOLO("best.pt")
And it should run! You can change the name of the model in the folder, and in this line if you need as well. Iâm using âbest.ptâ as an example as its the default output name of a trained model.
Iâd take a good look around but here is a YoloV5 implementation that might run a little slower than a v11 implementation - youâll find the model in the models folder, ignore the set up and running instructions you can use this guide and the method above.
There is also a file in this v11 repo called best.onnx. I think the set up from this guide can run ONNX? If not you should be able to use the NCNN conversion method to turn it into an NCNN model that it can run.
Let us know how you go! Itâs been on our list of videos to eventually get around to!
well guess what (x2)!?
- above code ran the best.onnx from that v11 repo, just downloaded itself some onnx stuff. 5fps.
- The modeling must have been very very deep, because it only sees âSadâ even when i fake a smile. . .
Iâmâa keep hunting.
Iâll try the onnx to NCNN conversion, jfk.
Let me know if you come across anything,please and kindly
Thanks. Once I succeed setting up my Raspberry Pi, I will give it a try.