Welcome to my project on Object Detection and Tracking using the latest YOLOv12 model! 🦾 YOLO (You Only Look Once) is one of the most powerful object detection models, and I’ve integrated it with live video feed tracking for a seamless and dynamic experience.
This project combines real-time object detection and tracking:
-
YOLOv12 detects objects in live frames.
-
Users can manually select an object to track using CSRT Tracker.
-
The system continues tracking the selected object and provides a live overlay, even as it moves.
YOLOv12 is the latest version of the YOLO series and was released just 4 days ago (🔥 cutting-edge technology!). It comes with improvements in:
• ⚡ Speed: Real-time performance with a balance between precision and speed.
• 🧠 Accuracy: Better object localization and class prediction.
• 📊 Versatility: Pretrained on the COCO dataset, making it a strong general-purpose model.
The COCO (Common Objects in Context) dataset is a benchmark dataset in computer vision, containing 80 diverse classes such as:
• 🏃 Person
• 🚗 Car
• 🐕 Dog
• 📱 Cell phone
• 🍕 Pizza
Many more!
Instead of creating a custom dataset, I utilized COCO because it provides a rich set of labeled data that perfectly aligns with the needs of this project. Custom datasets are useful for domain-specific tasks, but COCO's versatility made it ideal for this general application.
-
Live Video Feed 🎥: The system captures live video frames using OpenCV.
-
Object Detection 🧐: YOLOv12 processes each frame to detect objects and overlay bounding boxes.
-
Manual Object Selection 🖱️: Click on the object of interest, and the CSRT tracker initializes.
-
Tracking 🎯: The tracker updates the object's position frame by frame, ensuring dynamic tracking.
• YOLOv12 for object detection.
• CSRT Tracker for robust object tracking.
• OpenCV for video processing and visualization.
• Real-Time Processing: Instantaneous detection and tracking.
• Dynamic Tracking: Track any object selected in the live feed.
• Scalable: Can be extended to custom datasets if required.
• User-Friendly: Interactive and intuitive interface.
-
Clone the repository and ensure dependencies are installed.
-
Run the Python script.
-
Use the live video feed window to select an object by clicking on it.
-
Watch as the system tracks your selected object in real-time!
Below is a demonstration of how the application works: