Skip to content

bharath5673/yolov5_BEV

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

14 Commits
 
 
 
 
 
 
 
 

Repository files navigation

YOLOv5 Object Detection with Bird's Eye View and Tracking

This project utilizes the YOLOv5 deep learning model to perform real-time object detection for Advanced Driver Assistance Systems (ADAS). It provides a framework for detecting and tracking objects in the context of automotive safety and driver assistance applications. it provides a Bird's Eye View (BEV) visualization, which offers a top-down perspective of the detected objects.

demo

Features

  • Real-time object detection using the YOLOv5 model.
  • Detection of various objects relevant to ADAS, such as vehicles, pedestrians, cyclists, and traffic signs.
  • Object tracking to maintain continuity and trajectory of detected objects.
  • Bird's Eye View (BEV) visualization of the detected objects in a simulated environment.
  • Customizable confidence threshold and class filtering.
  • Simulated environment provides an intuitive top-down view of object positions and movements.
  • Supports both image and video input for object detection and tracking.
  • Easy integration with pre-trained YOLOv5 models.
  • Provides bounding box coordinates, class labels, and tracking IDs for detected objects.

Prerequisites

  • Python 3.x
  • OpenCV
  • PyTorch
  • NumPy

Installation

  1. Clone this repository.
  2. Install the required dependencies
pip3 install torch opencv numpy

Usage

  1. Download pre-trained YOLOv5 weights or train your own model.
  2. Provide the path to the YOLOv5 weights in the code.
  3. Run the script with the video file.
  4. View the object detection results and Bird's Eye View visualization.

For more detailed usage instructions and options, refer to the project documentation.

Run

python3 yoloV5_sim.py

Contributing

Contributions are welcome! If you find any issues or have suggestions for improvements, please open an issue or submit a pull request.

License

This project is licensed under the MIT License. See the LICENSE file for details.

Acknowledgments

Releases

No releases published

Packages

No packages published

Languages