This project aims to employ sensor fusion techniques utilizing data from lidar and camera sensors within the KITTI dataset to generate an informative birds-eye-view occupancy grid map enriched with semantic details
Key Features • Download • How To Use • Credits • License
-
Bird's Eye View Occupancy Grid Map Generation: Creation of a comprehensive occupancy grid map from LiDAR data providing a top-down view.
-
Semantic Segmentation Mask Prediction using BiSeNetv2: Employing BiSeNetv2 for predicting semantic segmentation masks of the identified objects.
-
Point Cloud Projection onto Image for Semantic Information: Projection of the point cloud overlapping the camera's field of view onto images to extract semantic information about objects.
-
DBSCAN-Based Object Clustering in Point Cloud: Implementation of DBSCAN for clustering objects within the point cloud data.
-
Temporal Tracking of Objects for Semantic Identity Retention: Tracking the identified objects in a temporal domain, ensuring semantic identity retention even if objects move out of the camera's field of view.
- Set the path to the downloaded KITTI dataset files in the
config/settings.yaml
file. - Set Bisenet Weights Path in
config/settings.yaml
file. - Execute the
main.py
file to start the project.python3 main.py
This project is licensed under the MIT License - see the LICENSE file for details.