"Construction-Hazard-Detection" is an AI-powered tool designed to enhance safety on construction sites. By leveraging the YOLO model for object detection, it identifies potential hazards such as:
- Workers without helmets
- Workers without safety vests
- Workers near machinery or vehicles
- Workers in restricted areas, restricted areas will be automatically generated by computing and clustering the coordinates of safety cones.
Post-processing algorithms further improve detection accuracy. The system is built for real-time deployment, offering instant analysis and alerts for identified hazards.
Additionally, the system integrates AI recognition results in real-time via a web interface. It can send notifications and real-time on-site images through messaging apps like LINE, Messenger, WeChat, and Telegram for prompt alerts and reminders. The system also supports multiple languages, enabling users to receive notifications and interact with the interface in their preferred language. Supported languages include:
- 🇹🇼 Traditional Chinese (Taiwan)
- 🇨🇳 Simplified Chinese (Mainland China)
- 🇫🇷 French
- 🇬🇧 English
- 🇹🇭 Thai
- 🇻🇳 Vietnamese
- 🇮🇩 Indonesian
This multi-language support makes the system accessible to a global audience, improving usability across different regions.
- Hazard Detection Examples
- Usage
- Additional Information
- Dataset Information
- Contributing
- Development Roadmap
- License
Below are examples of real-time hazard detection by the system:
Before running the application, you need to configure the system by specifying the details of the video streams and other parameters in a YAML configuration file. An example configuration file configuration.yaml
should look like this:
# This is a list of video configurations
- video_url: "rtsp://example1.com/stream" # URL of the video
site: "Construction Site" # Location of the monitoring system
stream_name: "Front Gate" # Number of the monitor
model_key: "yolo11n" # Model key for the video
notifications: # List of line tokens, and languages for notifications
line_token_1: language_1
line_token_2: language_2
detect_with_server: True # Run objection detection with server
expire_date: "2024-12-31T23:59:59" # Expire date in ISO 8601 format
detection_items: # Detection items
detect_no_safety_vest_or_helmet: True # Detect if no safety vest or helmet is worn
detect_near_machinery_or_vehicle: True # Detect if a person is near machinery or vehicle
detect_in_restricted_area: True # Detect if a person is in a restricted area
- video_url: "streaming URL"
site: "Factory_1"
stream_name: "camera_1"
model_key: "yolo11n"
notifications:
line_token_3: language_3
line_token_4: language_4
detect_with_server: False
expire_date: "No Expire Date"
detection_items:
detect_no_safety_vest_or_helmet: True
detect_near_machinery_or_vehicle: False
detect_in_restricted_area: True
Each object in the array represents a video stream configuration with the following fields:
-
video_url
: The URL of the live video stream. This can include:- Surveillance streams
- RTSP streams
- Secondary streams
- YouTube videos or live streams
- Discord streams
-
site
: The location of the monitoring system (e.g., construction site, factory). -
stream_name
: The name assigned to the camera or stream (e.g., "Front Gate", "Camera 1"). -
model_key
: The key identifier for the machine learning model to use (e.g., "yolo11n"). -
notifications
: A list of LINE messaging API tokens and corresponding languages for sending notifications.line_token_1
,line_token_2
, etc.: These are the LINE API tokens.language_1
,language_2
, etc.: The languages for the notifications (e.g., "en" for English, "zh-TW" for Traditional Chinese).
For information on how to obtain a LINE token, please refer to line_notify_guide_en.
-
detect_with_server
: Boolean value indicating whether to run object detection using a server API. IfTrue
, the system will use the server for object detection. IfFalse
, object detection will run locally on the machine. -
expire_date
: Expire date for the video stream configuration in ISO 8601 format (e.g., "2024-12-31T23:59:59"). If there is no expiration date, a string like "No Expire Date" can be used. -
detection_items
: Specifies the safety detection items for monitoring specific scenarios. Each item can be set toTrue
to enable orFalse
to disable. The available detection items are:detect_no_safety_vest_or_helmet
: Detects if a person is not wearing a safety vest or helmet. This is essential for monitoring compliance with safety gear requirements on sites where such equipment is mandatory for personnel protection.detect_near_machinery_or_vehicle
: Detects if a person is dangerously close to machinery or vehicles. This helps prevent accidents caused by close proximity to heavy equipment or moving vehicles, often encountered in construction sites or industrial areas.detect_in_restricted_area
: Detects if a person has entered a restricted or controlled area. Restricted areas may be dangerous for untrained personnel or may contain sensitive equipment, so this setting aids in controlling access to such zones.
Now, you could launch the hazard-detection system in Docker or Python env:
Docker
To run the hazard detection system, you need to have Docker and Docker Compose installed on your machine. Follow these steps to get the system up and running:
-
Clone the repository to your local machine.
git clone https://github.com/yihong1120/Construction-Hazard-Detection.git
-
Navigate to the cloned directory.
cd Construction-Hazard-Detection
-
Build and run the services using Docker Compose:
docker-compose up --build
-
To run the main application with a specific configuration file, use the following command:
docker-compose run main-application python main.py --config /path/in/container/configuration.yaml
Replace
/path/in/container/configuration.yaml
with the actual path to your configuration file inside the container. -
To stop the services, use the following command:
docker-compose down
Python
To run the hazard detection system with Python, follow these steps:
-
Clone the repository to your local machine:
git clone https://github.com/yihong1120/Construction-Hazard-Detection.git
-
Navigate to the cloned directory:
cd Construction-Hazard-Detection
-
Install required packages:
pip install -r requirements.txt
-
Install and launch MySQL service (if required):
For Ubuntu users:
sudo apt install mysql-server sudo systemctl start mysql.service
For others, you can download and install MySQL that works in your operation system in this link.
-
Start user management API:
gunicorn -w 1 -b 0.0.0.0:8000 "examples.user_management.app:user-managements-app"
-
Run object detection API:
gunicorn -w 1 -b 0.0.0.0:8001 "examples.YOLO_server_api.app:app"
-
Run the main application with a specific configuration file:
python3 main.py --config /path/to/your/configuration.yaml
Replace
/path/to/your/configuration.yaml
with the actual path to your configuration file. -
Start the streaming web service:
For linux users:
gunicorn -w 1 -k eventlet -b 127.0.0.1:8002 "examples.streaming_web.app:streaming-web-app"
For windows users:
waitress-serve --host=127.0.0.1 --port=8002 "examples.streaming_web.app:streaming-web-app"
- The system logs are available within the Docker container and can be accessed for debugging purposes.
- The output images with detections (if enabled) will be saved to the specified output path.
- Notifications will be sent through LINE messaging API during the specified hours if hazards are detected.
- Ensure that the
Dockerfile
is present in the root directory of the project and is properly configured as per your application's requirements. - The
-p 8080:8080
flag maps port 8080 of the container to port 8080 on your host machine, allowing you to access the application via the host's IP address and port number.
For more information on Docker usage and commands, refer to the Docker documentation.
The primary dataset for training this model is the Construction Site Safety Image Dataset from Roboflow. We have enriched this dataset with additional annotations and made it openly accessible on Roboflow. The enhanced dataset can be found here: Construction Hazard Detection on Roboflow. This dataset includes the following labels:
0: 'Hardhat'
1: 'Mask'
2: 'NO-Hardhat'
3: 'NO-Mask'
4: 'NO-Safety Vest'
5: 'Person'
6: 'Safety Cone'
7: 'Safety Vest'
8: 'Machinery'
9: 'Vehicle'
Models for detection
Model | size (pixels) |
mAPval 50 |
mAPval 50-95 |
params (M) |
FLOPs (B) |
---|---|---|---|---|---|
YOLO11n | 640 | 54.1 | 31.0 | 2.6 | 6.5 |
YOLO11s | 640 | 70.1 | 44.8 | 9.4 | 21.6 |
YOLO11m | 640 | // | // | 20.1 | 68.0 |
YOLO11l | 640 | // | // | 25.3 | 86.9 |
YOLO11x | 640 | 76.8 | 52.5 | 56.9 | 194.9 |
Our comprehensive dataset ensures that the model is well-equipped to identify a wide range of potential hazards commonly found in construction environments.
We welcome contributions to this project. Please follow these steps:
- Fork the repository.
- Make your changes.
- Submit a pull request with a clear description of your improvements.
- Data collection and preprocessing.
- Training YOLO model with construction site data.
- Developing post-processing techniques for enhanced accuracy.
- Implementing real-time analysis and alert system.
- Testing and validation in simulated environments.
- Deployment in actual construction sites for field testing.
- Ongoing maintenance and updates based on user feedback.
- Add support for WhatsApp notifications,
- Switch from Flask to Fastapi in server_api and streaming_web
This project is licensed under the AGPL-3.0 License.