Ensure safety and comfort of employees by monitoring ergonomics.
Ergonomics is the study of designing and arranging products, systems, and environments to fit the capabilities and limitations of people, with the goal of improving efficiency, safety, and comfort. The primary focus of ergonomics is to create environments that optimize human performance and well-being.
Vision AI based monitoring¶
Vision AI based monitors can be used to monitor productivity and workers health by providing real-time video feeds of different areas. These cameras can be used to monitor and track number of objects moved by persons from one place to the other, as well as the bending information of workers while performing this work.
Companies can put compliance policies in place to ensure that workers are made aware of work-related injuries and illnesses due to unnecessary bending.
It is important to note that these camera-based monitoring provides should be supplanted by strong compliance processes to ensure their accuracy and reliability. In addition, workers working in companies should always be trained on ergonomics and its significance for safety and its impact on human health.
VisionAI model's generated events would be:
- Bend count event per individual
Model training is carried out with Microsoft COCO: Common Objects in Context dataset. Person and Book classes are considered for model building. Person class is considered here because the problem of ergonomics is related to pose estimation and the object used to show movement is the book. It is the object carried by a person from one location to the other. Other objects can be considered as per the requirement.
The dataset is made up of a large number of images and it is curated to ensure a true representation of the real world for:
Indoor vs Outdoor environments
Variations in time
Different types of clothing for persons
Different distances from the camera
Various lighting conditions
Various camera angles, resolutions and calibrations
Using security camera feeds
The Yolov5 pre-trained model for detecting person and book (an example of an item) classes are used to build the model. DenseNet is employed to estimate each person's landmarks. These landmarks are used to estimate poses. This is mostly used to track a person's bending motion. To guarantee the counting of boxes, object tracking using a strong sort algorithm is also built.
This provides ergonomics data that may be utilised for a variety of tasks, such as alarm generation when the number of bends exceeds predetermined levels and productivity counting to determine how many objects were transported from one location to another.
The DenseNet Model for Landmark detection
|Dataset size||Version||Camera support||Precision||Recall||mAP|
The model is lightweight enough to be run on any edge device.
The business logic for this scenario is as follows:
We use existing camera feeds from the premises to monitor the presence of workers.
VisionAI system is able to run on edge devices. It uses camera feeds for processing.
We detect and track the number of objects transported and we monitor the total number of bending motions of a person while working.
=== "Test now with online Web-Cam" To test this model & scenario, you can use the following steps:
- Install the visionai package from PyPI ```console $ pip install visionai ``` - Test the scenario from your local web-cam ```console $ visionai scenario test ergonomics ``` Downloading models for scenario: ergonomics Starting scenario: ergonomics.. ``` - You should be able to see the events generated on your console window with the detections of firearms and knives event within the camera field of view.
VisionAI app is available at a Azure Market place, one can download and use it by following steps mentioned here
Training with custom data¶
The scenario is provided as part of our GPL-v3 package for VisionAI. If you wish to train this with custom datasets, please contact us and we can provide you with the training code. You can do custom training with your own datasets for free, as long as it complies with GPLv3 license (you give back the code to the community). If you are interested in a custom license, please [contact us] (../company/contact.md).