Using YOLOv5s to detect each person in the frame and use AlphaPose to get skeleton-pose and then use ST-GCN model to predict action from every 30 frames of each person tracks.
Which now support 7 actions: Standing, Walking, Sitting, Lying Down, Stand up, Sit down, Fall Down.
Using modfied YOLOv5 to detect helmet, cell phone, and smoking.
- Python >= 3.6
- torch >= 1.7.1
- opencv-python >= 4.1.2
You can also use "pip install -r requirements.txt" directly.
Original test run on: i7-8750H CPU @ 2.20GHz x12, GeForce RTX 2060 6GB, CUDA 10.2
Train with rotation augmented COCO person keypoints dataset for more robust person detection in a variant of angle pose.
For actions recognition used data from Le2i Fall detection Dataset (Coffee room, Home) extract skeleton-pose by AlphaPose and labeled each action frames by hand for training ST-GCN model.
-
open ld.so.conf file
sudo gedit /etc/ld.so.conf
-
add HikVision SDK lib path in new line
Project_path/HiKcamSDK/lib
-
update .so library list
sudo ldconfig
- Download all pre-trained models into ./Models folder.
- Run my_multi_thread.py or Run image_input.py
python my_multi_thread.py
- AlphaPose : https://github.com/Amanbhandula/AlphaPose
- ST-GCN : https://github.com/yysijie/st-gcn