Gesture Learning for Advanced Driver Assistance Systems
GLADAS is an open-sourced research platform for investigating Human-AV Interactions. Specifically, GLADAS is built for the training, testing, validation, and benchmarking of hand gesture recognition algorithms of self-driving cars. Although this repo provides an example algorithm for usage, research teams are free to test and benchmark their own algorithms.
Unreal Engine with AirSim Compatibility (link)
AirSim Package (installation tutorial)
AirSim-Python Package
pip install msgpack-rpc-python
(optional) 3D map for self-driving car research and driving
(optional) Pedestrian with Animated Gestures
BenchmarkMetrics: Measures algorithm performance.
*Includes support for Precision-Recall, PR Graphs, F1, Accuracy, TN, TP, FP, FN Metrics*
CarTest: Gives feedback and data on the hand gesture recognition algorithm's performance. Includes the Video Streamer. Resulting data saved to /data/ScenarioX.csv for each scenario.
CarSim: A real-time visualization of CarTest.
Classifier: The Gesture Classifier model used as an example for GLADAS testing. This, as long as most of CarTest, can be changed to fit each individual research team's requirements and specifications.
Full instructions and explanation of the code are coming soon.
Ethan Shaotran
Jonathan Cruz (Harvard University)
Vijay Janapa Reddi (Harvard University)
Ethan Shaotran (shaotran@mit.edu)