/tfg_mbenavent

Primary LanguagePythonGNU General Public License v3.0GPL-3.0

Context-Based Action Estimation with YOLO

This repository contains the developments started in the Bacherlor's Degree in Computer Science of Manuel Benavent-Lledo and extended for his Master's Thesis. The advances have been also published in international conferences.

The aim of this project is to provide a context-based action estimation based on the recognition of the objects and the hands in the scene.

Project Structure

The folders ADL, EGO-DAILY and EPIC-KITCHENS (we use Epic-Kitchens 55) contain the annotations and scripts to extract the relevant ones for those datasets, used to train YOLO models.

The MIT_INDOOR folder contains the annotations and code for fine-tuning and running VGG16 architecure for scene recognition.

The Pipeline folder contains the YOLO architecture and the action estimation architecture, further details are provided when accessing the folder.

Docker

A docker image and launch script is provided to run this architecture.

Citations

The following papers have been published based on the different versions of the project:

@INPROCEEDINGS{9892910,
  author={Benavent-Lledo, Manuel and Oprea, Sergiu and Castro-Vargas, John Alejandro and Mulero-Perez, David and Garcia-Rodriguez, Jose},
  booktitle={2022 International Joint Conference on Neural Networks (IJCNN)}, 
  title={Predicting Human-Object Interactions in Egocentric Videos}, 
  year={2022},
  volume={},
  number={},
  pages={1-7},
  doi={10.1109/IJCNN55064.2022.9892910}}

Author

Manuel Benavent-Lledo (mbenavent@dtic.ua.es)