/assigning-visual-words-to-places

Assigning visual words to places for loop closure detection

Primary LanguageMATLABMIT LicenseMIT

Assigning Visual Words to Places for Loop Closure Detection

This open source MATLAB algorith presents an efficient image-to-sequence appearance-based loop closure detection method. Using a voting scheme over the on-line generated visual words and coupling the method with a probability function, the pipeline is able to accurately detect revisited places. Dynamic sequence segmentation is performed on the incoming image stream formulating “places” on the robot’s navigated path. Subsequently, the accumulated local feature descriptors are processed by a growing neural gas clustering mechanism for the corresponding visual words' generation. When new query images enter to the pipeline, the extracted descriptors assign votes to the database sequences including their nearest neighboring visual words. The system uses a binomial probability density function to locate the proper candidate place and a nearest descriptor neighbor technique to identify image-to-image associations within the selected loop closing sequence. Temporal and geometrical consistency checks are performed between the query and candidate image, providing a higher level of discrimination.

Note that the given framework is a research code. The authors are not responsible for any errors it may contain.

Conditions of use

Assigning-visual-words-to-places is distributed under the terms of the MIT License.

Related publication

The details of the algorithm are explained in the following publication:

Assigning Visual Words to Places for Loop Closure Detection
Konstantinos A. Tsintotas, Loukas Bampis, and Antonios Gasteratos
IEEE International Conference on Robotics and Automation (ICRA)

If you use this code, please cite:

@inproceedings{tsintotas2018places,
  title={Assigning Visual Words to Places for Loop Closure Detection},  
  author={K. A. Tsintotas and L. Bampis and A. Gasteratos},   
  booktitle={IEEE International Conference on Robotics and Automation (ICRA)},
  pages={5979 - 5985},
  year={2018},   
  month={May},
  address={Brisbane, QLD, Australia},
  doi={10.1109/ICRA.2018.8461146} 
}

Contact

If you have problems or questions using this code, please contact the author (e-mail address: ktsintot@pme.duth.gr, ktsintotas@icloud.com). Ground truth matrices for many of the used datasets are a kindly offer from Dorian Galvez-Lopez. Contributions are totally welcome.