This letter presents an approach for semantic place categorization using data obtained from RGB cameras. Previous studies on visual place recognition and classification have shown that by considering features derived from pretrained convolutional neural networks (CNNs) in combination with part-based classification models, high recognition accuracy can be achieved, even in the presence of occlusions and severe viewpoint changes. Inspired by these works, we propose to exploit local deep representations, representing images as set of regions applying a Naïve Bayes nearest neighbor (NBNN) model for image classification. As opposed to previous methods, where CNNs are merely used as feature extractors, our approach seamlessly integrates the NBNN model into a fully CNN. Experimental results show that the proposed algorithm outperforms previous methods based on pretrained CNN models and that, when employed in challenging robot place recognition tasks, it is robust to occlusions, environmental and sensor changes.

Learning Deep NBNN Representations for Robust Place Categorization

Mancini, Massimiliano;Ricci, Elisa;Caputo, Barbara
2017-01-01

Abstract

This letter presents an approach for semantic place categorization using data obtained from RGB cameras. Previous studies on visual place recognition and classification have shown that by considering features derived from pretrained convolutional neural networks (CNNs) in combination with part-based classification models, high recognition accuracy can be achieved, even in the presence of occlusions and severe viewpoint changes. Inspired by these works, we propose to exploit local deep representations, representing images as set of regions applying a Naïve Bayes nearest neighbor (NBNN) model for image classification. As opposed to previous methods, where CNNs are merely used as feature extractors, our approach seamlessly integrates the NBNN model into a fully CNN. Experimental results show that the proposed algorithm outperforms previous methods based on pretrained CNN models and that, when employed in challenging robot place recognition tasks, it is robust to occlusions, environmental and sensor changes.
File in questo prodotto:
File Dimensione Formato  
ManRotRic_ARXIV_RAL_2017.pdf

solo utenti autorizzati

Descrizione: articolo principale
Tipologia: Documento in Post-print
Licenza: PUBBLICO - Pubblico con Copyright
Dimensione 783.06 kB
Formato Adobe PDF
783.06 kB Adobe PDF   Visualizza/Apri   Richiedi una copia

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/11582/310290
Citazioni
  • ???jsp.display-item.citation.pmc??? ND
social impact