English   español  
Please use this identifier to cite or link to this item: http://hdl.handle.net/10261/180083
Share/Impact:
Statistics
logo share SHARE   Add this article to your Mendeley library MendeleyBASE
Visualizar otros formatos: MARC | Dublin Core | RDF | ORE | MODS | METS | DIDL
Exportar a otros formatos:

Title

Tracking and approaching persons using deep learning techniques

AuthorsLaplaza Galindo, Javier
AdvisorSanfeliu, Alberto
Issue Date17-Sep-2018
PublisherCSIC-UPC - Instituto de Robótica e Informática Industrial (IRII)
Universidad Politécnica de Cataluña
AbstractThis project proposes a solution in order to enable a social robot to approach and follow one specific person using a vision based system. The idea is to introduce this features to the robot so that it will be able, int he futur, to autonomously interact with people. To do so, the algorithms developed in the project use Convolutional Neural Networks to identify where is the person given an image provided by the robot camera. In order to accomplish these tasks, two algorithms are created: an object detector using YOLO algorithm and an object tracker using a Siamese network. In order to fully understand how these algorithms work, the methods and architectures on which they are based are explained. Also, one of the algorithms is attempted to train with Google Colaboratory, even though the training results aren’t used in the final implementation. Different recordings are filmed teleoperating the robot simulating real approaching and following operations. Each recording is then labelled frame by frame to use this data in the training. Since the results of the training aren’t used in the final implementation, the recordings are used to test how well the final implementation is able to track the target in a frame. Therefore, different metrics are evaluated in the recordings, separating the ”approaching operation” videos from the ”following operation” videos, thus obtaining separated results for the two operations. Also, since the recordings where filmed considering different light conditions, it is possible to analyze how light variations affect the results when performing both approaching and following operations. These algorithms are run and tested in a Jetson TX2, using the GPU of the embedded device to enhance the performance of the algorithms. The implementation is build using PyTorch
Publisher version (URL)https://upcommons.upc.edu/handle/2117/127124
URIhttp://hdl.handle.net/10261/180083
Appears in Collections:(IRII) Tesis
Files in This Item:
File Description SizeFormat 
tfm-javier-laplaza-2018-1.pdf10,43 MBAdobe PDFThumbnail
View/Open
Show full item record
Review this work
 


WARNING: Items in Digital.CSIC are protected by copyright, with all rights reserved, unless otherwise indicated.