Mostrar el registro sencillo del ítem

dc.contributor.authorMallick, Arijit
dc.contributor.otherCervera Mateu, Enric
dc.contributor.otherKermorgant, Olivier
dc.contributor.otherUniversitat Jaume I. Departament d'Enginyeria i Ciència dels Computador
dc.date.accessioned2018-04-24T10:13:43Z
dc.date.available2018-04-24T10:13:43Z
dc.date.issued2017-07-13
dc.identifier.urihttp://hdl.handle.net/10234/174287
dc.descriptionTreball de Final de Màster Universitari Erasmus Mundus en Robòtica Avançada. Curs acadèmic 2016-2017ca_CA
dc.description.abstractIn the light of current advancement in deep learning, robot vision is not an exception. Many popular machine learning algorithms has already been proposed and implemented to solve intricate computer vision problems. The same has not been in the case of robot vision. Due to real time constraints and dynamic nature of environment such as illumination and processing power, very few algorithms are able to solve the object recognition problem at large. The primary objective of the thesis project is to converge into an accurate working algorithm for object recognition in a cluttered scene and subsequently helping the BAXTER robot to pick up the correct object among the clutter. Feature matching algorithms usually fail to identify most of the object having no texture, hence deep learning has been employed for better performance. The next step is to look for the object and localize it within the image frame. Although basic shallow Convolutional Neural Network easily identifies the presence of an object within a frame, it is very difficult to localize the object location within the frame. This work primarily focuses on finding a solution for an accurate localization. The first solution which comes to mind is to produce a bounding box surrounding the object. In literature, YOLO is found to be providing a very robust result on existing datasets. But this was not the case when it was tried on new objects belonging to the current thesis project work. Due to high inaccuracy and presence of a huge redundant area within the bounding box, an algorithm was needed which will segment the object accurately and make the picking task easier. This was done through semantic segmentation using deep CNNs. Although time consuming, RESNET has been found to be very efficient as its post processed output helps to identify items in a significantly difficult task environment. This work has been done in light of upcoming AMAZON robotic challenge where the robot successfully classified and distinguished everyday items from a cluttered scenario. In addition to this, a performance analysis study has also been done comparing YOLO and RESNET justifying the usage of the later algorithm with the help of performance metrics such IOU and ViG.ca_CA
dc.format.extent49 p.ca_CA
dc.format.mimetypeapplication/pdfca_CA
dc.language.isoengca_CA
dc.publisherUniversitat Jaume Ica_CA
dc.rightsAttribution-NonCommercial-NoDerivatives 4.0 Internacional*
dc.rights.urihttp://creativecommons.org/licenses/by-nc-nd/4.0/*
dc.subjectMàster Universitari Erasmus Mundus en Robòtica Avançadaca_CA
dc.subjectErasmus Mundus University Master's Degree in Advanced Roboticsca_CA
dc.subjectMáster Universitario Erasmus Mundus en Robótica Avanzadaca_CA
dc.titleDeep Learning for Object Recognition in picking tasksca_CA
dc.typeinfo:eu-repo/semantics/masterThesisca_CA
dc.educationLevelEstudios de Postgradoca_CA
dc.rights.accessRightsinfo:eu-repo/semantics/openAccessca_CA
dc.audience.educationlevelEstudios de Postgrado


Ficheros en el ítem

Thumbnail
Thumbnail

Este ítem aparece en la(s) siguiente(s) colección(ones)

Mostrar el registro sencillo del ítem

Attribution-NonCommercial-NoDerivatives 4.0 Internacional
Excepto si se señala otra cosa, la licencia del ítem se describe como: Attribution-NonCommercial-NoDerivatives 4.0 Internacional