Please use this identifier to cite or link to this item: http://10.1.7.192:80/jspui/handle/123456789/6227
Full metadata record
DC FieldValueLanguage
dc.contributor.authorSompura, Dipesh-
dc.date.accessioned2015-09-26T07:59:36Z-
dc.date.available2015-09-26T07:59:36Z-
dc.date.issued2015-06-01-
dc.identifier.urihttp://hdl.handle.net/123456789/6227-
dc.description.abstractObject detection performance, as measured on the canonical PASCAL VOC dataset, has plateaued in the last few years. The best-performing methods are complex ensemble systems that typically combine multiple low-level image features with high-level context. I implement scalable detection algorithm that improves mean average precision (mAP). Our approach combines two key insights: (1) one can apply high-capacity convolutional neural networks (CNNs) to bottom-up region proposals in order to localize and segment objects and (2) when labeled training data is scarce, supervised pre-training for an auxiliary task, followed by domain-specific fine-tuning, yields a significant performance boost. Since we combine region proposals with CNNs, we call our method R-CNN: Regions with CNN features. We also compare R-CNN to OverFeat, a recently proposed sliding-window detector based on a similar CNN architecture. Finally my goal is to recognition of realistic scene objects successfully, Such as Airplane, bicycle,cat,person etc.en_US
dc.publisherInstitute of Technologyen_US
dc.relation.ispartofseries12MICT48;-
dc.subjectComputer 2013en_US
dc.subjectProject Report 2013en_US
dc.subjectComputer Project Reporten_US
dc.subjectProject Reporten_US
dc.subject12MICTen_US
dc.subject12MICT48en_US
dc.subjectICTen_US
dc.subjectICT 2013en_US
dc.titleVisual Object Recognition Using Region Convolutional Neural Network (RCNN)en_US
dc.typeDissertationen_US
Appears in Collections:Dissertation, CE (ICT)

Files in This Item:
File Description SizeFormat 
12MICT48.pdf12MICT482.82 MBAdobe PDFThumbnail
View/Open


Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.