Please use this identifier to cite or link to this item: https://hdl.handle.net/2440/122927
Citations
Scopus Web of Science® Altmetric
?
?
Full metadata record
DC FieldValueLanguage
dc.contributor.authorNekrasov, V.-
dc.contributor.authorDharmasiri, T.-
dc.contributor.authorSpek, A.-
dc.contributor.authorDrummond, T.-
dc.contributor.authorShen, C.-
dc.contributor.authorReid, I.-
dc.contributor.editorHoward, A.-
dc.contributor.editorAlthoefer, K.-
dc.contributor.editorArai, F.-
dc.contributor.editorArrichiello, F.-
dc.contributor.editorCaputo, B.-
dc.contributor.editorCastellanos, J.-
dc.contributor.editorHauser, K.-
dc.contributor.editorIsler, V.-
dc.contributor.editorKim, J.-
dc.contributor.editorLiu, H.-
dc.contributor.editorOh, P.-
dc.contributor.editorSantos, V.-
dc.contributor.editorScaramuzza, D.-
dc.contributor.editorUde, A.-
dc.contributor.editorVoyles, R.-
dc.contributor.editorYamane, K.-
dc.contributor.editorOkamura, A.-
dc.date.issued2019-
dc.identifier.citationIEEE International Conference on Robotics and Automation, 2019 / Howard, A., Althoefer, K., Arai, F., Arrichiello, F., Caputo, B., Castellanos, J., Hauser, K., Isler, V., Kim, J., Liu, H., Oh, P., Santos, V., Scaramuzza, D., Ude, A., Voyles, R., Yamane, K., Okamura, A. (ed./s), vol.2019-May, pp.7101-7107-
dc.identifier.isbn153866027X-
dc.identifier.isbn9781538660270-
dc.identifier.issn1050-4729-
dc.identifier.issn2577-087X-
dc.identifier.urihttp://hdl.handle.net/2440/122927-
dc.description.abstractDeployment of deep learning models in robotics as sensory information extractors can be a daunting task to handle, even using generic GPU cards. Here, we address three of its most prominent hurdles, namely, i) the adaptation of a single model to perform multiple tasks at once (in this work, we consider depth estimation and semantic segmentation crucial for acquiring geometric and semantic understanding of the scene), while ii) doing it in real-time, and iii) using asymmetric datasets with uneven numbers of annotations per each modality. To overcome the first two issues, we adapt a recently proposed real-time semantic segmentation network, making changes to further reduce the number of floating point operations. To approach the third issue, we embrace a simple solution based on hard knowledge distillation under the assumption of having access to a powerful `teacher' network. We showcase how our system can be easily extended to handle more tasks, and more datasets, all at once, performing depth estimation and segmentation both indoors and outdoors with a single model. Quantitatively, we achieve results equivalent to (or better than) current state-of-the-art approaches with one forward pass costing just 13ms and 6.5 GFLOPs on 640×480 inputs. This efficiency allows us to directly incorporate the raw predictions of our network into the SemanticFusion framework [1] for dense 3D semantic reconstruction of the scene.-
dc.description.statementofresponsibilityVladimir Nekrasov, Thanuja Dharmasiri, Andrew Spek, Tom Drummond, Chunhua Shen and Ian Reid-
dc.language.isoen-
dc.publisherIEEE-
dc.relation.ispartofseriesIEEE International Conference on Robotics and Automation ICRA-
dc.rights©2019 IEEE-
dc.source.urihttps://ieeexplore.ieee.org/xpl/conhome/8780387/proceeding-
dc.titleReal-time joint semantic segmentation and depth estimation using asymmetric annotations-
dc.typeConference paper-
dc.contributor.conferenceIEEE International Conference on Robotics and Automation (ICRA) (20 May 2019 - 24 May 2019 : Montreal, Canada)-
dc.identifier.doi10.1109/ICRA.2019.8794220-
dc.relation.granthttp://purl.org/au-research/grants/arc/CE140100016-
dc.relation.granthttp://purl.org/au-research/grants/arc/FL130100102-
pubs.publication-statusPublished-
dc.identifier.orcidNekrasov, V. [0000-0001-9653-7539]-
dc.identifier.orcidReid, I. [0000-0001-7790-6423]-
Appears in Collections:Aurora harvest 8
Computer Science publications

Files in This Item:
There are no files associated with this item.


Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.