Please use this identifier to cite or link to this item: https://hdl.handle.net/2440/129989
Citations
Scopus Web of Science® Altmetric
?
?
Full metadata record
DC FieldValueLanguage
dc.contributor.authorDuan, X.-
dc.contributor.authorWu, Q.-
dc.contributor.authorGan, C.-
dc.contributor.authorZhang, Y.-
dc.contributor.authorHuang, W.-
dc.contributor.authorVan Den Hengel, A.-
dc.contributor.authorZhu, W.-
dc.date.issued2019-
dc.identifier.citationProceedings of the 27th ACM International Conference on Multimedia (ACM Multimedia 2019), MM '19, 2019, pp.1543-1551-
dc.identifier.isbn1450368891-
dc.identifier.isbn9781450368896-
dc.identifier.urihttp://hdl.handle.net/2440/129989-
dc.description.abstractHumans have a surprising capacity to induce general rules that describe the specific actions portrayed in a video sequence. The rules learned through this kind of process allow us to achieve similar goals to those shown in the video but in more general circumstances. Enabling an agent to achieve the same capacity represents a significant challenge. In this paper, we propose a Watch-Reason-Code (WRC) model to synthesise programs that describe the process carried out in a set of video sequences. The ‘watch’ stage is simply a video encoder that encodes videos to multiple feature vectors. The ‘reason’ stage takes as input the features from multiple diverse videos and generates a compact feature representation via a novel deviation-pooling method. The ‘code’ stage is a multi-round decoder that the first step leverages to generate a draft program layout with possible useful statements and perceptions. Further steps then take these outputs and generate a fully structured, compile-able and executable program. We evaluate the effectiveness of our model in two video-to-program synthesis environments, Karel and ViZdoom, showing that we can achieve the state-of-the-art under a variety of settings.-
dc.description.statementofresponsibilityXuguang Duan, Qi Wu, Chuang Gan, Yiwei Zhang, Wenbing Huang, Anton Van Den Hengel, Wenwu Zhu-
dc.language.isoen-
dc.publisherAssociation for Computing Machinery-
dc.rights© 2019 Association for Computing Machinery.-
dc.source.urihttps://dl.acm.org/doi/proceedings/10.1145/3343031-
dc.subjectvideo understanding; video embedding; video to program translation-
dc.titleWatch, reason and code: Learning to represent videos using program-
dc.typeConference paper-
dc.contributor.conference27th ACM International Conference on Multimedia (ACM Multimedia) (21 Oct 2019 - 25 Oct 2019 : Nice, France)-
dc.identifier.doi10.1145/3343031.3351094-
dc.publisher.placeonline-
dc.relation.granthttp://purl.org/au-research/grants/arc/DE190100539-
pubs.publication-statusPublished-
dc.identifier.orcidWu, Q. [0000-0003-3631-256X]-
dc.identifier.orcidVan Den Hengel, A. [0000-0003-3027-8364]-
Appears in Collections:Aurora harvest 4
Computer Science publications

Files in This Item:
File Description SizeFormat 
hdl_129989.pdfAccepted version1.71 MBAdobe PDFView/Open


Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.