Show simple item record

dc.contributor.authorBenotti, Luciana
dc.contributor.authorLau, Tessa
dc.contributor.authorVillalba, Martín Federico
dc.date.accessioned2023-03-22T15:00:04Z
dc.date.available2023-03-22T15:00:04Z
dc.date.issued2010
dc.identifier.urihttp://hdl.handle.net/11086/546748
dc.descriptionArtículo publicado finalmente en : ACM Transactions on Interactive Intelligent Systems, Vol. 4, No. 3, Article 13, Publication date: July 2014.es
dc.description.abstractWe define the problem of automatic instruction interpretation as follows. Given a natural language instruc- tion, can we automatically predict what an instruction follower, such as a robot, should do in the environment to follow that instruction? Previous approaches to automatic instruction interpretation have required either extensive domain-dependent rule writing or extensive manually annotated corpora. This article presents a novel approach that leverages a large amount of unannotated, easy-to-collect data from humans inter- acting in a game-like environment. Our approach uses an automatic annotation phase based on artificial intelligence planning, for which two different annotation strategies are compared: one based on behavioral information and the other based on visibility information. The resulting annotations are used as training data for different automatic classifiers. This algorithm is based on the intuition that the problem of inter- preting a situated instruction can be cast as a classification problem of choosing among the actions that are possible in the situation. Classification is done by combining language, vision, and behavior information. Our empirical analysis shows that machine learning classifiers achieve 77% accuracy on this task on avail- able English corpora and 74% on similar German corpora. Finally, the inclusion of human feedback in the interpretation process is shown to boost performance to 92% for the English corpus and 90% for the German corpus.es
dc.description.urihttp://dl.acm.org/citation.cfm?id=2629632
dc.language.isoenges
dc.relation.urihttps://doi.org/10.1145/2629632
dc.rightsAttribution-NonCommercial-ShareAlike 4.0 International*
dc.rights.urihttps://creativecommons.org/licenses/by-nc-sa/4.0/*
dc.subjectNatural language processinges
dc.subjectNatural language interpretationes
dc.subjectMulti-modal understandinges
dc.subjectAction recognitiones
dc.subjectVisual feedbackes
dc.subjectSituated virtual agentes
dc.subjectUnsupervised learninges
dc.titleInterpreting natural language instructions using language, vision and behaviores
dc.typearticlees
dc.description.versionacceptedVersionen
dc.description.filFil: Benotti, Luciana. Universidad Nacional de Córdoba. Facultad de Matemática, Astronomía y Física; Argentina.es
dc.description.filFil: Benotti, Luciana. Consejo Nacional de Investigaciones Científicas y Técnicas; Argentina.es
dc.description.filFil: Lau, Tessa. Savioke Incorporation; United States of America.en
dc.description.filFil: Villalba, Martín Federico. University of Potsdam; Germany.en
dc.journal.countryEstados Unidoses
dc.description.fieldCiencias de la Computación
dc.contributor.orcidhttps://orcid.org/0000-0001-7456-4333es


Files in this item

Thumbnail

This item appears in the following Collection(s)

Show simple item record

Attribution-NonCommercial-ShareAlike 4.0 International
Except where otherwise noted, this item's license is described as Attribution-NonCommercial-ShareAlike 4.0 International