{"@context":{"@vocab":"https://cir.nii.ac.jp/schema/1.0/","rdfs":"http://www.w3.org/2000/01/rdf-schema#","dc":"http://purl.org/dc/elements/1.1/","dcterms":"http://purl.org/dc/terms/","foaf":"http://xmlns.com/foaf/0.1/","prism":"http://prismstandard.org/namespaces/basic/2.0/","cinii":"http://ci.nii.ac.jp/ns/1.0/","datacite":"https://schema.datacite.org/meta/kernel-4/","ndl":"http://ndl.go.jp/dcndl/terms/","jpcoar":"https://github.com/JPCOAR/schema/blob/master/2.0/"},"@id":"https://cir.nii.ac.jp/crid/1360021390775713408.json","@type":"Article","productIdentifier":[{"identifier":{"@type":"DOI","@value":"10.3389/frobt.2024.1362463"}},{"identifier":{"@type":"URI","@value":"https://www.frontiersin.org/articles/10.3389/frobt.2024.1362463/full"}}],"resourceType":"学術雑誌論文(journal article)","dc:title":[{"@value":"Evaluation of co-speech gestures grounded in word-distributed representation"}],"description":[{"type":"abstract","notation":[{"@value":"<jats:p>The condition for artificial agents to possess perceivable intentions can be considered that they have resolved a form of the symbol grounding problem. Here, the symbol grounding is considered an achievement of the state where the language used by the agent is endowed with some quantitative meaning extracted from the physical world. To achieve this type of symbol grounding, we adopt a method for characterizing robot gestures with quantitative meaning calculated from word-distributed representations constructed from a large corpus of text. In this method, a “size image” of a word is generated by defining an axis (index) that discriminates the “size” of the word in the word-distributed vector space. The generated size images are converted into gestures generated by a physical artificial agent (robot). The robot’s gesture can be set to reflect either the size of the word in terms of the amount of movement or in terms of its posture. To examine the perception of communicative intention in the robot that performs the gestures generated as described above, the authors examine human ratings on “the naturalness” obtained through an online survey, yielding results that partially validate our proposed method. Based on the results, the authors argue for the possibility of developing advanced artifacts that achieve human-like symbolic grounding.</jats:p>"}]}],"creator":[{"@id":"https://cir.nii.ac.jp/crid/1380021390775713545","@type":"Researcher","foaf:name":[{"@value":"Kosuke Sasaki"}]},{"@id":"https://cir.nii.ac.jp/crid/1380021390775713409","@type":"Researcher","foaf:name":[{"@value":"Junya Morita"}]},{"@id":"https://cir.nii.ac.jp/crid/1380021390775713417","@type":"Researcher","foaf:name":[{"@value":"Jumpei Nishikawa"}]}],"publication":{"publicationIdentifier":[{"@type":"EISSN","@value":"22969144"}],"prism:publicationName":[{"@value":"Frontiers in Robotics and AI"}],"dc:publisher":[{"@value":"Frontiers Media SA"}],"prism:publicationDate":"2024-04-25","prism:volume":"11"},"reviewed":"false","dcterms:accessRights":"http://purl.org/coar/access_right/c_abf2","dc:rights":["https://creativecommons.org/licenses/by/4.0/"],"url":[{"@id":"https://www.frontiersin.org/articles/10.3389/frobt.2024.1362463/full"}],"createdAt":"2024-04-25","modifiedAt":"2024-04-25","foaf:topic":[{"@id":"https://cir.nii.ac.jp/all?q=word-distributed%20representation","dc:title":"word-distributed representation"},{"@id":"https://cir.nii.ac.jp/all?q=robotics","dc:title":"robotics"},{"@id":"https://cir.nii.ac.jp/all?q=Robotics%20and%20AI","dc:title":"Robotics and AI"},{"@id":"https://cir.nii.ac.jp/all?q=Electronic%20computers.%20Computer%20science","dc:title":"Electronic computers. Computer science"},{"@id":"https://cir.nii.ac.jp/all?q=natural%20language%20processing%20(NLP)","dc:title":"natural language processing (NLP)"},{"@id":"https://cir.nii.ac.jp/all?q=TJ1-1570","dc:title":"TJ1-1570"},{"@id":"https://cir.nii.ac.jp/all?q=Mechanical%20engineering%20and%20machinery","dc:title":"Mechanical engineering and machinery"},{"@id":"https://cir.nii.ac.jp/all?q=QA75.5-76.95","dc:title":"QA75.5-76.95"},{"@id":"https://cir.nii.ac.jp/all?q=human-robot%20interaction%20(HRI)","dc:title":"human-robot interaction (HRI)"},{"@id":"https://cir.nii.ac.jp/all?q=co-speech%20iconic%20gesture","dc:title":"co-speech iconic gesture"}],"project":[{"@id":"https://cir.nii.ac.jp/crid/1040291932565078272","@type":"Project","projectIdentifier":[{"@type":"KAKEN","@value":"22H04861"},{"@type":"JGN","@value":"JP22H04861"},{"@type":"URI","@value":"https://kaken.nii.ac.jp/grant/KAKENHI-PUBLICLY-22H04861/"}],"notation":[{"@language":"ja","@value":"対話的モデルベース回想法の自然化と社会的接地"},{"@language":"en","@value":"Naturalization and Societal Grounding of Model-based Reminiscence"}]}],"relatedProduct":[{"@id":"https://cir.nii.ac.jp/crid/1360001113983667456","@type":"Article","relationType":["references"],"jpcoar:relatedTitle":[{"@value":"An Experimental Study of Apparent Behavior"}]},{"@id":"https://cir.nii.ac.jp/crid/1360016867506179200","@type":"Article","relationType":["references"],"jpcoar:relatedTitle":[{"@value":"Revisiting Human-Agent Communication: The Importance of Joint Co-construction and Understanding Mental States"}]},{"@id":"https://cir.nii.ac.jp/crid/1360290617639229696","@type":"Article","resourceType":"学術雑誌論文(journal article)","relationType":["references"],"jpcoar:relatedTitle":[{"@value":"Exploring What Is Encoded in Distributional Word Vectors: A Neurobiologically Motivated Analysis"}]},{"@id":"https://cir.nii.ac.jp/crid/1360298341433787136","@type":"Article","relationType":["references"],"jpcoar:relatedTitle":[{"@value":"ConceptNet 5.5: An Open Multilingual Graph of General Knowledge"}]},{"@id":"https://cir.nii.ac.jp/crid/1360584343257448192","@type":"Article","relationType":["references"],"jpcoar:relatedTitle":[{"@value":"Iconicity ratings for 14,000+ English words"}]},{"@id":"https://cir.nii.ac.jp/crid/1360584343257490176","@type":"Article","relationType":["references"],"jpcoar:relatedTitle":[{"@value":"Situating Language in the Real-World: The Role of Multimodal Iconicity and Indexicality"}]},{"@id":"https://cir.nii.ac.jp/crid/1360584346296923136","@type":"Article","relationType":["references"],"jpcoar:relatedTitle":[{"@value":"Iconicity ratings across the Japanese lexicon: A comparative study with English"}]},{"@id":"https://cir.nii.ac.jp/crid/1360584346476536064","@type":"Article","relationType":["references"],"jpcoar:relatedTitle":[{"@value":"A systematic survey of multiple social robots as a passive- and interactive-social medium"}]},{"@id":"https://cir.nii.ac.jp/crid/1360865817580491520","@type":"Article","relationType":["references"],"jpcoar:relatedTitle":[{"@value":"The developing role of prosody in novel word interpretation"}]},{"@id":"https://cir.nii.ac.jp/crid/1360865817581264256","@type":"Article","relationType":["references"],"jpcoar:relatedTitle":[{"@value":"Mind in Motion"}]},{"@id":"https://cir.nii.ac.jp/crid/1360865820895686784","@type":"Article","relationType":["references"],"jpcoar:relatedTitle":[{"@value":"Semantic projection recovers rich human knowledge of multiple object features from word embeddings"}]},{"@id":"https://cir.nii.ac.jp/crid/1361137045696104960","@type":"Article","relationType":["references"],"jpcoar:relatedTitle":[{"@value":"The symbol grounding problem"}]},{"@id":"https://cir.nii.ac.jp/crid/1362544419721415296","@type":"Article","relationType":["references"],"jpcoar:relatedTitle":[{"@value":"The uncanny advantage of using androids in cognitive and social science research"}]},{"@id":"https://cir.nii.ac.jp/crid/1362825894691784704","@type":"Article","relationType":["references"],"jpcoar:relatedTitle":[{"@value":"Contextual correlates of synonymy"}]},{"@id":"https://cir.nii.ac.jp/crid/1363388844170699520","@type":"Article","relationType":["references"],"jpcoar:relatedTitle":[{"@value":"WordNet"}]},{"@id":"https://cir.nii.ac.jp/crid/1364233268973125248","@type":"Article","relationType":["references"],"jpcoar:relatedTitle":[{"@value":"Machines and Mindlessness: Social Responses to Computers"}]},{"@id":"https://cir.nii.ac.jp/crid/1364233271059965056","@type":"Article","relationType":["references"],"jpcoar:relatedTitle":[{"@value":"The Proteus Effect: The Effect of Transformed Self-Representation on Behavior"}]},{"@id":"https://cir.nii.ac.jp/crid/1370021390775713413","@type":"Product","relationType":["references"],"jpcoar:relatedTitle":[{"@value":"A transition model for cognitions about agency"}]},{"@id":"https://cir.nii.ac.jp/crid/1370021390775713414","@type":"Product","relationType":["references"],"jpcoar:relatedTitle":[{"@value":"Robots learn social skills: end-to-end learning of co-speech gesture generation for humanoid robots"}]},{"@id":"https://cir.nii.ac.jp/crid/1370021390775713416","@type":"Product","relationType":["references"],"jpcoar:relatedTitle":[{"@value":"Wikipedia kiji ni taisuru kakucyo koyu hyogen label no tajyu fuyo (in Japanese)"}]},{"@id":"https://cir.nii.ac.jp/crid/1370021390775713417","@type":"Product","relationType":["references"],"jpcoar:relatedTitle":[{"@value":"Human motion modeling using dvgans"}]},{"@id":"https://cir.nii.ac.jp/crid/1370021390775713420","@type":"Product","relationType":["references"],"jpcoar:relatedTitle":[{"@value":"Efficient estimation of word representations in vector space"}]},{"@id":"https://cir.nii.ac.jp/crid/1370021390775713421","@type":"Product","relationType":["references"],"jpcoar:relatedTitle":[{"@value":"The distributional hypothesis"}]},{"@id":"https://cir.nii.ac.jp/crid/1370021390775713422","@type":"Product","relationType":["references"],"jpcoar:relatedTitle":[{"@value":"Motion and meaning: data-driven analyses of the relationship between gesture and communicative semantics"}]},{"@id":"https://cir.nii.ac.jp/crid/1370021390775713423","@type":"Product","relationType":["references"],"jpcoar:relatedTitle":[{"@value":"A neural probabilistic language model"}]},{"@id":"https://cir.nii.ac.jp/crid/1370021390775713424","@type":"Product","relationType":["references"],"jpcoar:relatedTitle":[{"@value":"Body movement construction based on distributionally represented size image"}]},{"@id":"https://cir.nii.ac.jp/crid/1370021390775713425","@type":"Product","relationType":["references"],"jpcoar:relatedTitle":[{"@value":"Extending the Japanese wordnet"}]},{"@id":"https://cir.nii.ac.jp/crid/1370021390775713426","@type":"Product","relationType":["references"],"jpcoar:relatedTitle":[{"@value":"Computers are social actors"}]},{"@id":"https://cir.nii.ac.jp/crid/1370021390775713427","@type":"Product","relationType":["references"],"jpcoar:relatedTitle":[{"@value":"From text to motion: grounding gpt-4 in a humanoid robot “alter3”"}]},{"@id":"https://cir.nii.ac.jp/crid/1370021390775713428","@type":"Product","relationType":["references"],"jpcoar:relatedTitle":[{"@value":"Generating body motions using spoken language in dialogue"}]},{"@id":"https://cir.nii.ac.jp/crid/1370021390775713429","@type":"Product","relationType":["references"],"jpcoar:relatedTitle":[{"@value":"Language models are few-shot learners"}]},{"@id":"https://cir.nii.ac.jp/crid/1370021390775713430","@type":"Product","relationType":["references"],"jpcoar:relatedTitle":[{"@value":"Large language models in textual analysis for gesture selection"}]},{"@id":"https://cir.nii.ac.jp/crid/1370021390775713537","@type":"Product","relationType":["references"],"jpcoar:relatedTitle":[{"@value":"Language2pose: natural language grounded pose forecasting"}]},{"@id":"https://cir.nii.ac.jp/crid/1370021390775713538","@type":"Product","relationType":["references"],"jpcoar:relatedTitle":[{"@value":"GPT-4 technical report"}]},{"@id":"https://cir.nii.ac.jp/crid/1370021390775713543","@type":"Product","relationType":["references"],"jpcoar:relatedTitle":[{"@value":"Human motion diffusion model"}]},{"@id":"https://cir.nii.ac.jp/crid/1370021390775713545","@type":"Product","relationType":["references"],"jpcoar:relatedTitle":[{"@value":"A robot for reconstructing presentation behavior in lecture"}]},{"@id":"https://cir.nii.ac.jp/crid/1370021390775713546","@type":"Product","relationType":["references"],"jpcoar:relatedTitle":[{"@value":"Palm: scaling language modeling with pathways"}]}],"dataSourceIdentifier":[{"@type":"CROSSREF","@value":"10.3389/frobt.2024.1362463"},{"@type":"KAKEN","@value":"PRODUCT-25163599"},{"@type":"OPENAIRE","@value":"doi_dedup___::c7b69b0d15214590f971bcde67c01d51"}]}