{"@context":{"@vocab":"https://cir.nii.ac.jp/schema/1.0/","rdfs":"http://www.w3.org/2000/01/rdf-schema#","dc":"http://purl.org/dc/elements/1.1/","dcterms":"http://purl.org/dc/terms/","foaf":"http://xmlns.com/foaf/0.1/","prism":"http://prismstandard.org/namespaces/basic/2.0/","cinii":"http://ci.nii.ac.jp/ns/1.0/","datacite":"https://schema.datacite.org/meta/kernel-4/","ndl":"http://ndl.go.jp/dcndl/terms/","jpcoar":"https://github.com/JPCOAR/schema/blob/master/2.0/"},"@id":"https://cir.nii.ac.jp/crid/1360285705668188288.json","@type":"Article","productIdentifier":[{"identifier":{"@type":"DOI","@value":"10.1007/978-3-642-34014-7_5"}},{"identifier":{"@type":"URI","@value":"http://link.springer.com/content/pdf/10.1007/978-3-642-34014-7_5.pdf"}}],"resourceType":"学術雑誌論文(journal article)","dc:title":[{"@value":"Using Speech Data to Recognize Emotion in Human Gait"}],"creator":[{"@id":"https://cir.nii.ac.jp/crid/1380285705668188419","@type":"Researcher","foaf:name":[{"@value":"Angelica Lim"}]},{"@id":"https://cir.nii.ac.jp/crid/1380285705668188416","@type":"Researcher","foaf:name":[{"@value":"Hiroshi G. Okuno"}]}],"publication":{"publicationIdentifier":[{"@type":"PISSN","@value":"03029743"},{"@type":"EISSN","@value":"16113349"},{"@type":"ISBN","@value":"9783642340130"},{"@type":"ISBN","@value":"9783642340147"}],"prism:publicationName":[{"@value":"Lecture Notes in Computer Science"}],"dc:publisher":[{"@value":"Springer Berlin Heidelberg"}],"prism:publicationDate":"2012","prism:startingPage":"52","prism:endingPage":"64"},"reviewed":"false","url":[{"@id":"http://link.springer.com/content/pdf/10.1007/978-3-642-34014-7_5.pdf"}],"createdAt":"2012-09-18","modifiedAt":"2022-01-29","project":[{"@id":"https://cir.nii.ac.jp/crid/1040282257168118144","@type":"Project","projectIdentifier":[{"@type":"KAKEN","@value":"24220006"},{"@type":"JGN","@value":"JP24220006"},{"@type":"URI","@value":"https://kaken.nii.ac.jp/grant/KAKENHI-PROJECT-24220006/"}],"notation":[{"@language":"ja","@value":"ロボット聴覚の実環境理解に向けた多面的展開"},{"@language":"en","@value":"Deployment of Robot Audition Toward Understanding Real World"}]}],"relatedProduct":[{"@id":"https://cir.nii.ac.jp/crid/1050564285744979712","@type":"Article","resourceType":"学術雑誌論文(journal article)","relationType":["references"],"jpcoar:relatedTitle":[{"@language":"en","@value":"Towards expressive musical robots: A cross-modal framework for emotional gesture, voice and music"}]},{"@id":"https://cir.nii.ac.jp/crid/1360002217955484160","@type":"Article","resourceType":"学術雑誌論文(journal article)","relationType":["references"],"jpcoar:relatedTitle":[{"@value":"Converting emotional voice to motion for robot telepresence"}]},{"@id":"https://cir.nii.ac.jp/crid/1360011144063681408","@type":"Article","relationType":["references"],"jpcoar:relatedTitle":[{"@value":"Recognition of Vocal Expressions of Emotion"}]},{"@id":"https://cir.nii.ac.jp/crid/1360011144475092096","@type":"Article","relationType":["references"],"jpcoar:relatedTitle":[{"@value":"Fourier principles for emotion-based human figure animation"}]},{"@id":"https://cir.nii.ac.jp/crid/1360292619092238976","@type":"Article","relationType":["references"],"jpcoar:relatedTitle":[{"@value":"Communicating expressiveness and affect in multimodal interactive systems"}]},{"@id":"https://cir.nii.ac.jp/crid/1360574095742091776","@type":"Article","relationType":["references"],"jpcoar:relatedTitle":[{"@value":"Critical features for the perception of emotion from gait"}]},{"@id":"https://cir.nii.ac.jp/crid/1360845538926265600","@type":"Article","relationType":["references"],"jpcoar:relatedTitle":[{"@value":"Emotion recognition in human-computer interaction"}]},{"@id":"https://cir.nii.ac.jp/crid/1360855568788224512","@type":"Article","relationType":["references"],"jpcoar:relatedTitle":[{"@value":"Changing Musical Emotion: A Computational Rule System for Modifying Score and Performance"}]},{"@id":"https://cir.nii.ac.jp/crid/1361137043657591808","@type":"Article","relationType":["references"],"jpcoar:relatedTitle":[{"@value":"Recognition of Affect Based on Gait Patterns"}]},{"@id":"https://cir.nii.ac.jp/crid/1361137046416271744","@type":"Article","relationType":["references"],"jpcoar:relatedTitle":[{"@value":"The HUMAINE Database: Addressing the Collection and Annotation of Naturalistic and Induced Emotional Data"}]},{"@id":"https://cir.nii.ac.jp/crid/1361699996181190912","@type":"Article","relationType":["references"],"jpcoar:relatedTitle":[{"@value":"Designing Sociable Robots"}]},{"@id":"https://cir.nii.ac.jp/crid/1361981471130197760","@type":"Article","relationType":["references"],"jpcoar:relatedTitle":[{"@value":"Score-Independent Audio Features for Description of Music Expression"}]},{"@id":"https://cir.nii.ac.jp/crid/1362544419192693632","@type":"Article","relationType":["references"],"jpcoar:relatedTitle":[{"@value":"Vocal affect expression: A review and a model for future research."}]},{"@id":"https://cir.nii.ac.jp/crid/1362544419665570816","@type":"Article","relationType":["references"],"jpcoar:relatedTitle":[{"@value":"The identification of emotions from gait information"}]},{"@id":"https://cir.nii.ac.jp/crid/1362544420218322688","@type":"Article","relationType":["references"],"jpcoar:relatedTitle":[{"@value":"Classical and novel discriminant features for affect recognition from speech"}]},{"@id":"https://cir.nii.ac.jp/crid/1362544420426045056","@type":"Article","relationType":["references"],"jpcoar:relatedTitle":[{"@value":"Studies on gesture expressivity for a virtual agent"}]},{"@id":"https://cir.nii.ac.jp/crid/1362825893412358016","@type":"Article","relationType":["references"],"jpcoar:relatedTitle":[{"@value":"Communication of emotions in vocal expression and music performance: Different channels, same code?"}]},{"@id":"https://cir.nii.ac.jp/crid/1363107369422784896","@type":"Article","relationType":["references"],"jpcoar:relatedTitle":[{"@value":"Recognition of Emotions in Gait Patterns by Means of Artificial Neural Nets"}]},{"@id":"https://cir.nii.ac.jp/crid/1363107371075292160","@type":"Article","relationType":["references"],"jpcoar:relatedTitle":[{"@value":"A motion capture library for the study of identity, gender, and emotion perception from biological motion"}]},{"@id":"https://cir.nii.ac.jp/crid/1363951795965685120","@type":"Article","relationType":["references"],"jpcoar:relatedTitle":[{"@value":"The Use of Body Movements and Gestures as Cues to Emotions in Younger and Older Adults"}]}],"dataSourceIdentifier":[{"@type":"CROSSREF","@value":"10.1007/978-3-642-34014-7_5"},{"@type":"KAKEN","@value":"PRODUCT-14471808"}]}