{"created":"2023-07-25T10:26:23.195506+00:00","id":4958,"links":{},"metadata":{"_buckets":{"deposit":"73961fcf-2a19-4cda-836c-f74dd8725808"},"_deposit":{"created_by":4,"id":"4958","owners":[4],"pid":{"revision_id":0,"type":"depid","value":"4958"},"status":"published"},"_oai":{"id":"oai:naist.repo.nii.ac.jp:00004958","sets":["34:36"]},"author_link":["13309","328","311","222"],"item_1698715929687":{"attribute_name":"会議情報","attribute_value_mlt":[{"subitem_conference_country":"CHN","subitem_conference_date":{"subitem_conference_date_language":"en","subitem_conference_period":"Jul 14-18, 2014"},"subitem_conference_names":[{"subitem_conference_name":"ICME 2014 : IEEE International Conference on Multimedia and Expo ","subitem_conference_name_language":"en"}],"subitem_conference_places":[{"subitem_conference_place":"Chengdu","subitem_conference_place_language":"en"}]}]},"item_9_biblio_info_7":{"attribute_name":"書誌情報","attribute_value_mlt":[{"bibliographicIssueDates":{"bibliographicIssueDate":"2014","bibliographicIssueDateType":"Issued"},"bibliographicPageEnd":"6","bibliographicPageStart":"1"}]},"item_9_description_15":{"attribute_name":"フォーマット","attribute_value_mlt":[{"subitem_description":"application/pdf","subitem_description_type":"Other"}]},"item_9_description_5":{"attribute_name":"抄録","attribute_value_mlt":[{"subitem_description":"When observing a person (an actor) performing or demonstrating some activity for the purpose of learning the action, it is best for the viewers to be present at the same time and place as the actor. Otherwise, a video must be recorded. However, conventional video only provides two-dimensional (2D) motion, which lacks the original third dimension of motion. In the presence of some ambiguity, it may be hard for the viewer to comprehend the action with only two dimensions, making it harder to learn the action. This paper proposes an augmented reality system to reenact such actions at any time the viewer wants, in order to aid comprehension of 3D motion. In the proposed system, a user first captures the actor's motion and appearance, using a single RGB-D camera. Upon a viewer's request, our system displays the motion from an arbitrary viewpoint using a rough 3D model of the subject, made up of cylinders, and selecting the most appropriate textures based on the viewpoint and the subject's pose. We evaluate the usefulness of the system and the quality of the displayed images by user study.","subitem_description_language":"en","subitem_description_type":"Abstract"}]},"item_9_publisher_8":{"attribute_name":"出版者","attribute_value_mlt":[{"subitem_publisher":"IEEE","subitem_publisher_language":"en"}]},"item_9_relation_13":{"attribute_name":"DOI","attribute_value_mlt":[{"subitem_relation_type":"isVersionOf","subitem_relation_type_id":{"subitem_relation_type_id_text":"https://doi.org/10.1109/ICME.2014.6890243","subitem_relation_type_select":"DOI"}}]},"item_9_rights_14":{"attribute_name":"権利","attribute_value_mlt":[{"subitem_rights":"Copyright c 2014 IEEE","subitem_rights_language":"en"}]},"item_9_version_type_16":{"attribute_name":"著者版フラグ","attribute_value_mlt":[{"subitem_version_resource":"http://purl.org/coar/version/c_ab4af688f83e57aa","subitem_version_type":"AM"}]},"item_access_right":{"attribute_name":"アクセス権","attribute_value_mlt":[{"subitem_access_right":"open access","subitem_access_right_uri":"http://purl.org/coar/access_right/c_abf2"}]},"item_creator":{"attribute_name":"著者","attribute_type":"creator","attribute_value_mlt":[{"creatorNames":[{"creatorName":"Fabian, Lorenzo Dayrit","creatorNameLang":"en"}],"nameIdentifiers":[{"nameIdentifier":"13309","nameIdentifierScheme":"WEKO"}]},{"creatorNames":[{"creatorName":"Nakashima, Yuta","creatorNameLang":"en"}],"nameIdentifiers":[{"nameIdentifier":"328","nameIdentifierScheme":"WEKO"},{"nameIdentifier":"70633551","nameIdentifierScheme":"e-Rad","nameIdentifierURI":"https://kaken.nii.ac.jp/ja/search/?qm=70633551"}]},{"creatorNames":[{"creatorName":"Sato, Tomokazu","creatorNameLang":"en"}],"nameIdentifiers":[{"nameIdentifier":"311","nameIdentifierScheme":"WEKO"},{"nameIdentifier":"50362835","nameIdentifierScheme":"e-Rad","nameIdentifierURI":"https://kaken.nii.ac.jp/ja/search/?qm=50362835"}]},{"creatorNames":[{"creatorName":"Yokoya, Naokazu","creatorNameLang":"en"}],"nameIdentifiers":[{"nameIdentifier":"222","nameIdentifierScheme":"WEKO"},{"nameIdentifier":"10252834","nameIdentifierScheme":"e-Rad","nameIdentifierURI":"https://kaken.nii.ac.jp/ja/search/?qm=10252834"}]}]},"item_files":{"attribute_name":"ファイル情報","attribute_type":"file","attribute_value_mlt":[{"accessrole":"open_date","date":[{"dateType":"Available","dateValue":"2023-03-02"}],"displaytype":"detail","filename":"1353_Paper 634.pdf","filesize":[{"value":"1.2 MB"}],"format":"application/pdf","licensetype":"license_note","mimetype":"application/pdf","url":{"label":"fulltext","objectType":"fulltext","url":"https://naist.repo.nii.ac.jp/record/4958/files/1353_Paper 634.pdf"},"version_id":"9572dd62-5cb1-4c6f-b171-7acb73bca08e"}]},"item_keyword":{"attribute_name":"キーワード","attribute_value_mlt":[{"subitem_subject":"augmented reality","subitem_subject_language":"en","subitem_subject_scheme":"Other"},{"subitem_subject":"image colour analysis","subitem_subject_language":"en","subitem_subject_scheme":"Other"},{"subitem_subject":"image motion analysis","subitem_subject_language":"en","subitem_subject_scheme":"Other"},{"subitem_subject":"image sensors","subitem_subject_language":"en","subitem_subject_scheme":"Other"},{"subitem_subject":"image texture","subitem_subject_language":"en","subitem_subject_scheme":"Other"},{"subitem_subject":"solid modelling","subitem_subject_language":"en","subitem_subject_scheme":"Other"},{"subitem_subject":"video signal processing","subitem_subject_language":"en","subitem_subject_scheme":"Other"},{"subitem_subject":"3D motion comprehension","subitem_subject_language":"en","subitem_subject_scheme":"Other"},{"subitem_subject":"action learning","subitem_subject_language":"en","subitem_subject_scheme":"Other"},{"subitem_subject":"augmented reality system","subitem_subject_language":"en","subitem_subject_scheme":"Other"},{"subitem_subject":"free-viewpoint AR human-motion reenactment","subitem_subject_language":"en","subitem_subject_scheme":"Other"},{"subitem_subject":"rough 3D model","subitem_subject_language":"en","subitem_subject_scheme":"Other"},{"subitem_subject":"single RGB-D camera","subitem_subject_language":"en","subitem_subject_scheme":"Other"},{"subitem_subject":"single RGB-D video stream","subitem_subject_language":"en","subitem_subject_scheme":"Other"},{"subitem_subject":"texture selection","subitem_subject_language":"en","subitem_subject_scheme":"Other"},{"subitem_subject":"Cameras","subitem_subject_language":"en","subitem_subject_scheme":"Other"},{"subitem_subject":"Joints","subitem_subject_language":"en","subitem_subject_scheme":"Other"},{"subitem_subject":"Sensors","subitem_subject_language":"en","subitem_subject_scheme":"Other"},{"subitem_subject":"Solid modeling","subitem_subject_language":"en","subitem_subject_scheme":"Other"},{"subitem_subject":"Streaming media","subitem_subject_language":"en","subitem_subject_scheme":"Other"},{"subitem_subject":"Three-dimensional displays","subitem_subject_language":"en","subitem_subject_scheme":"Other"},{"subitem_subject":"Augmented reality","subitem_subject_language":"en","subitem_subject_scheme":"Other"},{"subitem_subject":"free-viewpoint image generation","subitem_subject_language":"en","subitem_subject_scheme":"Other"},{"subitem_subject":"human motion capture","subitem_subject_language":"en","subitem_subject_scheme":"Other"}]},"item_language":{"attribute_name":"言語","attribute_value_mlt":[{"subitem_language":"eng"}]},"item_resource_type":{"attribute_name":"資源タイプ","attribute_value_mlt":[{"resourcetype":"conference paper","resourceuri":"http://purl.org/coar/resource_type/c_5794"}]},"item_title":"Free-viewpoint AR human-motion reenactment based on a single RGB-D video stream","item_titles":{"attribute_name":"タイトル","attribute_value_mlt":[{"subitem_title":"Free-viewpoint AR human-motion reenactment based on a single RGB-D video stream","subitem_title_language":"en"}]},"item_type_id":"9","owner":"4","path":["36"],"pubdate":{"attribute_name":"PubDate","attribute_value":"2016-12-26"},"publish_date":"2016-12-26","publish_status":"0","recid":"4958","relation_version_is_last":true,"title":["Free-viewpoint AR human-motion reenactment based on a single RGB-D video stream"],"weko_creator_id":"4","weko_shared_id":-1},"updated":"2023-11-10T05:00:17.442902+00:00"}