{"created":"2021-03-01T06:06:46.341258+00:00","id":3028,"links":{},"metadata":{"_buckets":{"deposit":"550da76d-ab76-4b7a-ac52-1cda2d2304f0"},"_deposit":{"id":"3028","owners":[],"pid":{"revision_id":0,"type":"depid","value":"3028"},"status":"published"},"_oai":{"id":"oai:niigata-u.repo.nii.ac.jp:00003028","sets":["423:424:425","453:454"]},"item_5_alternative_title_1":{"attribute_name":"その他のタイトル","attribute_value_mlt":[{"subitem_alternative_title":"Inverse Animation : An Approach to Image-based Motion Caputuring, Gesture Recognition and Performance Animation (Special Section on Seeing Human)"}]},"item_5_biblio_info_6":{"attribute_name":"書誌情報","attribute_value_mlt":[{"bibliographicIssueDates":{"bibliographicIssueDate":"2006-06","bibliographicIssueDateType":"Issued"},"bibliographicIssueNumber":"SIG9","bibliographicPageEnd":"31","bibliographicPageStart":"21","bibliographicVolumeNumber":"47","bibliographic_titles":[{"bibliographic_title":"情報処理学会論文誌. コンピュータビジョンとイメージメディア"},{"bibliographic_title":"情報処理学会論文誌. コンピュータビジョンとイメージメディア","bibliographic_titleLang":"en"}]}]},"item_5_description_4":{"attribute_name":"抄録","attribute_value_mlt":[{"subitem_description":"映像からの動作測定法には2つの大きな利点がある.それは,身体に接触することなく自然な動作を測定することができることと,過去の人物でもその映像が残ってさえいれば,その動作を測定し再現できることである.本論文では,映像から動作を測定し動作を認識し,さらに映像制作などで動作を再利用することのできる一貫した方式を提案する.本方式では動作を表現し記述するために多関節モデルを使用する.動作の認識とアニメーションとで共通の動作モデルを使用しているので,互いに蓄積したデータを利用することができる.動作データの蓄積が多くなれば,信頼性のある認識と豊かな映像表現が期待できる.本論文で提案した方式をインバースアニメーションと呼ぶことにする.この呼称はリアルなレンダリングのために,画像からその生成過程を推定する手法がインバースレンダリングと総称されていることにヒントを得た.","subitem_description_type":"Abstract"},{"subitem_description":"An image-based motion capturing has two major advantages. First, it can measure natural movement of human body non-invasively. Second, for even a person passed away, as far as his/her action is recorded on a film or video, it can reconstruct his/her movement. This paper proposes a novel approach to motion capturing, gesture recognition and creating performance animation. This approach uses an articulated model to describe a human body in motion. The common model makes it possible for gesture recognition and performance animation to share each other's action data. The richer the action databases are, one can expect the more reliable gesture recognition and the richer performance in animation. We shall call the proposed approach an inverse animation. This naming is inspired by inverse rendering which is technique to estimate a process generating an image.","subitem_description_type":"Abstract"}]},"item_5_full_name_3":{"attribute_name":"著者別名","attribute_value_mlt":[{"nameIdentifiers":[{"nameIdentifier":"40475","nameIdentifierScheme":"WEKO"}],"names":[{"name":"Yamamoto, Masanobu"}]}]},"item_5_publisher_7":{"attribute_name":"出版者","attribute_value_mlt":[{"subitem_publisher":"情報処理学会"}]},"item_5_relation_31":{"attribute_name":"異版である","attribute_value_mlt":[{"subitem_relation_type":"isVersionOf","subitem_relation_type_id":{"subitem_relation_type_id_text":"http://ci.nii.ac.jp/naid/110006391060","subitem_relation_type_select":"URI"}}]},"item_5_rights_15":{"attribute_name":"権利","attribute_value_mlt":[{"subitem_rights":"一般社団法人情報処理学会"},{"subitem_rights":"本文データは学協会の許諾に基づきCiNiiから複製したものである"}]},"item_5_select_19":{"attribute_name":"著者版フラグ","attribute_value_mlt":[{"subitem_select_item":"publisher"}]},"item_5_source_id_11":{"attribute_name":"書誌レコードID","attribute_value_mlt":[{"subitem_source_identifier":"AA11560603","subitem_source_identifier_type":"NCID"}]},"item_5_source_id_9":{"attribute_name":"ISSN","attribute_value_mlt":[{"subitem_source_identifier":"03875806","subitem_source_identifier_type":"ISSN"}]},"item_creator":{"attribute_name":"著者","attribute_type":"creator","attribute_value_mlt":[{"creatorNames":[{"creatorName":"山本, 正信"}],"nameIdentifiers":[{"nameIdentifier":"40474","nameIdentifierScheme":"WEKO"}]}]},"item_files":{"attribute_name":"ファイル情報","attribute_type":"file","attribute_value_mlt":[{"accessrole":"open_date","date":[{"dateType":"Available","dateValue":"2019-07-30"}],"displaytype":"detail","filename":"110006391060.pdf","filesize":[{"value":"2.0 MB"}],"format":"application/pdf","licensetype":"license_note","mimetype":"application/pdf","url":{"label":"110006391060.pdf","url":"https://niigata-u.repo.nii.ac.jp/record/3028/files/110006391060.pdf"},"version_id":"0a6cfc33-80be-4838-9a49-4e914669044f"}]},"item_language":{"attribute_name":"言語","attribute_value_mlt":[{"subitem_language":"jpn"}]},"item_resource_type":{"attribute_name":"資源タイプ","attribute_value_mlt":[{"resourcetype":"journal article","resourceuri":"http://purl.org/coar/resource_type/c_6501"}]},"item_title":"インバースアニメーション : 映像からの動作の計測・認識・再利用(<特集>人をみる)","item_titles":{"attribute_name":"タイトル","attribute_value_mlt":[{"subitem_title":"インバースアニメーション : 映像からの動作の計測・認識・再利用(<特集>人をみる)"},{"subitem_title":"インバースアニメーション : 映像からの動作の計測・認識・再利用(<特集>人をみる)","subitem_title_language":"en"}]},"item_type_id":"5","owner":"1","path":["454","425"],"pubdate":{"attribute_name":"公開日","attribute_value":"2012-08-31"},"publish_date":"2012-08-31","publish_status":"0","recid":"3028","relation_version_is_last":true,"title":["インバースアニメーション : 映像からの動作の計測・認識・再利用(<特集>人をみる)"],"weko_creator_id":"1","weko_shared_id":null},"updated":"2022-12-15T03:36:14.720871+00:00"}