{"metadata":{"_oai":{"id":"oai:ipsj.ixsq.nii.ac.jp:00219703","sets":["6164:6165:6640:11008"]},"path":["11008"],"owner":"44499","recid":"219703","title":["A study on estimating the accurate head IMU motion from Video"],"pubdate":{"attribute_name":"公開日","attribute_value":"2022-07-06"},"_buckets":{"deposit":"a2ce5c2e-eb68-48ee-a73f-934dda57a7c0"},"_deposit":{"id":"219703","pid":{"type":"depid","value":"219703","revision_id":0},"owners":[44499],"status":"published","created_by":44499},"item_title":"A study on estimating the accurate head IMU motion from Video","author_link":["573300","573297","573304","573305","573301","573303","573299","573306","573302","573298"],"item_titles":{"attribute_name":"タイトル","attribute_value_mlt":[{"subitem_title":"A study on estimating the accurate head IMU motion from Video"},{"subitem_title":"A study on estimating the accurate head IMU motion from Video","subitem_title_language":"en"}]},"item_type_id":"18","publish_date":"2022-07-06","item_language":{"attribute_name":"言語","attribute_value_mlt":[{"subitem_language":"eng"}]},"item_18_text_3":{"attribute_name":"著者所属","attribute_value_mlt":[{"subitem_text_value":"九州大学"},{"subitem_text_value":"九州大学"},{"subitem_text_value":"公立はこだて未来大学"},{"subitem_text_value":"九州大学"},{"subitem_text_value":"九州大学"}]},"item_18_text_4":{"attribute_name":"著者所属(英)","attribute_value_mlt":[{"subitem_text_value":"Kyushu University","subitem_text_language":"en"},{"subitem_text_value":"Kyushu University","subitem_text_language":"en"},{"subitem_text_value":"Future University Hakodate","subitem_text_language":"en"},{"subitem_text_value":"Kyushu University","subitem_text_language":"en"},{"subitem_text_value":"Kyushu University","subitem_text_language":"en"}]},"item_publisher":{"attribute_name":"出版者","attribute_value_mlt":[{"subitem_publisher":"情報処理学会","subitem_publisher_language":"ja"}]},"publish_status":"0","weko_shared_id":-1,"item_file_price":{"attribute_name":"Billing file","attribute_type":"file","attribute_value_mlt":[{"url":{"url":"https://ipsj.ixsq.nii.ac.jp/record/219703/files/IPSJ-DICOMO2022129.pdf","label":"IPSJ-DICOMO2022129.pdf"},"date":[{"dateType":"Available","dateValue":"2024-07-06"}],"format":"application/pdf","billing":["billing_file"],"filename":"IPSJ-DICOMO2022129.pdf","filesize":[{"value":"1.5 MB"}],"mimetype":"application/pdf","priceinfo":[{"tax":["include_tax"],"price":"660","billingrole":"5"},{"tax":["include_tax"],"price":"330","billingrole":"6"},{"tax":["include_tax"],"price":"0","billingrole":"44"}],"accessrole":"open_date","version_id":"fa6ae885-ad25-426f-9f2a-b2572ae602ee","displaytype":"detail","licensetype":"license_note","license_note":"Copyright (c) 2022 by the Information Processing Society of Japan"}]},"item_18_creator_5":{"attribute_name":"著者名","attribute_type":"creator","attribute_value_mlt":[{"creatorNames":[{"creatorName":"MinYen, Lu"}],"nameIdentifiers":[{}]},{"creatorNames":[{"creatorName":"ChenHao, Chen"}],"nameIdentifiers":[{}]},{"creatorNames":[{"creatorName":"石田, 繁巳"}],"nameIdentifiers":[{}]},{"creatorNames":[{"creatorName":"中村, 優吾"}],"nameIdentifiers":[{}]},{"creatorNames":[{"creatorName":"荒川, 豊"}],"nameIdentifiers":[{}]}]},"item_18_creator_6":{"attribute_name":"著者名(英)","attribute_type":"creator","attribute_value_mlt":[{"creatorNames":[{"creatorName":"Minyen, Lu","creatorNameLang":"en"}],"nameIdentifiers":[{}]},{"creatorNames":[{"creatorName":"Chenhao, Chen","creatorNameLang":"en"}],"nameIdentifiers":[{}]},{"creatorNames":[{"creatorName":"Shigemi, Ishida","creatorNameLang":"en"}],"nameIdentifiers":[{}]},{"creatorNames":[{"creatorName":"Yugo, Nakamura","creatorNameLang":"en"}],"nameIdentifiers":[{}]},{"creatorNames":[{"creatorName":"Yutaka, Arakawa","creatorNameLang":"en"}],"nameIdentifiers":[{}]}]},"item_resource_type":{"attribute_name":"資源タイプ","attribute_value_mlt":[{"resourceuri":"http://purl.org/coar/resource_type/c_5794","resourcetype":"conference paper"}]},"item_18_description_7":{"attribute_name":"論文抄録","attribute_value_mlt":[{"subitem_description":"Inertial measurement unit (IMU) data have been utilized in human activity recognition (HAR). In recent studies, deep learning recognition for IMU data has caught researchers' attention for the capability of automatic feature extraction and accurate prediction. On the other hand, the challenge of data collection and labeling discourages researchers to step into it. IMUTube provides a solution by building up a pipeline to estimate virtual IMU data from YouTube videos for body motion. For head motion data, several methods, such as OpenFace 2.0 provide the function of predicting facial landmarks and calculating head facing angle from video. However, to our knowledge, there is no study focusing on estimating IMU data from human head motion. In our previous work DisCaaS, we created the M3B dataset which contains IMU and 360-degree video data from the meeting. We exploit head motion data extraction models to predict participants' nodding and speaking gestures. In order to further improve the performance of nodding recognition, in this paper, we are interested in understanding the quality of estimated gyro data calculated from these existing head motion models. We investigate the difference between the motion data estimated from video and those measured by a 9-axis sensor not only in the time domain but also in the frequency domain. Finally, we discuss the future direction of the result.","subitem_description_type":"Other"}]},"item_18_description_8":{"attribute_name":"論文抄録(英)","attribute_value_mlt":[{"subitem_description":"Inertial measurement unit (IMU) data have been utilized in human activity recognition (HAR). In recent studies, deep learning recognition for IMU data has caught researchers' attention for the capability of automatic feature extraction and accurate prediction. On the other hand, the challenge of data collection and labeling discourages researchers to step into it. IMUTube provides a solution by building up a pipeline to estimate virtual IMU data from YouTube videos for body motion. For head motion data, several methods, such as OpenFace 2.0 provide the function of predicting facial landmarks and calculating head facing angle from video. However, to our knowledge, there is no study focusing on estimating IMU data from human head motion. In our previous work DisCaaS, we created the M3B dataset which contains IMU and 360-degree video data from the meeting. We exploit head motion data extraction models to predict participants' nodding and speaking gestures. In order to further improve the performance of nodding recognition, in this paper, we are interested in understanding the quality of estimated gyro data calculated from these existing head motion models. We investigate the difference between the motion data estimated from video and those measured by a 9-axis sensor not only in the time domain but also in the frequency domain. Finally, we discuss the future direction of the result.","subitem_description_type":"Other"}]},"item_18_biblio_info_10":{"attribute_name":"書誌情報","attribute_value_mlt":[{"bibliographicPageEnd":"923","bibliographic_titles":[{"bibliographic_title":"マルチメディア,分散,協調とモバイルシンポジウム2022論文集"}],"bibliographicPageStart":"918","bibliographicIssueDates":{"bibliographicIssueDate":"2022-07-06","bibliographicIssueDateType":"Issued"},"bibliographicVolumeNumber":"2022"}]},"relation_version_is_last":true,"weko_creator_id":"44499"},"id":219703,"updated":"2025-01-19T14:48:20.901094+00:00","links":{},"created":"2025-01-19T01:19:46.058069+00:00"}