{"updated":"2025-01-21T14:19:33.818674+00:00","metadata":{"_oai":{"id":"oai:ipsj.ixsq.nii.ac.jp:00094851","sets":["1164:4619:6988:7247"]},"path":["7247"],"owner":"11","recid":"94851","title":["映像閲覧行動の時空間ずれ構造モデルを用いた注視点予測"],"pubdate":{"attribute_name":"公開日","attribute_value":"2013-08-26"},"_buckets":{"deposit":"4dc3548b-80ed-4e68-99f7-759513060810"},"_deposit":{"id":"94851","pid":{"type":"depid","value":"94851","revision_id":0},"owners":[11],"status":"published","created_by":11},"item_title":"映像閲覧行動の時空間ずれ構造モデルを用いた注視点予測","author_link":["0","0"],"item_titles":{"attribute_name":"タイトル","attribute_value_mlt":[{"subitem_title":"映像閲覧行動の時空間ずれ構造モデルを用いた注視点予測"},{"subitem_title":"Gaze Point Prediction with Gap Structure Models","subitem_title_language":"en"}]},"item_type_id":"4","publish_date":"2013-08-26","item_4_text_3":{"attribute_name":"著者所属","attribute_value_mlt":[{"subitem_text_value":"京都大学大学院情報学研究科"},{"subitem_text_value":"京都大学大学院情報学研究科"},{"subitem_text_value":"京都大学大学院情報学研究科"}]},"item_4_text_4":{"attribute_name":"著者所属(英)","attribute_value_mlt":[{"subitem_text_value":"Graduate School of Informatics, Kyoto University","subitem_text_language":"en"},{"subitem_text_value":"Graduate School of Informatics, Kyoto University","subitem_text_language":"en"},{"subitem_text_value":"Graduate School of Informatics, Kyoto University","subitem_text_language":"en"}]},"item_language":{"attribute_name":"言語","attribute_value_mlt":[{"subitem_language":"jpn"}]},"item_publisher":{"attribute_name":"出版者","attribute_value_mlt":[{"subitem_publisher":"情報処理学会","subitem_publisher_language":"ja"}]},"publish_status":"0","weko_shared_id":-1,"item_file_price":{"attribute_name":"Billing file","attribute_type":"file","attribute_value_mlt":[{"url":{"url":"https://ipsj.ixsq.nii.ac.jp/record/94851/files/IPSJ-CVIM13188010.pdf"},"date":[{"dateType":"Available","dateValue":"2100-01-01"}],"format":"application/pdf","billing":["billing_file"],"filename":"IPSJ-CVIM13188010.pdf","filesize":[{"value":"737.9 kB"}],"mimetype":"application/pdf","priceinfo":[{"tax":["include_tax"],"price":"0","billingrole":"20"},{"tax":["include_tax"],"price":"0","billingrole":"44"}],"accessrole":"open_date","version_id":"78f53d2d-8d41-47b9-afdf-504448bfc981","displaytype":"detail","licensetype":"license_note","license_note":"Copyright (c) 2013 by the Institute of Electronics, Information and Communication Engineers\nThis SIG report is only available to those in membership of the SIG."}]},"item_4_creator_5":{"attribute_name":"著者名","attribute_type":"creator","attribute_value_mlt":[{"creatorNames":[{"creatorName":"米谷, 竜"},{"creatorName":"川嶋, 宏彰"},{"creatorName":"松山, 隆司"}],"nameIdentifiers":[{}]}]},"item_4_creator_6":{"attribute_name":"著者名(英)","attribute_type":"creator","attribute_value_mlt":[{"creatorNames":[{"creatorName":"Ryo, Yonetani","creatorNameLang":"en"},{"creatorName":"Hiroaki, Kawashima","creatorNameLang":"en"},{"creatorName":"Takashi, Matsuyama","creatorNameLang":"en"}],"nameIdentifiers":[{}]}]},"item_4_source_id_9":{"attribute_name":"書誌レコードID","attribute_value_mlt":[{"subitem_source_identifier":"AA11131797","subitem_source_identifier_type":"NCID"}]},"item_4_textarea_12":{"attribute_name":"Notice","attribute_value_mlt":[{"subitem_textarea_value":"SIG Technical Reports are nonrefereed and hence may later appear in any journals, conferences, symposia, etc."}]},"item_resource_type":{"attribute_name":"資源タイプ","attribute_value_mlt":[{"resourceuri":"http://purl.org/coar/resource_type/c_18gh","resourcetype":"technical report"}]},"item_4_description_7":{"attribute_name":"論文抄録","attribute_value_mlt":[{"subitem_description":"人間が一般の映像を閲覧する状況では,映像中において人間の視覚的注意を引きつける顕著領域と,実際の注視点との間に時空間的なずれが生まれる.このずれは,顕著領域の変動 (顕著性変動) に対する反応遅れや動き予測に起因するものであり,変動の種類によってその程度が異なる.本研究では,顕著領域・注視点間のずれが持つ構造と顕著性変動の関係性に基づいて,映像から注視点を予測する手法を提案する.提案手法は,注視点に対して時空間ずれをともなって観測される顕著性変動パターンを,単純な変動パターンの組み合わせによって表現する.これにより,「注視点の近傍でどのような顕著性変動パターンがどのような時空間ずれとともに観測されやすいか」 をデータドリブンに学習することによる注視点予測が実現できる.","subitem_description_type":"Other"}]},"item_4_description_8":{"attribute_name":"論文抄録(英)","attribute_value_mlt":[{"subitem_description":"When we are watching videos, there exist spatiotemporal gaps between salient regions that attract visual attention and actual points of gaze. These gaps result from reaction delays or anticipation in eye movements toward dynamic changes of the salient regions (saliency dynamics). In this paper, we exploit the underlying structures of the gaps for gaze point prediction from videos. The proposed method involves the relationships between the gaps and the saliency dynamics by introducing localized primitive patterns of salient regions for the description of saliency dynamics. It enables us to learn “what types of saliency dynamics can be observed around the points of gaze” in a data-driven fashion for the gaze point prediction.","subitem_description_type":"Other"}]},"item_4_biblio_info_10":{"attribute_name":"書誌情報","attribute_value_mlt":[{"bibliographicPageEnd":"6","bibliographic_titles":[{"bibliographic_title":"研究報告コンピュータビジョンとイメージメディア(CVIM)"}],"bibliographicPageStart":"1","bibliographicIssueDates":{"bibliographicIssueDate":"2013-08-26","bibliographicIssueDateType":"Issued"},"bibliographicIssueNumber":"10","bibliographicVolumeNumber":"2013-CVIM-188"}]},"relation_version_is_last":true,"weko_creator_id":"11"},"created":"2025-01-18T23:42:04.139379+00:00","id":94851,"links":{}}