{"id":53450,"created":"2025-01-18T23:17:32.847443+00:00","metadata":{"_oai":{"id":"oai:ipsj.ixsq.nii.ac.jp:00053450","sets":["1164:4619:4727:4728"]},"path":["4728"],"owner":"1","recid":"53450","title":["動画像からの歩行者検出法"],"pubdate":{"attribute_name":"公開日","attribute_value":"1991-11-22"},"_buckets":{"deposit":"0a589b01-2046-47b6-9306-27793ec7ecb8"},"_deposit":{"id":"53450","pid":{"type":"depid","value":"53450","revision_id":0},"owners":[1],"status":"published","created_by":1},"item_title":"動画像からの歩行者検出法","author_link":["0","0"],"item_titles":{"attribute_name":"タイトル","attribute_value_mlt":[{"subitem_title":"動画像からの歩行者検出法"},{"subitem_title":"Segmentation of People in Motion","subitem_title_language":"en"}]},"item_type_id":"4","publish_date":"1991-11-22","item_4_text_3":{"attribute_name":"著者所属","attribute_value_mlt":[{"subitem_text_value":"NTTヒューマンインタフェース研究所"},{"subitem_text_value":"カリフォルニア大学アーバイン校"}]},"item_4_text_4":{"attribute_name":"著者所属(英)","attribute_value_mlt":[{"subitem_text_value":"NTT Human Interface Laboratories Nippon Telegraph and Telephone Corp.","subitem_text_language":"en"},{"subitem_text_value":"Department of Electrical and Computer Engineering University of California, Irvine","subitem_text_language":"en"}]},"item_language":{"attribute_name":"言語","attribute_value_mlt":[{"subitem_language":"jpn"}]},"item_publisher":{"attribute_name":"出版者","attribute_value_mlt":[{"subitem_publisher":"情報処理学会","subitem_publisher_language":"ja"}]},"publish_status":"0","weko_shared_id":-1,"item_file_price":{"attribute_name":"Billing file","attribute_type":"file","attribute_value_mlt":[{"url":{"url":"https://ipsj.ixsq.nii.ac.jp/record/53450/files/IPSJ-CVIM91075005.pdf"},"date":[{"dateType":"Available","dateValue":"1993-11-22"}],"format":"application/pdf","billing":["billing_file"],"filename":"IPSJ-CVIM91075005.pdf","filesize":[{"value":"1.6 MB"}],"mimetype":"application/pdf","priceinfo":[{"tax":["include_tax"],"price":"660","billingrole":"5"},{"tax":["include_tax"],"price":"330","billingrole":"6"},{"tax":["include_tax"],"price":"0","billingrole":"20"},{"tax":["include_tax"],"price":"0","billingrole":"44"}],"accessrole":"open_date","version_id":"95b8bd60-2d06-47fc-b890-c66e64885033","displaytype":"detail","licensetype":"license_note","license_note":"Copyright (c) 1991 by the Information Processing Society of Japan"}]},"item_4_creator_5":{"attribute_name":"著者名","attribute_type":"creator","attribute_value_mlt":[{"creatorNames":[{"creatorName":"塩, 昭夫"},{"creatorName":"ジャックスクランスキー"}],"nameIdentifiers":[{}]}]},"item_4_creator_6":{"attribute_name":"著者名(英)","attribute_type":"creator","attribute_value_mlt":[{"creatorNames":[{"creatorName":"Akio, Shio","creatorNameLang":"en"},{"creatorName":"Jack, Sklansky","creatorNameLang":"en"}],"nameIdentifiers":[{}]}]},"item_4_source_id_9":{"attribute_name":"書誌レコードID","attribute_value_mlt":[{"subitem_source_identifier":"AA11131797","subitem_source_identifier_type":"NCID"}]},"item_4_textarea_12":{"attribute_name":"Notice","attribute_value_mlt":[{"subitem_textarea_value":"SIG Technical Reports are nonrefereed and hence may later appear in any journals, conferences, symposia, etc."}]},"item_resource_type":{"attribute_name":"資源タイプ","attribute_value_mlt":[{"resourceuri":"http://purl.org/coar/resource_type/c_18gh","resourcetype":"technical report"}]},"item_4_description_7":{"attribute_name":"論文抄録","attribute_value_mlt":[{"subitem_description":"単眼視による動画像のシーケンスから動いている歩行者を自動検出する方法について述べる。この方法は、画像強度、動き情報、および物体モデル (動いている歩行者の画像のモデル) に基づいている。画像中の人物は、類似色 (強度)の服を着ていることが多く、しかも他の人物に遮蔽されることも多い。また、個々の人物は髪の毛、皮膚、服など色 (強度) の異なった複数の部分に分割される。このため、画像から個々の人物を検出するのは容易ではない。しかし、同一人物に属する各部は協調して動くことを利用すれば、動き情報に基づいて領域の分割/統合が可能となる。この領域統合処理は、人間の運動視に於ける知覚的統合 (rceptual groupi) に対応する。複雑な実画像のシーケンスを用いた実験の結果、提案する歩行者検出法の妥当性を裏付ける良好な結果を得た。","subitem_description_type":"Other"}]},"item_4_description_8":{"attribute_name":"論文抄録(英)","attribute_value_mlt":[{"subitem_description":"A method for segmenting monocular images of people in motion from a cinematic sequence of frames is described. This method is based on image intensities, motion, and an object model-i.e., a model of the image of a person in motion. People in an image may be occluded by other people, and usually it is not easy to detect their boundaries. These boundaries can be detected with motion information if they move in different directions, even if there are almost no apparent differences among object intensities or colors. Each image of a person in a scene usually can be divided into several parts, each with distinct intensities or colors. The parts of a person can be merged into a single group by an iterative merging algorithm based on the object model and the motion information because the parts move coherently. This merging is analogous to the property of perceptual grouping in human visual perception of motion. Experiments based on a sequence of complex real scenes produced results that are supportive of our approach to the segmentation of people in motion.","subitem_description_type":"Other"}]},"item_4_biblio_info_10":{"attribute_name":"書誌情報","attribute_value_mlt":[{"bibliographicPageEnd":"8","bibliographic_titles":[{"bibliographic_title":"情報処理学会研究報告コンピュータビジョンとイメージメディア(CVIM)"}],"bibliographicPageStart":"1","bibliographicIssueDates":{"bibliographicIssueDate":"1991-11-22","bibliographicIssueDateType":"Issued"},"bibliographicIssueNumber":"101(1991-CVIM-075)","bibliographicVolumeNumber":"1991"}]},"relation_version_is_last":true,"weko_creator_id":"1"},"updated":"2025-01-22T06:19:23.144363+00:00","links":{}}