{"created":"2025-01-19T01:39:35.293317+00:00","updated":"2025-01-19T08:59:07.595557+00:00","metadata":{"_oai":{"id":"oai:ipsj.ixsq.nii.ac.jp:00237130","sets":["1164:6389:11481:11662"]},"path":["11662"],"owner":"44499","recid":"237130","title":["The OU-ISIR Multimodal Biometric Database and Its Performance Evaluation"],"pubdate":{"attribute_name":"公開日","attribute_value":"2024-07-15"},"_buckets":{"deposit":"7a9117e5-e5fc-4c8c-97df-82fd0ce95142"},"_deposit":{"id":"237130","pid":{"type":"depid","value":"237130","revision_id":0},"owners":[44499],"status":"published","created_by":44499},"item_title":"The OU-ISIR Multimodal Biometric Database and Its Performance Evaluation","author_link":["648453","648454","648449","648451","648452","648455","648450","648448"],"item_titles":{"attribute_name":"タイトル","attribute_value_mlt":[{"subitem_title":"The OU-ISIR Multimodal Biometric Database and Its Performance Evaluation"},{"subitem_title":"The OU-ISIR Multimodal Biometric Database and Its Performance Evaluation","subitem_title_language":"en"}]},"item_keyword":{"attribute_name":"キーワード","attribute_value_mlt":[{"subitem_subject":"BioX ムーショットセッション","subitem_subject_scheme":"Other"}]},"item_type_id":"4","publish_date":"2024-07-15","item_4_text_3":{"attribute_name":"著者所属","attribute_value_mlt":[{"subitem_text_value":"SANKEN, Osaka University"},{"subitem_text_value":"SANKEN, Osaka University"}]},"item_4_text_4":{"attribute_name":"著者所属(英)","attribute_value_mlt":[{"subitem_text_value":"SANKEN, Osaka University","subitem_text_language":"en"},{"subitem_text_value":"SANKEN, Osaka University","subitem_text_language":"en"}]},"item_language":{"attribute_name":"言語","attribute_value_mlt":[{"subitem_language":"eng"}]},"item_publisher":{"attribute_name":"出版者","attribute_value_mlt":[{"subitem_publisher":"情報処理学会","subitem_publisher_language":"ja"}]},"publish_status":"0","weko_shared_id":-1,"item_file_price":{"attribute_name":"Billing file","attribute_type":"file","attribute_value_mlt":[{"url":{"url":"https://ipsj.ixsq.nii.ac.jp/record/237130/files/IPSJ-SPT24056009.pdf","label":"IPSJ-SPT24056009.pdf"},"date":[{"dateType":"Available","dateValue":"2026-07-15"}],"format":"application/pdf","billing":["billing_file"],"filename":"IPSJ-SPT24056009.pdf","filesize":[{"value":"1.8 MB"}],"mimetype":"application/pdf","priceinfo":[{"tax":["include_tax"],"price":"0","billingrole":"46"},{"tax":["include_tax"],"price":"0","billingrole":"44"}],"accessrole":"open_date","version_id":"e5c9d097-e452-4ab2-b1c4-646c04feb022","displaytype":"detail","licensetype":"license_note","license_note":"Copyright (c) 2024 by the Institute of Electronics, Information and Communication Engineers This SIG report is only available to those in membership of the SIG."}]},"item_4_creator_5":{"attribute_name":"著者名","attribute_type":"creator","attribute_value_mlt":[{"creatorNames":[{"creatorName":"Chi, Xu"}],"nameIdentifiers":[{}]},{"creatorNames":[{"creatorName":"Xiang, Li"}],"nameIdentifiers":[{}]},{"creatorNames":[{"creatorName":"Allam, Shehata"}],"nameIdentifiers":[{}]},{"creatorNames":[{"creatorName":"Yasushi, Yagi"}],"nameIdentifiers":[{}]}]},"item_4_creator_6":{"attribute_name":"著者名(英)","attribute_type":"creator","attribute_value_mlt":[{"creatorNames":[{"creatorName":"Chi, Xu","creatorNameLang":"en"}],"nameIdentifiers":[{}]},{"creatorNames":[{"creatorName":"Xiang, Li","creatorNameLang":"en"}],"nameIdentifiers":[{}]},{"creatorNames":[{"creatorName":"Allam, Shehata","creatorNameLang":"en"}],"nameIdentifiers":[{}]},{"creatorNames":[{"creatorName":"Yasushi, Yagi","creatorNameLang":"en"}],"nameIdentifiers":[{}]}]},"item_4_source_id_9":{"attribute_name":"書誌レコードID","attribute_value_mlt":[{"subitem_source_identifier":"AA12628305","subitem_source_identifier_type":"NCID"}]},"item_4_textarea_12":{"attribute_name":"Notice","attribute_value_mlt":[{"subitem_textarea_value":"SIG Technical Reports are nonrefereed and hence may later appear in any journals, conferences, symposia, etc."}]},"item_resource_type":{"attribute_name":"資源タイプ","attribute_value_mlt":[{"resourceuri":"http://purl.org/coar/resource_type/c_18gh","resourcetype":"technical report"}]},"item_4_source_id_11":{"attribute_name":"ISSN","attribute_value_mlt":[{"subitem_source_identifier":"2188-8671","subitem_source_identifier_type":"ISSN"}]},"item_4_description_7":{"attribute_name":"論文抄録","attribute_value_mlt":[{"subitem_description":"In this paper, we describe a new multimodal biometric database named “OU-ISIR Multimodal Biometric Database”. An early version of this database consists of more than 100 subjects, which will increase to around 400 subjects in the future. Eleven biometric modalities are provided in this database, which, to the best of our knowledge, is the largest number of modalities among existing multimodal databases. Specifically, for each subject, we collected his/her iris, palm veins, 2D face images, signature images, gait videos, and speech data, which are typically included in existing multimodal databases. Additionally, some modalities not commonly considered in previous datasets are also included, that is, full-body images, online signature time series data, brain signals, inertial data (e.g., acceleration), and health data (e.g., heartbeat). We provide baseline results by evaluating benchmark algorithms on some individual modalities, and discuss possible future works using this database. We believe this database can facilitate future research on person authentication using unimodal, multimodal, and even cross-modal approaches, as well as research on brain signal and health status analysis.","subitem_description_type":"Other"}]},"item_4_description_8":{"attribute_name":"論文抄録(英)","attribute_value_mlt":[{"subitem_description":"In this paper, we describe a new multimodal biometric database named “OU-ISIR Multimodal Biometric Database”. An early version of this database consists of more than 100 subjects, which will increase to around 400 subjects in the future. Eleven biometric modalities are provided in this database, which, to the best of our knowledge, is the largest number of modalities among existing multimodal databases. Specifically, for each subject, we collected his/her iris, palm veins, 2D face images, signature images, gait videos, and speech data, which are typically included in existing multimodal databases. Additionally, some modalities not commonly considered in previous datasets are also included, that is, full-body images, online signature time series data, brain signals, inertial data (e.g., acceleration), and health data (e.g., heartbeat). We provide baseline results by evaluating benchmark algorithms on some individual modalities, and discuss possible future works using this database. We believe this database can facilitate future research on person authentication using unimodal, multimodal, and even cross-modal approaches, as well as research on brain signal and health status analysis.","subitem_description_type":"Other"}]},"item_4_biblio_info_10":{"attribute_name":"書誌情報","attribute_value_mlt":[{"bibliographicPageEnd":"6","bibliographic_titles":[{"bibliographic_title":"研究報告セキュリティ心理学とトラスト(SPT)"}],"bibliographicPageStart":"1","bibliographicIssueDates":{"bibliographicIssueDate":"2024-07-15","bibliographicIssueDateType":"Issued"},"bibliographicIssueNumber":"9","bibliographicVolumeNumber":"2024-SPT-56"}]},"relation_version_is_last":true,"weko_creator_id":"44499"},"id":237130,"links":{}}