{"metadata":{"_oai":{"id":"oai:ipsj.ixsq.nii.ac.jp:00209828","sets":["1164:4619:10416:10532"]},"path":["10532"],"owner":"44499","recid":"209828","title":["Deep Embedding Networkに基づくモノラル混合音声からの音声分離性能の向上"],"pubdate":{"attribute_name":"公開日","attribute_value":"2021-02-25"},"_buckets":{"deposit":"0570ece5-d2c6-45c1-b6d6-431690eb9226"},"_deposit":{"id":"209828","pid":{"type":"depid","value":"209828","revision_id":0},"owners":[44499],"status":"published","created_by":44499},"item_title":"Deep Embedding Networkに基づくモノラル混合音声からの音声分離性能の向上","author_link":["530030","530033","530032","530037","530036","530031","530035","530034"],"item_titles":{"attribute_name":"タイトル","attribute_value_mlt":[{"subitem_title":"Deep Embedding Networkに基づくモノラル混合音声からの音声分離性能の向上"},{"subitem_title":"Improved Speech Separation Performance from Monaural Mixed Speech Based on Deep Embedding Network","subitem_title_language":"en"}]},"item_keyword":{"attribute_name":"キーワード","attribute_value_mlt":[{"subitem_subject":"セッション4-2","subitem_subject_scheme":"Other"}]},"item_type_id":"4","publish_date":"2021-02-25","item_4_text_3":{"attribute_name":"著者所属","attribute_value_mlt":[{"subitem_text_value":"名古屋大学大学院情報学研究科"},{"subitem_text_value":"名古屋大学大学院情報学研究科"},{"subitem_text_value":"名古屋大学大学院情報学研究科"},{"subitem_text_value":"大同大学情報学部"}]},"item_4_text_4":{"attribute_name":"著者所属(英)","attribute_value_mlt":[{"subitem_text_value":"Graduate School of Informatics, Nagoya University","subitem_text_language":"en"},{"subitem_text_value":"Graduate School of Informatics, Nagoya University","subitem_text_language":"en"},{"subitem_text_value":"Graduate School of Informatics, Nagoya University","subitem_text_language":"en"},{"subitem_text_value":"School of Informatics, Daido University","subitem_text_language":"en"}]},"item_language":{"attribute_name":"言語","attribute_value_mlt":[{"subitem_language":"eng"}]},"item_publisher":{"attribute_name":"出版者","attribute_value_mlt":[{"subitem_publisher":"情報処理学会","subitem_publisher_language":"ja"}]},"publish_status":"0","weko_shared_id":-1,"item_file_price":{"attribute_name":"Billing file","attribute_type":"file","attribute_value_mlt":[{"url":{"url":"https://ipsj.ixsq.nii.ac.jp/record/209828/files/IPSJ-CVIM21225030.pdf","label":"IPSJ-CVIM21225030.pdf"},"format":"application/pdf","billing":["billing_file"],"filename":"IPSJ-CVIM21225030.pdf","filesize":[{"value":"4.5 MB"}],"mimetype":"application/pdf","priceinfo":[{"tax":["include_tax"],"price":"0","billingrole":"20"},{"tax":["include_tax"],"price":"0","billingrole":"44"}],"accessrole":"open_login","version_id":"33c6c8bf-fbfe-4a8a-84fe-62021a3a33ff","displaytype":"detail","licensetype":"license_note","license_note":"Copyright (c) 2021 by the Institute of Electronics, Information and Communication Engineers This SIG report is only available to those in membership of the SIG."}]},"item_4_creator_5":{"attribute_name":"著者名","attribute_type":"creator","attribute_value_mlt":[{"creatorNames":[{"creatorName":"党, 少翔"}],"nameIdentifiers":[{}]},{"creatorNames":[{"creatorName":"松本, 哲也"}],"nameIdentifiers":[{}]},{"creatorNames":[{"creatorName":"工藤, 博章"}],"nameIdentifiers":[{}]},{"creatorNames":[{"creatorName":"竹内, 義則"}],"nameIdentifiers":[{}]}]},"item_4_creator_6":{"attribute_name":"著者名(英)","attribute_type":"creator","attribute_value_mlt":[{"creatorNames":[{"creatorName":"Shaoxiang, Dang","creatorNameLang":"en"}],"nameIdentifiers":[{}]},{"creatorNames":[{"creatorName":"Tetsuya, Matsumoto","creatorNameLang":"en"}],"nameIdentifiers":[{}]},{"creatorNames":[{"creatorName":"Hiroaki, Kudo","creatorNameLang":"en"}],"nameIdentifiers":[{}]},{"creatorNames":[{"creatorName":"Yoshinori, Takeuchi","creatorNameLang":"en"}],"nameIdentifiers":[{}]}]},"item_4_source_id_9":{"attribute_name":"書誌レコードID","attribute_value_mlt":[{"subitem_source_identifier":"AA11131797","subitem_source_identifier_type":"NCID"}]},"item_4_textarea_12":{"attribute_name":"Notice","attribute_value_mlt":[{"subitem_textarea_value":"SIG Technical Reports are nonrefereed and hence may later appear in any journals, conferences, symposia, etc."}]},"item_resource_type":{"attribute_name":"資源タイプ","attribute_value_mlt":[{"resourceuri":"http://purl.org/coar/resource_type/c_18gh","resourcetype":"technical report"}]},"item_4_source_id_11":{"attribute_name":"ISSN","attribute_value_mlt":[{"subitem_source_identifier":"2188-8701","subitem_source_identifier_type":"ISSN"}]},"item_4_description_7":{"attribute_name":"論文抄録","attribute_value_mlt":[{"subitem_description":"音声分離は,複数の話者が同時に話している状況での発話の分離である.これに対して,deep clustering (DC) の手法として deep embedding network を用いることが提案されている.オーディオデータを低次元多様体に埋め込むことで,同様の性質を持つデータは,その空間で密に分布しており,クラスタリングアルゴリズムによって,分かれて分布するデータを容易に分離することができる.このモデルは,話者についてアノテートされたデータを示す binary mask で構成された ideal affinity matrix によって学習される.しかし,binary mask の利用が,混合スペクトログラム内の個々のビンに対して,対応するマスク内の同じ位置のビンに 0 か 1 を割り当てることになり,システム全体としてボトルネックになる.ここでは,より正確なマスクを使用することによって binary mask の欠点について改善を行う.deep embedding network を拡張し二段階処理による方法を提案する.第一段階で DC を行い,第二段階で permutation 問題を避けるための permutation invariant training の処理を行った.実験結果として,平均で,SNR で 1.55dB,SDR で 4.45dB,SDRi で 4.41dB,STOI で 0.16,PESQ で 0.3,元の DC の値より,向上が見られた.提案手法により,DC で生じたスペクトルグラム内の欠落部分を修復することも見られた.","subitem_description_type":"Other"}]},"item_4_description_8":{"attribute_name":"論文抄録(英)","attribute_value_mlt":[{"subitem_description":"Speech separation refers to the separation of utterances in which multiple people are speaking simultaneously. The idea of deep clustering (DC) is put forward by using a deep embedding network to embed audio data in the underlying manifold, and data with similar property gathers tightly in the embedding space. Then this model uses a clustering algorithm because the clustering algorithm can easily separate distributed data. Regarding the learning process, the model is supervised by an ideal affinity matrix constructed of binary masks of annotation data. However, the binary mask gives a bottleneck to the entire system since the same position of bins in masks are assigned to 0 or 1 according to the contribution of individual utterances to mixed spectrograms. Thus, we propose an extended two-stage version of network based on the deep embedding. The network can eliminate the shortcomings by using various more accurate masks. We employ DC as our first stage, and conduct a permutation invariant training approach to prevent permutation problem in the second stage. As a result, the results according to our experiment outperforms the original DC model by 1.55dB in SNR by 4.45dB in SDR, 4.41dB in SDRi, 0.16 in STOI, and 0.3 in PESQ on average. We also observe that the proposed method can recover the defects in the spectrograms brought in by DC.","subitem_description_type":"Other"}]},"item_4_biblio_info_10":{"attribute_name":"書誌情報","attribute_value_mlt":[{"bibliographicPageEnd":"6","bibliographic_titles":[{"bibliographic_title":"研究報告コンピュータビジョンとイメージメディア(CVIM)"}],"bibliographicPageStart":"1","bibliographicIssueDates":{"bibliographicIssueDate":"2021-02-25","bibliographicIssueDateType":"Issued"},"bibliographicIssueNumber":"30","bibliographicVolumeNumber":"2021-CVIM-225"}]},"relation_version_is_last":true,"weko_creator_id":"44499"},"id":209828,"updated":"2025-01-19T18:22:38.860153+00:00","links":{},"created":"2025-01-19T01:11:06.773296+00:00"}