{"id":241600,"links":{},"created":"2025-01-19T01:46:18.027815+00:00","metadata":{"_oai":{"id":"oai:ipsj.ixsq.nii.ac.jp:00241600","sets":["1164:4179:11560:11869"]},"path":["11869"],"owner":"44499","recid":"241600","title":["自己教師あり学習モデルとT5モデルを用いたG2Pフリー音声合成に関する研究"],"pubdate":{"attribute_name":"公開日","attribute_value":"2024-12-05"},"_buckets":{"deposit":"eb241803-ee8b-4f8b-86e4-d27ee3920608"},"_deposit":{"id":"241600","pid":{"type":"depid","value":"241600","revision_id":0},"owners":[44499],"status":"published","created_by":44499},"item_title":"自己教師あり学習モデルとT5モデルを用いたG2Pフリー音声合成に関する研究","author_link":["665434","665433","665432"],"item_titles":{"attribute_name":"タイトル","attribute_value_mlt":[{"subitem_title":"自己教師あり学習モデルとT5モデルを用いたG2Pフリー音声合成に関する研究"},{"subitem_title":"A Study on G2P-free Speech Synthesis using Speech Self-Supervised Learning and T5 Model","subitem_title_language":"en"}]},"item_keyword":{"attribute_name":"キーワード","attribute_value_mlt":[{"subitem_subject":"特別セッション:分野横断","subitem_subject_scheme":"Other"}]},"item_type_id":"4","publish_date":"2024-12-05","item_4_text_3":{"attribute_name":"著者所属","attribute_value_mlt":[{"subitem_text_value":"東京大学大学院電気系工学専攻"},{"subitem_text_value":"東京大学大学院電気系工学専攻"},{"subitem_text_value":"東京大学大学院電気系工学専攻"}]},"item_language":{"attribute_name":"言語","attribute_value_mlt":[{"subitem_language":"jpn"}]},"item_publisher":{"attribute_name":"出版者","attribute_value_mlt":[{"subitem_publisher":"情報処理学会","subitem_publisher_language":"ja"}]},"publish_status":"0","weko_shared_id":-1,"item_file_price":{"attribute_name":"Billing file","attribute_type":"file","attribute_value_mlt":[{"url":{"url":"https://ipsj.ixsq.nii.ac.jp/record/241600/files/IPSJ-NL24262025.pdf","label":"IPSJ-NL24262025.pdf"},"date":[{"dateType":"Available","dateValue":"2026-12-05"}],"format":"application/pdf","billing":["billing_file"],"filename":"IPSJ-NL24262025.pdf","filesize":[{"value":"1.6 MB"}],"mimetype":"application/pdf","priceinfo":[{"tax":["include_tax"],"price":"660","billingrole":"5"},{"tax":["include_tax"],"price":"330","billingrole":"6"},{"tax":["include_tax"],"price":"0","billingrole":"23"},{"tax":["include_tax"],"price":"0","billingrole":"44"}],"accessrole":"open_date","version_id":"19088ab4-9ccd-477d-9459-d38a53a09e9c","displaytype":"detail","licensetype":"license_note","license_note":"Copyright (c) 2024 by the Information Processing Society of Japan"}]},"item_4_creator_5":{"attribute_name":"著者名","attribute_type":"creator","attribute_value_mlt":[{"creatorNames":[{"creatorName":"朴, 浚鎔"}],"nameIdentifiers":[{}]},{"creatorNames":[{"creatorName":"齋藤, 大輔"}],"nameIdentifiers":[{}]},{"creatorNames":[{"creatorName":"峯松, 信明"}],"nameIdentifiers":[{}]}]},"item_4_source_id_9":{"attribute_name":"書誌レコードID","attribute_value_mlt":[{"subitem_source_identifier":"AN10115061","subitem_source_identifier_type":"NCID"}]},"item_4_textarea_12":{"attribute_name":"Notice","attribute_value_mlt":[{"subitem_textarea_value":"SIG Technical Reports are nonrefereed and hence may later appear in any journals, conferences, symposia, etc."}]},"item_resource_type":{"attribute_name":"資源タイプ","attribute_value_mlt":[{"resourceuri":"http://purl.org/coar/resource_type/c_18gh","resourcetype":"technical report"}]},"item_4_source_id_11":{"attribute_name":"ISSN","attribute_value_mlt":[{"subitem_source_identifier":"2188-8779","subitem_source_identifier_type":"ISSN"}]},"item_4_description_7":{"attribute_name":"論文抄録","attribute_value_mlt":[{"subitem_description":"本研究は,従来のグラフェム-音素変換(G2P)を代替する,音声合成における新しいアプローチを提案する.具体的には,音声から直接離散的なトークンを生成するディープラーニングベースのモデルを使用する.事前学習された音声自己教師あり学習(SSL)モデルを活用し,T5 エンコーダを訓練して,漢字や仮名などの混在スクリプトから疑似的な言語ラベルを生成する.この手法により,手動による音声表記の転写が不要となり,特に未転写の大規模音声データセットにおいてコスト削減とスケーラビリティの向上が実現される.本モデルは,従来の G2P ベースのテキスト音声合成システムと同等のパフォーマンスを達成し,アクセントやイントネーションなど,自然な言語的およびパラ言語的特徴を保持した音声を合成することが可能である.","subitem_description_type":"Other"}]},"item_4_description_8":{"attribute_name":"論文抄録(英)","attribute_value_mlt":[{"subitem_description":"This study presents a novel approach to voice synthesis that can substitute the traditional grapheme-to-phoneme (G2P) conversion by using a deep learning-based model that generates discrete tokens directly from speech. Utilizing a pre-trained voice SSL model, we train a T5 encoder to produce pseudo-language labels from mixed-script texts (e.g., containing Kanji and Kana). This method eliminates the need for manual phonetic transcription, reducing costs and enhancing scalability, especially for large non-transcribed audio datasets. Our model matches the performance of conventional G2P-based text-to-speech systems and is capable of synthesizing speech that retains natural linguistic and paralinguistic features, such as accents and intonations.","subitem_description_type":"Other"}]},"item_4_biblio_info_10":{"attribute_name":"書誌情報","attribute_value_mlt":[{"bibliographicPageEnd":"6","bibliographic_titles":[{"bibliographic_title":"研究報告自然言語処理(NL)"}],"bibliographicPageStart":"1","bibliographicIssueDates":{"bibliographicIssueDate":"2024-12-05","bibliographicIssueDateType":"Issued"},"bibliographicIssueNumber":"25","bibliographicVolumeNumber":"2024-NL-262"}]},"relation_version_is_last":true,"weko_creator_id":"44499"},"updated":"2025-01-19T07:36:29.313470+00:00"}