{"id":241635,"updated":"2025-01-19T07:35:47.155358+00:00","links":{},"created":"2025-01-19T01:46:21.348840+00:00","metadata":{"_oai":{"id":"oai:ipsj.ixsq.nii.ac.jp:00241635","sets":["1164:5159:11541:11870"]},"path":["11870"],"owner":"44499","recid":"241635","title":["音声感情キャプショニングのためのデータ作成とモデル構築の検討"],"pubdate":{"attribute_name":"公開日","attribute_value":"2024-12-05"},"_buckets":{"deposit":"f3c2425d-5992-4090-a772-286f3e531eb2"},"_deposit":{"id":"241635","pid":{"type":"depid","value":"241635","revision_id":0},"owners":[44499],"status":"published","created_by":44499},"item_title":"音声感情キャプショニングのためのデータ作成とモデル構築の検討","author_link":["665588","665589","665591","665590","665586","665587"],"item_titles":{"attribute_name":"タイトル","attribute_value_mlt":[{"subitem_title":"音声感情キャプショニングのためのデータ作成とモデル構築の検討"},{"subitem_title":"Study on Data Creation and Model Construction for Speech Emotion Captioning","subitem_title_language":"en"}]},"item_keyword":{"attribute_name":"キーワード","attribute_value_mlt":[{"subitem_subject":"ポスターセッション","subitem_subject_scheme":"Other"}]},"item_type_id":"4","publish_date":"2024-12-05","item_4_text_3":{"attribute_name":"著者所属","attribute_value_mlt":[{"subitem_text_value":"立命館大学大学院情報理工学研究科"},{"subitem_text_value":"立命館大学情報理工学部"},{"subitem_text_value":"立命館大学情報理工学部"}]},"item_4_text_4":{"attribute_name":"著者所属(英)","attribute_value_mlt":[{"subitem_text_value":"Graduate School of Information Science and Engineering, Ritsumeikan University","subitem_text_language":"en"},{"subitem_text_value":"Collage of Information Science and Engineering, Ritsumeikan University","subitem_text_language":"en"},{"subitem_text_value":"Collage of Information Science and Engineering, Ritsumeikan University","subitem_text_language":"en"}]},"item_language":{"attribute_name":"言語","attribute_value_mlt":[{"subitem_language":"jpn"}]},"item_publisher":{"attribute_name":"出版者","attribute_value_mlt":[{"subitem_publisher":"情報処理学会","subitem_publisher_language":"ja"}]},"publish_status":"0","weko_shared_id":-1,"item_file_price":{"attribute_name":"Billing file","attribute_type":"file","attribute_value_mlt":[{"url":{"url":"https://ipsj.ixsq.nii.ac.jp/record/241635/files/IPSJ-SLP24154015.pdf","label":"IPSJ-SLP24154015.pdf"},"format":"application/pdf","billing":["billing_file"],"filename":"IPSJ-SLP24154015.pdf","filesize":[{"value":"2.3 MB"}],"mimetype":"application/pdf","priceinfo":[{"tax":["include_tax"],"price":"0","billingrole":"22"},{"tax":["include_tax"],"price":"0","billingrole":"44"}],"accessrole":"open_login","version_id":"a1901f55-bb45-4eca-96ff-b13e8c7f46dc","displaytype":"detail","licensetype":"license_note","license_note":"Copyright (c) 2024 by the Institute of Electronics, Information and Communication Engineers This SIG report is only available to those in membership of the SIG."}]},"item_4_creator_5":{"attribute_name":"著者名","attribute_type":"creator","attribute_value_mlt":[{"creatorNames":[{"creatorName":"永瀬, 亮太郎"}],"nameIdentifiers":[{}]},{"creatorNames":[{"creatorName":"福森, 隆寛"}],"nameIdentifiers":[{}]},{"creatorNames":[{"creatorName":"山下, 洋一"}],"nameIdentifiers":[{}]}]},"item_4_creator_6":{"attribute_name":"著者名(英)","attribute_type":"creator","attribute_value_mlt":[{"creatorNames":[{"creatorName":"Ryotaro, Nagase","creatorNameLang":"en"}],"nameIdentifiers":[{}]},{"creatorNames":[{"creatorName":"Takahiro, Fukumori","creatorNameLang":"en"}],"nameIdentifiers":[{}]},{"creatorNames":[{"creatorName":"Yoichi, Yamashita","creatorNameLang":"en"}],"nameIdentifiers":[{}]}]},"item_4_source_id_9":{"attribute_name":"書誌レコードID","attribute_value_mlt":[{"subitem_source_identifier":"AN10442647","subitem_source_identifier_type":"NCID"}]},"item_4_textarea_12":{"attribute_name":"Notice","attribute_value_mlt":[{"subitem_textarea_value":"SIG Technical Reports are nonrefereed and hence may later appear in any journals, conferences, symposia, etc."}]},"item_resource_type":{"attribute_name":"資源タイプ","attribute_value_mlt":[{"resourceuri":"http://purl.org/coar/resource_type/c_18gh","resourcetype":"technical report"}]},"item_4_source_id_11":{"attribute_name":"ISSN","attribute_value_mlt":[{"subitem_source_identifier":"2188-8663","subitem_source_identifier_type":"ISSN"}]},"item_4_description_7":{"attribute_name":"論文抄録","attribute_value_mlt":[{"subitem_description":"従来の音声感情認識では,予測結果をカテゴリ感情や次元感情で表現する.このような感情の表現形式では発話者が示す複雑な感情を具体的に示すことが容易ではない.故に,本研究では感情の説明文を予測結果として出力する音声感情キャプショニングの研究に取り組む.初めに,既存の音声データに GPT4 とクラウドソーシングを用いて感情キャプションデータを付与し,人が記述した感情キャプションと比較した.次に,収集したデータを用いて, LLM を活用した音声感情キャプショニングのモデルを学習した.結果,本研究で実施した手順で収集した感情キャプションと人が記述した感情キャプションの相違点を明らかにした.また,LLM をデコーダとして用いることで音声感情キャプショニングの性能が向上することが分かった.","subitem_description_type":"Other"}]},"item_4_description_8":{"attribute_name":"論文抄録(英)","attribute_value_mlt":[{"subitem_description":"In previous studies on speech emotion recognition, the results of the prediction are represented by categorical or dimensional emotions. These descriptions of emotions make it difficult to represent the complex emotions conveyd by speech in detail. To resolve this problem, we study speech emotion captioning, which outputs an emotion caption as a prediction. Firstly, we added emotion captions to existing speech data using GPT4 and crowdsourcing, and compared them with emotion captions written by humans. Secondly, we used the collected emotion captions to train the model of speech emotion captioning with LLM. As a result, we clarified the differences between the emotion captions collected by the procedure in this study and those written by humans. We also found that using LLM as a decoder improved the performance of speech emotion captioning.","subitem_description_type":"Other"}]},"item_4_biblio_info_10":{"attribute_name":"書誌情報","attribute_value_mlt":[{"bibliographicPageEnd":"6","bibliographic_titles":[{"bibliographic_title":"研究報告音声言語情報処理(SLP)"}],"bibliographicPageStart":"1","bibliographicIssueDates":{"bibliographicIssueDate":"2024-12-05","bibliographicIssueDateType":"Issued"},"bibliographicIssueNumber":"15","bibliographicVolumeNumber":"2024-SLP-154"}]},"relation_version_is_last":true,"weko_creator_id":"44499"}}