{"id":235093,"updated":"2025-01-19T09:36:50.444318+00:00","links":{},"created":"2025-01-19T01:37:05.130275+00:00","metadata":{"_oai":{"id":"oai:ipsj.ixsq.nii.ac.jp:00235093","sets":["1164:4179:11560:11676"]},"path":["11676"],"owner":"44499","recid":"235093","title":["文字言語モデルからの単語言語モデルの教師なし合成"],"pubdate":{"attribute_name":"公開日","attribute_value":"2024-06-21"},"_buckets":{"deposit":"ecf9c117-834f-49d0-867e-1a515d2e9bc4"},"_deposit":{"id":"235093","pid":{"type":"depid","value":"235093","revision_id":0},"owners":[44499],"status":"published","created_by":44499},"item_title":"文字言語モデルからの単語言語モデルの教師なし合成","author_link":["641850"],"item_titles":{"attribute_name":"タイトル","attribute_value_mlt":[{"subitem_title":"文字言語モデルからの単語言語モデルの教師なし合成"},{"subitem_title":"Unsupervised Synthesis of Word Language Models from Pretrained Character Language Models","subitem_title_language":"en"}]},"item_keyword":{"attribute_name":"キーワード","attribute_value_mlt":[{"subitem_subject":"LLM基礎","subitem_subject_scheme":"Other"}]},"item_type_id":"4","publish_date":"2024-06-21","item_4_text_3":{"attribute_name":"著者所属","attribute_value_mlt":[{"subitem_text_value":"京都大学"}]},"item_language":{"attribute_name":"言語","attribute_value_mlt":[{"subitem_language":"jpn"}]},"item_publisher":{"attribute_name":"出版者","attribute_value_mlt":[{"subitem_publisher":"情報処理学会","subitem_publisher_language":"ja"}]},"publish_status":"0","weko_shared_id":-1,"item_file_price":{"attribute_name":"Billing file","attribute_type":"file","attribute_value_mlt":[{"url":{"url":"https://ipsj.ixsq.nii.ac.jp/record/235093/files/IPSJ-NL24260002.pdf","label":"IPSJ-NL24260002.pdf"},"date":[{"dateType":"Available","dateValue":"2026-06-21"}],"format":"application/pdf","billing":["billing_file"],"filename":"IPSJ-NL24260002.pdf","filesize":[{"value":"1.4 MB"}],"mimetype":"application/pdf","priceinfo":[{"tax":["include_tax"],"price":"660","billingrole":"5"},{"tax":["include_tax"],"price":"330","billingrole":"6"},{"tax":["include_tax"],"price":"0","billingrole":"23"},{"tax":["include_tax"],"price":"0","billingrole":"44"}],"accessrole":"open_date","version_id":"cb371a98-6780-453c-9376-f5f897572b7a","displaytype":"detail","licensetype":"license_note","license_note":"Copyright (c) 2024 by the Information Processing Society of Japan"}]},"item_4_creator_5":{"attribute_name":"著者名","attribute_type":"creator","attribute_value_mlt":[{"creatorNames":[{"creatorName":"村脇, 有吾"}],"nameIdentifiers":[{}]}]},"item_4_source_id_9":{"attribute_name":"書誌レコードID","attribute_value_mlt":[{"subitem_source_identifier":"AN10115061","subitem_source_identifier_type":"NCID"}]},"item_4_textarea_12":{"attribute_name":"Notice","attribute_value_mlt":[{"subitem_textarea_value":"SIG Technical Reports are nonrefereed and hence may later appear in any journals, conferences, symposia, etc."}]},"item_resource_type":{"attribute_name":"資源タイプ","attribute_value_mlt":[{"resourceuri":"http://purl.org/coar/resource_type/c_18gh","resourcetype":"technical report"}]},"item_4_source_id_11":{"attribute_name":"ISSN","attribute_value_mlt":[{"subitem_source_identifier":"2188-8779","subitem_source_identifier_type":"ISSN"}]},"item_4_description_7":{"attribute_name":"論文抄録","attribute_value_mlt":[{"subitem_description":"Transformer に基づく事前訓練文字言語モデルから単語言語モデルへの教師なし合成の実現可能性を示す.教師なし単語分割における興味の中心は学習を実現する帰納バイアスを明らかにすることだが,ニューラル言語モデルに基づく場合,アーキテクチャ上の制約から有限語彙を前提とするにもかかわらず,教師なし単語分割においては語彙を事前に決定できないという問題への取り組みを中心に据えざるを得ない.本稿では,この問題を解決するために多段階の訓練手続きを提案する.提案手法は単語境界における確率的不確実性を取っ掛かりとして利用しており,幼児の初期学習との関連が示唆される.","subitem_description_type":"Other"}]},"item_4_description_8":{"attribute_name":"論文抄録(英)","attribute_value_mlt":[{"subitem_description":"We demonstrate the feasibility of transforming a Transformer-based pretrained character language model into a word language model without explicit supervision on word segmentation. While the main interest in unsupervised word segmentation lies in identifying the inductive biases that facilitate language acquisition, neural language models face technical challenges due to their architectural constraints requiring a fixed vocabulary, despite the inability to predefine this vocabulary in unsupervised settings. To address this issue, we propose a multi-stage training procedure. Our method leverages the stochastic uncertainty pertaining to word boundaries to bootstrap the process, suggesting a connection to early learning in infants.","subitem_description_type":"Other"}]},"item_4_biblio_info_10":{"attribute_name":"書誌情報","attribute_value_mlt":[{"bibliographicPageEnd":"14","bibliographic_titles":[{"bibliographic_title":"研究報告自然言語処理(NL)"}],"bibliographicPageStart":"1","bibliographicIssueDates":{"bibliographicIssueDate":"2024-06-21","bibliographicIssueDateType":"Issued"},"bibliographicIssueNumber":"2","bibliographicVolumeNumber":"2024-NL-260"}]},"relation_version_is_last":true,"weko_creator_id":"44499"}}