{"updated":"2025-01-21T18:55:56.758082+00:00","metadata":{"_oai":{"id":"oai:ipsj.ixsq.nii.ac.jp:00082612","sets":["581:6644:6801"]},"path":["6801"],"owner":"11","recid":"82612","title":["マルコフ決定過程のロールプレイングゲームへの適用"],"pubdate":{"attribute_name":"公開日","attribute_value":"2012-06-15"},"_buckets":{"deposit":"5a47d78d-41d9-48a4-8646-b5da73c707b4"},"_deposit":{"id":"82612","pid":{"type":"depid","value":"82612","revision_id":0},"owners":[11],"status":"published","created_by":11},"item_title":"マルコフ決定過程のロールプレイングゲームへの適用","author_link":["0","0"],"item_titles":{"attribute_name":"タイトル","attribute_value_mlt":[{"subitem_title":"マルコフ決定過程のロールプレイングゲームへの適用"},{"subitem_title":"Applying Markov Decision Processes to Role-playing Game","subitem_title_language":"en"}]},"item_keyword":{"attribute_name":"キーワード","attribute_value_mlt":[{"subitem_subject":"[特集:パズルの数理] マルコフ決定過程,ロールプレイングゲーム,統計的決定理論,動的計画法,ベイズ基準","subitem_subject_scheme":"Other"}]},"item_type_id":"2","publish_date":"2012-06-15","item_2_text_3":{"attribute_name":"著者所属","attribute_value_mlt":[{"subitem_text_value":"北見工業大学"},{"subitem_text_value":"北見工業大学"},{"subitem_text_value":"北見工業大学"},{"subitem_text_value":"北見工業大学"},{"subitem_text_value":"北見工業大学"}]},"item_2_text_4":{"attribute_name":"著者所属(英)","attribute_value_mlt":[{"subitem_text_value":"Kitami Institute of Technology","subitem_text_language":"en"},{"subitem_text_value":"Kitami Institute of Technology","subitem_text_language":"en"},{"subitem_text_value":"Kitami Institute of Technology","subitem_text_language":"en"},{"subitem_text_value":"Kitami Institute of Technology","subitem_text_language":"en"},{"subitem_text_value":"Kitami Institute of Technology","subitem_text_language":"en"}]},"item_language":{"attribute_name":"言語","attribute_value_mlt":[{"subitem_language":"jpn"}]},"publish_status":"0","weko_shared_id":-1,"item_file_price":{"attribute_name":"Billing file","attribute_type":"file","attribute_value_mlt":[{"url":{"url":"https://ipsj.ixsq.nii.ac.jp/record/82612/files/IPSJ-JNL5306015.pdf"},"date":[{"dateType":"Available","dateValue":"2014-06-15"}],"format":"application/pdf","billing":["billing_file"],"filename":"IPSJ-JNL5306015.pdf","filesize":[{"value":"599.8 kB"}],"mimetype":"application/pdf","priceinfo":[{"tax":["include_tax"],"price":"660","billingrole":"5"},{"tax":["include_tax"],"price":"330","billingrole":"6"},{"tax":["include_tax"],"price":"0","billingrole":"8"},{"tax":["include_tax"],"price":"0","billingrole":"44"}],"accessrole":"open_date","version_id":"69f4a7f9-9ae4-49d5-a660-ff5e041c8d13","displaytype":"detail","licensetype":"license_note","license_note":"Copyright (c) 2012 by the Information Processing Society of Japan"}]},"item_2_creator_5":{"attribute_name":"著者名","attribute_type":"creator","attribute_value_mlt":[{"creatorNames":[{"creatorName":"前田, 康成"},{"creatorName":"後藤文太朗"},{"creatorName":"升井, 洋志"},{"creatorName":"桝井, 文人"},{"creatorName":"鈴木, 正清"}],"nameIdentifiers":[{}]}]},"item_2_creator_6":{"attribute_name":"著者名(英)","attribute_type":"creator","attribute_value_mlt":[{"creatorNames":[{"creatorName":"Yasunari, Maeda","creatorNameLang":"en"},{"creatorName":"Fumitaro, Goto","creatorNameLang":"en"},{"creatorName":"Hiroshi, Masui","creatorNameLang":"en"},{"creatorName":"Fumito, Masui","creatorNameLang":"en"},{"creatorName":"Masakiyo, Suzuki","creatorNameLang":"en"}],"nameIdentifiers":[{}]}]},"item_2_source_id_9":{"attribute_name":"書誌レコードID","attribute_value_mlt":[{"subitem_source_identifier":"AN00116647","subitem_source_identifier_type":"NCID"}]},"item_resource_type":{"attribute_name":"資源タイプ","attribute_value_mlt":[{"resourceuri":"http://purl.org/coar/resource_type/c_6501","resourcetype":"journal article"}]},"item_2_source_id_11":{"attribute_name":"ISSN","attribute_value_mlt":[{"subitem_source_identifier":"1882-7764","subitem_source_identifier_type":"ISSN"}]},"item_2_description_7":{"attribute_name":"論文抄録","attribute_value_mlt":[{"subitem_description":"従来からマルコフ決定過程(MDP)を用いたロールプレイングゲーム(RPG)のモデル化が行われている.従来研究ではRPGが部分的にモデル化されている.本研究では,MDPを用いてより一般的なRPGのモデル化を行う.最初にMDPの真のパラメータ既知の場合に相当するRPGについて,報酬の期待値を最大にするアルゴリズムを提案する.次にMDPの真のパラメータ未知の場合に相当するRPGについて,ベイズ基準のもとで報酬を最大にするアルゴリズムを提案する.次にMDPの真のパラメータ未知の場合に相当するRPGについて,学習データを用いて報酬を近似的に最大にするアルゴリズムを提案する.","subitem_description_type":"Other"}]},"item_2_description_8":{"attribute_name":"論文抄録(英)","attribute_value_mlt":[{"subitem_description":"In previous research a part of role-playing game (RPG) is represented with Markov decision processes (MDP). In this research we represent a more general PRG with MDP. We maximize an expected total reward under the condition that the true parameter of MDP is known in the first proposition. We maximize an expected total reward with respect to a Bayes criterion under the condition that the true parameter of MDP is unknown in the second proposition. We approximately maximize an expected total reward using learning data under the condition that the true parameter of MDP is unknown in the third proposition.","subitem_description_type":"Other"}]},"item_2_biblio_info_10":{"attribute_name":"書誌情報","attribute_value_mlt":[{"bibliographicPageEnd":"1616","bibliographic_titles":[{"bibliographic_title":"情報処理学会論文誌"}],"bibliographicPageStart":"1608","bibliographicIssueDates":{"bibliographicIssueDate":"2012-06-15","bibliographicIssueDateType":"Issued"},"bibliographicIssueNumber":"6","bibliographicVolumeNumber":"53"}]},"relation_version_is_last":true,"weko_creator_id":"11"},"created":"2025-01-18T23:36:27.091896+00:00","id":82612,"links":{}}