{"created":"2025-01-19T01:36:19.971370+00:00","metadata":{"_oai":{"id":"oai:ipsj.ixsq.nii.ac.jp:00234547","sets":["1164:1579:11464:11617"]},"path":["11617"],"owner":"44499","recid":"234547","title":["A Study of Deep Reinforcement Learning Based Resource Allocation Strategy in Multi-access Edge Computing Environment"],"pubdate":{"attribute_name":"公開日","attribute_value":"2024-06-03"},"_buckets":{"deposit":"92fc8557-e35c-4aa3-bcbc-b24b48362488"},"_deposit":{"id":"234547","pid":{"type":"depid","value":"234547","revision_id":0},"owners":[44499],"status":"published","created_by":44499},"item_title":"A Study of Deep Reinforcement Learning Based Resource Allocation Strategy in Multi-access Edge Computing Environment","author_link":["639254","639251","639256","639253","639255","639252"],"item_titles":{"attribute_name":"タイトル","attribute_value_mlt":[{"subitem_title":"A Study of Deep Reinforcement Learning Based Resource Allocation Strategy in Multi-access Edge Computing Environment"},{"subitem_title":"A Study of Deep Reinforcement Learning Based Resource Allocation Strategy in Multi-access Edge Computing Environment","subitem_title_language":"en"}]},"item_keyword":{"attribute_name":"キーワード","attribute_value_mlt":[{"subitem_subject":"低消費電力","subitem_subject_scheme":"Other"}]},"item_type_id":"4","publish_date":"2024-06-03","item_4_text_3":{"attribute_name":"著者所属","attribute_value_mlt":[{"subitem_text_value":"The University of Tokyo"},{"subitem_text_value":"The University of Tokyo"},{"subitem_text_value":"The University of Tokyo"}]},"item_4_text_4":{"attribute_name":"著者所属(英)","attribute_value_mlt":[{"subitem_text_value":"The University of Tokyo","subitem_text_language":"en"},{"subitem_text_value":"The University of Tokyo","subitem_text_language":"en"},{"subitem_text_value":"The University of Tokyo","subitem_text_language":"en"}]},"item_language":{"attribute_name":"言語","attribute_value_mlt":[{"subitem_language":"eng"}]},"item_publisher":{"attribute_name":"出版者","attribute_value_mlt":[{"subitem_publisher":"情報処理学会","subitem_publisher_language":"ja"}]},"publish_status":"0","weko_shared_id":-1,"item_file_price":{"attribute_name":"Billing file","attribute_type":"file","attribute_value_mlt":[{"url":{"url":"https://ipsj.ixsq.nii.ac.jp/record/234547/files/IPSJ-ARC24257003.pdf","label":"IPSJ-ARC24257003.pdf"},"date":[{"dateType":"Available","dateValue":"2026-06-03"}],"format":"application/pdf","billing":["billing_file"],"filename":"IPSJ-ARC24257003.pdf","filesize":[{"value":"1.5 MB"}],"mimetype":"application/pdf","priceinfo":[{"tax":["include_tax"],"price":"660","billingrole":"5"},{"tax":["include_tax"],"price":"330","billingrole":"6"},{"tax":["include_tax"],"price":"0","billingrole":"16"},{"tax":["include_tax"],"price":"0","billingrole":"44"}],"accessrole":"open_date","version_id":"5596d518-4b10-4167-b4f6-895d6f914c48","displaytype":"detail","licensetype":"license_note","license_note":"Copyright (c) 2024 by the Information Processing Society of Japan"}]},"item_4_creator_5":{"attribute_name":"著者名","attribute_type":"creator","attribute_value_mlt":[{"creatorNames":[{"creatorName":"Huilin, Li"}],"nameIdentifiers":[{}]},{"creatorNames":[{"creatorName":"Hiroshi, Nakamura"}],"nameIdentifiers":[{}]},{"creatorNames":[{"creatorName":"Hideki, Takase"}],"nameIdentifiers":[{}]}]},"item_4_creator_6":{"attribute_name":"著者名(英)","attribute_type":"creator","attribute_value_mlt":[{"creatorNames":[{"creatorName":"Huilin, Li","creatorNameLang":"en"}],"nameIdentifiers":[{}]},{"creatorNames":[{"creatorName":"Hiroshi, Nakamura","creatorNameLang":"en"}],"nameIdentifiers":[{}]},{"creatorNames":[{"creatorName":"Hideki, Takase","creatorNameLang":"en"}],"nameIdentifiers":[{}]}]},"item_4_source_id_9":{"attribute_name":"書誌レコードID","attribute_value_mlt":[{"subitem_source_identifier":"AN10096105","subitem_source_identifier_type":"NCID"}]},"item_4_textarea_12":{"attribute_name":"Notice","attribute_value_mlt":[{"subitem_textarea_value":"SIG Technical Reports are nonrefereed and hence may later appear in any journals, conferences, symposia, etc."}]},"item_resource_type":{"attribute_name":"資源タイプ","attribute_value_mlt":[{"resourceuri":"http://purl.org/coar/resource_type/c_18gh","resourcetype":"technical report"}]},"item_4_source_id_11":{"attribute_name":"ISSN","attribute_value_mlt":[{"subitem_source_identifier":"2188-8574","subitem_source_identifier_type":"ISSN"}]},"item_4_description_7":{"attribute_name":"論文抄録","attribute_value_mlt":[{"subitem_description":"In order to help user terminal devices (UTDs) efficiently handle computation-intensive and delay-sensitive computing tasks, the use of multi-access edge computing (MEC) has been proposed. However, due to shortcomings in the performance of UTDs and limited resources in networks, the optimization problem of the energy consumption of UTDs under the constraint of a certain delay condition is still a focus of research. This paper studies the optimization strategy of resource allocation for the federated learning system in the MEC network in multi-user and single-server scenarios. Firstly, we formulate the optimization problem as a constrained nonlinear programming (CNLP) problem to minimize the sum of time delay and energy consumption of UTDs under a certain delay constraint, by adjusting the local computing frequency, channel selection and power allocation dynamically. Secondly, we propose a reinforcement learning based algorithm, which uses an actor network to predict the next action based on the current state and a critic network to evaluate the quality of the predicted value. Both actor and critic networks use the gradient descent method to reduce the loss and update their network parameters. Within a limited number of training episodes, we can obtain a stable convergence solution to the original optimization problem. Finally, the performance of the proposed algorithm is compared with other classical methods.","subitem_description_type":"Other"}]},"item_4_description_8":{"attribute_name":"論文抄録(英)","attribute_value_mlt":[{"subitem_description":"In order to help user terminal devices (UTDs) efficiently handle computation-intensive and delay-sensitive computing tasks, the use of multi-access edge computing (MEC) has been proposed. However, due to shortcomings in the performance of UTDs and limited resources in networks, the optimization problem of the energy consumption of UTDs under the constraint of a certain delay condition is still a focus of research. This paper studies the optimization strategy of resource allocation for the federated learning system in the MEC network in multi-user and single-server scenarios. Firstly, we formulate the optimization problem as a constrained nonlinear programming (CNLP) problem to minimize the sum of time delay and energy consumption of UTDs under a certain delay constraint, by adjusting the local computing frequency, channel selection and power allocation dynamically. Secondly, we propose a reinforcement learning based algorithm, which uses an actor network to predict the next action based on the current state and a critic network to evaluate the quality of the predicted value. Both actor and critic networks use the gradient descent method to reduce the loss and update their network parameters. Within a limited number of training episodes, we can obtain a stable convergence solution to the original optimization problem. Finally, the performance of the proposed algorithm is compared with other classical methods.","subitem_description_type":"Other"}]},"item_4_biblio_info_10":{"attribute_name":"書誌情報","attribute_value_mlt":[{"bibliographicPageEnd":"8","bibliographic_titles":[{"bibliographic_title":"研究報告システム・アーキテクチャ(ARC)"}],"bibliographicPageStart":"1","bibliographicIssueDates":{"bibliographicIssueDate":"2024-06-03","bibliographicIssueDateType":"Issued"},"bibliographicIssueNumber":"3","bibliographicVolumeNumber":"2024-ARC-257"}]},"relation_version_is_last":true,"weko_creator_id":"44499"},"id":234547,"updated":"2025-01-19T09:46:27.249010+00:00","links":{}}