ログイン 新規登録
言語:

WEKO3

  • トップ
  • ランキング
To
lat lon distance
To

Field does not validate



インデックスリンク

インデックスツリー

メールアドレスを入力してください。

WEKO

One fine body…

WEKO

One fine body…

アイテム

  1. 論文誌(トランザクション)
  2. プログラミング(PRO)
  3. Vol.10
  4. No.4

Performance Modeling of Task Parallel Programs

https://ipsj.ixsq.nii.ac.jp/records/182752
https://ipsj.ixsq.nii.ac.jp/records/182752
56f36c05-3a43-4334-b989-078922b8ef6d
名前 / ファイル ライセンス アクション
IPSJ-TPRO1004005.pdf IPSJ-TPRO1004005.pdf (29.5 kB)
Copyright (c) 2017 by the Information Processing Society of Japan
オープンアクセス
Item type Trans(1)
公開日 2017-07-21
タイトル
タイトル Performance Modeling of Task Parallel Programs
タイトル
言語 en
タイトル Performance Modeling of Task Parallel Programs
言語
言語 eng
キーワード
主題Scheme Other
主題 [発表概要]
資源タイプ
資源タイプ識別子 http://purl.org/coar/resource_type/c_6501
資源タイプ journal article
著者所属
Graduate School of Information Science and Technology, The University of Tokyo
著者所属
Graduate School of Information Science and Technology, The University of Tokyo
著者所属(英)
en
Graduate School of Information Science and Technology, The University of Tokyo
著者所属(英)
en
Graduate School of Information Science and Technology, The University of Tokyo
著者名 Byambajav, Namsraijav

× Byambajav, Namsraijav

Byambajav, Namsraijav

Search repository
Kenjiro, Taura

× Kenjiro, Taura

Kenjiro, Taura

Search repository
著者名(英) Byambajav, Namsraijav

× Byambajav, Namsraijav

en Byambajav, Namsraijav

Search repository
Kenjiro, Taura

× Kenjiro, Taura

en Kenjiro, Taura

Search repository
論文抄録
内容記述タイプ Other
内容記述 To estimate the execution time of an application for given parameters on a given hardware, people often use analytical models if they are familiar with the application logic. But such approaches become increasingly difficult as the application becomes complicated and the number of input parameters increase. Besides for task parallel applications, the performance is highly dependent on the characteristics of the underlying dynamic task scheduler runtime. Therefore, it is extremely challenging to formulate analytical models for task parallel applications as the performance depends on not only the application logic but also the underlying hardware properties and the task parallel runtime's performance. Machine learning techniques can be used to build a performance model when formulating reliable analytical model is unfeasible. First, we run the target application multiple times to learn the execution time for differing input parameters and worker counts. Then a machine learning model is built using that training data. However, past such models, mainly developed for well load-balanced applications, perform poorly when applied to task parallel applications where there is much load-imbalance. Also, the accuracy of those models significantly decreases when the number of workers used in the prediction target execution becomes bigger than the number of maximum workers used during the training. We investigate whether applying modern machine learning techniques can address these challenges. In this presentation, we build performance models for task parallel programs using Lars-Lasso and deep neural network regression. We also exploit some additional information which we can gather during the training, such as the hardware performance counter values, the number of tasks created, and the task-stealing overhead, to guide these models. We evaluate the proposed models for BOTS and PARSEC benchmark applications running on multiple task parallel runtimes.
論文抄録(英)
内容記述タイプ Other
内容記述 To estimate the execution time of an application for given parameters on a given hardware, people often use analytical models if they are familiar with the application logic. But such approaches become increasingly difficult as the application becomes complicated and the number of input parameters increase. Besides for task parallel applications, the performance is highly dependent on the characteristics of the underlying dynamic task scheduler runtime. Therefore, it is extremely challenging to formulate analytical models for task parallel applications as the performance depends on not only the application logic but also the underlying hardware properties and the task parallel runtime's performance. Machine learning techniques can be used to build a performance model when formulating reliable analytical model is unfeasible. First, we run the target application multiple times to learn the execution time for differing input parameters and worker counts. Then a machine learning model is built using that training data. However, past such models, mainly developed for well load-balanced applications, perform poorly when applied to task parallel applications where there is much load-imbalance. Also, the accuracy of those models significantly decreases when the number of workers used in the prediction target execution becomes bigger than the number of maximum workers used during the training. We investigate whether applying modern machine learning techniques can address these challenges. In this presentation, we build performance models for task parallel programs using Lars-Lasso and deep neural network regression. We also exploit some additional information which we can gather during the training, such as the hardware performance counter values, the number of tasks created, and the task-stealing overhead, to guide these models. We evaluate the proposed models for BOTS and PARSEC benchmark applications running on multiple task parallel runtimes.
書誌レコードID
収録物識別子タイプ NCID
収録物識別子 AA11464814
書誌情報 情報処理学会論文誌プログラミング(PRO)

巻 10, 号 4, p. 28-28, 発行日 2017-07-21
ISSN
収録物識別子タイプ ISSN
収録物識別子 1882-7802
出版者
言語 ja
出版者 情報処理学会
戻る
0
views
See details
Views

Versions

Ver.1 2025-01-20 03:57:46.789513
Show All versions

Share

Mendeley Twitter Facebook Print Addthis

Cite as

エクスポート

OAI-PMH
  • OAI-PMH JPCOAR
  • OAI-PMH DublinCore
  • OAI-PMH DDI
Other Formats
  • JSON
  • BIBTEX

Confirm


Powered by WEKO3


Powered by WEKO3