<?xml version='1.0' encoding='UTF-8'?>
<OAI-PMH xmlns="http://www.openarchives.org/OAI/2.0/" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" xsi:schemaLocation="http://www.openarchives.org/OAI/2.0/ http://www.openarchives.org/OAI/2.0/OAI-PMH.xsd">
  <responseDate>2026-03-09T09:48:02Z</responseDate>
  <request metadataPrefix="jpcoar_1.0" verb="GetRecord" identifier="oai:ipsj.ixsq.nii.ac.jp:00232552">https://ipsj.ixsq.nii.ac.jp/oai</request>
  <GetRecord>
    <record>
      <header>
        <identifier>oai:ipsj.ixsq.nii.ac.jp:00232552</identifier>
        <datestamp>2025-01-19T10:24:35Z</datestamp>
        <setSpec>1164:5159:11541:11549</setSpec>
      </header>
      <metadata>
        <jpcoar:jpcoar xmlns:datacite="https://schema.datacite.org/meta/kernel-4/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:dcndl="http://ndl.go.jp/dcndl/terms/" xmlns:dcterms="http://purl.org/dc/terms/" xmlns:jpcoar="https://github.com/JPCOAR/schema/blob/master/1.0/" xmlns:oaire="http://namespace.openaire.eu/schema/oaire/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:rioxxterms="http://www.rioxx.net/schema/v2.0/rioxxterms/" xmlns:xs="http://www.w3.org/2001/XMLSchema" xmlns="https://github.com/JPCOAR/schema/blob/master/1.0/" xsi:schemaLocation="https://github.com/JPCOAR/schema/blob/master/1.0/jpcoar_scm.xsd">
          <dc:title>音光変換デバイス「ブリンキー」の分散配置に基づく発話検出のシミュレーション評価</dc:title>
          <dc:title xml:lang="en">Simulation Evaluation of Speech Detection Based on Distributed Sound-to-Light Conversion Device Blinkies</dc:title>
          <jpcoar:creator>
            <jpcoar:creatorName>本山, 智司</jpcoar:creatorName>
          </jpcoar:creator>
          <jpcoar:creator>
            <jpcoar:creatorName>植野, 夏樹</jpcoar:creatorName>
          </jpcoar:creator>
          <jpcoar:creator>
            <jpcoar:creatorName>安田, 昌弘</jpcoar:creatorName>
          </jpcoar:creator>
          <jpcoar:creator>
            <jpcoar:creatorName>木下, 裕磨</jpcoar:creatorName>
          </jpcoar:creator>
          <jpcoar:creator>
            <jpcoar:creatorName>小野, 順貴</jpcoar:creatorName>
          </jpcoar:creator>
          <jpcoar:creator>
            <jpcoar:creatorName xml:lang="en">Satoshi, Motoyama</jpcoar:creatorName>
          </jpcoar:creator>
          <jpcoar:creator>
            <jpcoar:creatorName xml:lang="en">Natsuki, Ueno</jpcoar:creatorName>
          </jpcoar:creator>
          <jpcoar:creator>
            <jpcoar:creatorName xml:lang="en">Masahiro, Yasuda</jpcoar:creatorName>
          </jpcoar:creator>
          <jpcoar:creator>
            <jpcoar:creatorName xml:lang="en">Yuma, Kinoshita</jpcoar:creatorName>
          </jpcoar:creator>
          <jpcoar:creator>
            <jpcoar:creatorName xml:lang="en">Nobutaka, Ono</jpcoar:creatorName>
          </jpcoar:creator>
          <jpcoar:subject subjectScheme="Other">ポスターセッション3 EA/SIP</jpcoar:subject>
          <datacite:description descriptionType="Other">本研究では，分散配置した音光変換デバイス「ブリンキー」を活用した発話検出を目的とする．本稿では初期検討として，ブリンキーにおける音光変換を音響シーン分類の損失に基づき End-to-End 学習するフレームワークを発話検出に応用する．音光変換に用いる deep neural network (DNN) の学習のために室内音響伝搬をシミュレーションして作成したブリンキーの観測音と正解ラベルからなるデータセットの作成を行った．作成したデータを使用して発話検出に使用するノードの個数を変更して実験を行い，提案手法においてより多くのノードを用いることで発話検出の精度が向上することを確認した．</datacite:description>
          <datacite:description descriptionType="Other">The purpose of this study is speech detection using the distributed sound-to-light conversion device Blinkies. As an initial study, this paper applies a framework for End-to-End learning of sound-to-light conversion in acoustic scene recognition using Blinkies to speech detection based on the loss of acoustic scene classiﬁcation. To train the deep neural network for sound-to-light conversion, we created a dataset consisting of Blinky’s observed sounds created by simulating room acoustic propagation and labels. Experiments were conducted using the created data with diﬀerent numbers of nodes used for speech detection, and it was conﬁrmed that the accuracy of speech detection is improved by using more nodes in the proposed method.</datacite:description>
          <dc:publisher xml:lang="ja">情報処理学会</dc:publisher>
          <datacite:date dateType="Issued">2024-02-22</datacite:date>
          <dc:language>jpn</dc:language>
          <dc:type rdf:resource="http://purl.org/coar/resource_type/c_18gh">technical report</dc:type>
          <jpcoar:identifier identifierType="URI">https://ipsj.ixsq.nii.ac.jp/records/232552</jpcoar:identifier>
          <jpcoar:sourceIdentifier identifierType="ISSN">2188-8663</jpcoar:sourceIdentifier>
          <jpcoar:sourceIdentifier identifierType="NCID">AN10442647</jpcoar:sourceIdentifier>
          <jpcoar:sourceTitle>研究報告音声言語情報処理（SLP）</jpcoar:sourceTitle>
          <jpcoar:volume>2024-SLP-151</jpcoar:volume>
          <jpcoar:issue>82</jpcoar:issue>
          <jpcoar:pageStart>1</jpcoar:pageStart>
          <jpcoar:pageEnd>6</jpcoar:pageEnd>
          <jpcoar:file>
            <jpcoar:URI label="IPSJ-SLP24151082.pdf">https://ipsj.ixsq.nii.ac.jp/record/232552/files/IPSJ-SLP24151082.pdf</jpcoar:URI>
            <jpcoar:mimeType>application/pdf</jpcoar:mimeType>
            <jpcoar:extent>2.2 MB</jpcoar:extent>
          </jpcoar:file>
        </jpcoar:jpcoar>
      </metadata>
    </record>
  </GetRecord>
</OAI-PMH>
