ログイン
Language:

WEKO3

  • トップ
  • ランキング
To
lat lon distance
To

Field does not validate



インデックスリンク

インデックスツリー

メールアドレスを入力してください。

WEKO

One fine body…

WEKO

One fine body…

アイテム

  1. 02 情報科学
  2. 02 国際会議論文

Multi-label Learning with Random Circular Vectors

http://hdl.handle.net/10061/0002001118
http://hdl.handle.net/10061/0002001118
df7101a4-33d9-49dd-a41f-a32313499119
アイテムタイプ 会議発表論文 / Conference Paper(1)
公開日 2025-08-18
タイトル
タイトル Multi-label Learning with Random Circular Vectors
言語
言語 eng
資源タイプ
資源タイプ conference paper
アクセス権
アクセス権 open access
著者 Nishida, Ken

× Nishida, Ken

en Nishida, Ken

Search repository
Machi, Kojiro

× Machi, Kojiro

en Machi, Kojiro

Search repository
Onishi, Kazuma

× Onishi, Kazuma

en Onishi, Kazuma

Search repository
Hayashi, Katsuhiko

× Hayashi, Katsuhiko

en Hayashi, Katsuhiko

Search repository
上垣外, 英剛

× 上垣外, 英剛

ja 上垣外, 英剛

ja-Kana カミガイト, ヒデタカ

en Kamigaito, Hidetaka

Search repository
抄録
内容記述タイプ Abstract
内容記述 The extreme multi-label classification (XMC) task involves learning a classifier that can predict from a large label set the most relevant subset of labels for a data instance. While deep neural networks (DNNs) have demonstrated remarkable success in XMC problems, the task is still challenging because it must deal with a large number of output labels, which make the DNN training computationally expensive. This paper addresses the issue by exploring the use of random circular vectors, where each vector component is represented as a complex amplitude. In our framework, we can develop an output layer and loss function of DNNs for XMC by representing the final output layer as a fully connected layer that directly predicts a low-dimensional circular vector encoding a set of labels for a data instance. We conducted experiments on synthetic datasets to verify that circular vectors have better label encoding capacity and retrieval ability than normal real-valued vectors. Then, we conducted experiments on actual XMC datasets and found that these appealing properties of circular vectors contribute to significant improvements in task performance compared with a previous model using random real-valued vectors, while reducing the size of the output layers by up to 99%.
書誌情報 en : Proceedings of the 9th Workshop on Representation Learning for NLP (RepL4NLP-2024)

p. 245-255, ページ数 11, 発行日 2024-08-15
会議情報
会議名 The 9th Workshop on Representation Learning for NLP (RepL4NLP-2024)
開始年 2024
開始月 08
開始日 15
終了年 2024
終了月 08
終了日 15
開催期間 2024-08-15 - 2024-08-15
開催地 Bangkok, Thailand
開催国 THA
出版者
出版者 Association for Computational Linguistics
出版者版URI
関連タイプ isReplacedBy
識別子タイプ URI
関連識別子 https://aclanthology.org/2024.repl4nlp-1.18/
権利
権利情報Resource https://creativecommons.org/licenses/by/4.0/
権利情報 $00A92024 Association for Computational Linguistics. ACL materials are Copyright $00A9 1963$20132025 ACL; other materials are copyrighted by their respective copyright holders. Materials prior to 2016 here are licensed under the Creative Commons Attribution-NonCommercial-ShareAlike 3.0 International License. Permission is granted to make copies for the purposes of teaching and research. Materials published in or after 2016 are licensed on a Creative Commons Attribution 4.0 International License.
著者版フラグ
出版タイプ NA
戻る
0
views
See details
Views

Versions

Ver.1 2025-08-18 04:12:59.471583
Show All versions

Share

Share
tweet

Cite as

Other

print

エクスポート

OAI-PMH
  • OAI-PMH JPCOAR 2.0
  • OAI-PMH JPCOAR 1.0
  • OAI-PMH DublinCore
  • OAI-PMH DDI
Other Formats
  • JSON
  • BIBTEX
  • ZIP

コミュニティ

確認

確認

確認


Powered by WEKO3


Powered by WEKO3