Analyzing Transformers via Value Matrices
-
- Yoshida Minoru
- Tokushima University
-
- Matsumoto Kazuyuki
- Tokushima University
-
- Kita Kenji
- Tokushima University
Bibliographic Information
- Other Title
-
- Value行列を手掛かりとした Transformerの分析
Description
<p>We propose a new method to analyze Transformer language models. In Transformer self-attention modules, attention weights are calculated from the query vectors and key vectors. Then, output vectors are obtained by taking the weighted sum of value vectors. While existing works on analysis of Transformer have focused on attention weights, this work focused on value and output matrices. We obtain joint matrices by multiplying both matrices, and show that the trace of the joint matrices are correlated with word co-occurences.</p>
Journal
-
- Transactions of the Japanese Society for Artificial Intelligence
-
Transactions of the Japanese Society for Artificial Intelligence 38 (2), n/a-, 2023-03-01
The Japanese Society for Artificial Intelligence
- Tweet
Keywords
Details 詳細情報について
-
- CRID
- 1390013795251121280
-
- ISSN
- 13468030
- 13460714
-
- Text Lang
- ja
-
- Data Source
-
- JaLC
- Crossref
- KAKEN
- OpenAIRE
-
- Abstract License Flag
- Disallowed