Softmax Bottleneck Makes Language Models Unable to Represent Multi-mode Word Distributions
H. Chang, and A. McCallum. Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), page 8048--8073. Dublin, Ireland, Association for Computational Linguistics, (May 2022)
DOI: 10.18653/v1/2022.acl-long.554
Abstract
Neural language models (LMs) such as GPT-2 estimate the probability distribution over the next word by a softmax over the vocabulary. The softmax layer produces the distribution based on the dot products of a single hidden state and the embeddings of words in the vocabulary. However, we discover that this single hidden state cannot produce all probability distributions regardless of the LM size or training data size because the single hidden state embedding cannot be close to the embeddings of all the possible next words simultaneously when there are other interfering word embeddings between them. In this work, we demonstrate the importance of this limitation both theoretically and practically. Our work not only deepens our understanding of softmax bottleneck and mixture of softmax (MoS) but also inspires us to propose multi-facet softmax (MFS) to address the limitations of MoS. Extensive empirical analyses confirm our findings and show that against MoS, the proposed MFS achieves two-fold improvements in the perplexity of GPT-2 and BERT.
%0 Conference Paper
%1 chang-mccallum-2022-softmax
%A Chang, Haw-Shiuan
%A McCallum, Andrew
%B Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)
%C Dublin, Ireland
%D 2022
%E Muresan, Smaranda
%E Nakov, Preslav
%E Villavicencio, Aline
%I Association for Computational Linguistics
%K llm nlp reading softmax
%P 8048--8073
%R 10.18653/v1/2022.acl-long.554
%T Softmax Bottleneck Makes Language Models Unable to Represent Multi-mode Word Distributions
%U https://aclanthology.org/2022.acl-long.554
%X Neural language models (LMs) such as GPT-2 estimate the probability distribution over the next word by a softmax over the vocabulary. The softmax layer produces the distribution based on the dot products of a single hidden state and the embeddings of words in the vocabulary. However, we discover that this single hidden state cannot produce all probability distributions regardless of the LM size or training data size because the single hidden state embedding cannot be close to the embeddings of all the possible next words simultaneously when there are other interfering word embeddings between them. In this work, we demonstrate the importance of this limitation both theoretically and practically. Our work not only deepens our understanding of softmax bottleneck and mixture of softmax (MoS) but also inspires us to propose multi-facet softmax (MFS) to address the limitations of MoS. Extensive empirical analyses confirm our findings and show that against MoS, the proposed MFS achieves two-fold improvements in the perplexity of GPT-2 and BERT.
@inproceedings{chang-mccallum-2022-softmax,
abstract = {Neural language models (LMs) such as GPT-2 estimate the probability distribution over the next word by a softmax over the vocabulary. The softmax layer produces the distribution based on the dot products of a single hidden state and the embeddings of words in the vocabulary. However, we discover that this single hidden state cannot produce all probability distributions regardless of the LM size or training data size because the single hidden state embedding cannot be close to the embeddings of all the possible next words simultaneously when there are other interfering word embeddings between them. In this work, we demonstrate the importance of this limitation both theoretically and practically. Our work not only deepens our understanding of softmax bottleneck and mixture of softmax (MoS) but also inspires us to propose multi-facet softmax (MFS) to address the limitations of MoS. Extensive empirical analyses confirm our findings and show that against MoS, the proposed MFS achieves two-fold improvements in the perplexity of GPT-2 and BERT.},
added-at = {2024-04-19T09:30:29.000+0200},
address = {Dublin, Ireland},
author = {Chang, Haw-Shiuan and McCallum, Andrew},
biburl = {https://www.bibsonomy.org/bibtex/23ec496ab62f4df8106616fbe3eef23f6/tobias.koopmann},
booktitle = {Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)},
doi = {10.18653/v1/2022.acl-long.554},
editor = {Muresan, Smaranda and Nakov, Preslav and Villavicencio, Aline},
interhash = {ce2e301c39be360a0eeec21a6439d24a},
intrahash = {3ec496ab62f4df8106616fbe3eef23f6},
keywords = {llm nlp reading softmax},
month = may,
pages = {8048--8073},
publisher = {Association for Computational Linguistics},
timestamp = {2024-04-19T09:30:29.000+0200},
title = {Softmax Bottleneck Makes Language Models Unable to Represent Multi-mode Word Distributions},
url = {https://aclanthology.org/2022.acl-long.554},
year = 2022
}