It has been argued that BERT ``rediscovers the traditional NLP
pipeline'', with lower layers extracting morphosyntactic features and
higher layers creating holistic sentence-level representations.
In this paper, we critically examine this assumption through a
principle-component-guided analysis, extracing sets of inputs that
correspond to specific activation patterns in BERT sentence representations.
We find that even in higher layers, the model mostly picks up on a
variegated bunch of low-level features, many related to sentence
complexity, that presumably arise from its specific pre-training
objectives.
%0 Conference Paper
%1 nikolaev23:_universe
%A Nikolaev, Dmitry
%A Padó, Sebastian
%B Proceedings of IWCS
%C Nancy, France
%D 2023
%K conference myown
%T The Universe of Utterances According to BERT
%U https://aclanthology.org/2023.iwcs-1.12.pdf
%X It has been argued that BERT ``rediscovers the traditional NLP
pipeline'', with lower layers extracting morphosyntactic features and
higher layers creating holistic sentence-level representations.
In this paper, we critically examine this assumption through a
principle-component-guided analysis, extracing sets of inputs that
correspond to specific activation patterns in BERT sentence representations.
We find that even in higher layers, the model mostly picks up on a
variegated bunch of low-level features, many related to sentence
complexity, that presumably arise from its specific pre-training
objectives.
@inproceedings{nikolaev23:_universe,
abstract = {It has been argued that BERT ``rediscovers the traditional NLP
pipeline'', with lower layers extracting morphosyntactic features and
higher layers creating holistic sentence-level representations.
In this paper, we critically examine this assumption through a
principle-component-guided analysis, extracing sets of inputs that
correspond to specific activation patterns in BERT sentence representations.
We find that even in higher layers, the model mostly picks up on a
variegated bunch of low-level features, many related to sentence
complexity, that presumably arise from its specific pre-training
objectives.},
added-at = {2023-04-26T18:18:19.000+0200},
address = {Nancy, France},
author = {Nikolaev, Dmitry and Padó, Sebastian},
biburl = {https://puma.ub.uni-stuttgart.de/bibtex/2995882df8792875cb3a67bbea378a90a/sp},
booktitle = {Proceedings of IWCS},
interhash = {9dabc394be5ffe8d8b9818192791f99a},
intrahash = {995882df8792875cb3a67bbea378a90a},
keywords = {conference myown},
timestamp = {2024-02-22T12:31:40.000+0100},
title = {The Universe of Utterances According to {BERT}},
url = {https://aclanthology.org/2023.iwcs-1.12.pdf},
year = 2023
}