Probing Tasks Under Pressure

Abstract

Probing tasks are frequently used to evaluate whether the representations of Neural Language Models (NLMs) encode linguistic information. However, it is still questioned if probing classification tasks really enable such investigation or they simply hint for surface patterns in the data. We present a method to investigate this question by comparing the accuracies of a set of probing tasks on gold and automatically generated control datasets. Our results suggest that probing tasks can be used as reliable diagnostic methods to investigate the linguistic information encoded in NLMs representations.

Publication
In Proceedings of the Eighth Italian Conference on Computational Linguistics (CLiC-it 2021)
Alessio Miaschi
Alessio Miaschi
PhD Candidate in Computer Science