Semantic noise, the effect ensuing from the denotative and thus functional variability exhibited by different terms in different contexts, is a common concern in natural language processing (NLP). While unarguably problematic in specific applications (e.g., certain translation tasks), the main argument of this paper is that failing to observe this linguistic matter of fact as a generative effect rather than as an obstacle, leads to actual obstacles in instances where language model outputs are presented as neutral. Given that a common and long-standing challenge in NLP is the interpretation of ambiguous–i.e., semantically noisy–cases, this article focuses on an exemplar ambiguity-resolution task in NLP: the problem of anaphora in Winograd schemas. The main question considered is: to what extent is the standard approach to disambiguation in NLP subject to a stagnant “image of language”? And, can a transdisciplinary, dynamic approach combining linguistics and philosophy elucidate new perspectives on these possible conceptual shortcomings? In order to answer these questions we explore the term and concept of noise, particularly in its presentation as semantic noise. Owing to its definitional plurality, and sometimes even desirable unspecificity, the term noise is thus used as proof of concept for semantic generativity being an inherent characteristic in linguistic representation, and its concept is used to interrogate assumptions admitted in the resolution of Winograd schemas. The argument is speculative and theoretical in method, and the result is an analysis which provides an account of the fundamentally dialogical and necessarily open-ended effects of semantic noise in natural language.
CITATION STYLE
de Jager, S. (2023). Semantic Noise and Conceptual Stagnation in Natural Language Processing. Angelaki - Journal of the Theoretical Humanities, 28(3), 111–132. https://doi.org/10.1080/0969725X.2023.2216555
Mendeley helps you to discover research relevant for your work.