Computational approaches to semantic change

Computational approaches to semantic change
Author: Nina Tahmasebi
Publisher: Language Science Press
Total Pages: 396
Release: 2021-08-30
Genre: Language Arts & Disciplines
ISBN: 3961103127

Semantic change — how the meanings of words change over time — has preoccupied scholars since well before modern linguistics emerged in the late 19th and early 20th century, ushering in a new methodological turn in the study of language change. Compared to changes in sound and grammar, semantic change is the least understood. Ever since, the study of semantic change has progressed steadily, accumulating a vast store of knowledge for over a century, encompassing many languages and language families. Historical linguists also early on realized the potential of computers as research tools, with papers at the very first international conferences in computational linguistics in the 1960s. Such computational studies still tended to be small-scale, method-oriented, and qualitative. However, recent years have witnessed a sea-change in this regard. Big-data empirical quantitative investigations are now coming to the forefront, enabled by enormous advances in storage capability and processing power. Diachronic corpora have grown beyond imagination, defying exploration by traditional manual qualitative methods, and language technology has become increasingly data-driven and semantics-oriented. These developments present a golden opportunity for the empirical study of semantic change over both long and short time spans. A major challenge presently is to integrate the hard-earned knowledge and expertise of traditional historical linguistics with cutting-edge methodology explored primarily in computational linguistics. The idea for the present volume came out of a concrete response to this challenge. The 1st International Workshop on Computational Approaches to Historical Language Change (LChange'19), at ACL 2019, brought together scholars from both fields. This volume offers a survey of this exciting new direction in the study of semantic change, a discussion of the many remaining challenges that we face in pursuing it, and considerably updated and extended versions of a selection of the contributions to the LChange'19 workshop, addressing both more theoretical problems — e.g., discovery of "laws of semantic change" — and practical applications, such as information retrieval in longitudinal text archives.

Computational approaches to semantic change

Computational approaches to semantic change
Author: Nina Tahmasebi
Publisher: BoD – Books on Demand
Total Pages: 397
Release: 2021-08-10
Genre: Language Arts & Disciplines
ISBN: 398554008X

Semantic change — how the meanings of words change over time — has preoccupied scholars since well before modern linguistics emerged in the late 19th and early 20th century, ushering in a new methodological turn in the study of language change. Compared to changes in sound and grammar, semantic change is the least understood. Ever since, the study of semantic change has progressed steadily, accumulating a vast store of knowledge for over a century, encompassing many languages and language families. Historical linguists also early on realized the potential of computers as research tools, with papers at the very first international conferences in computational linguistics in the 1960s. Such computational studies still tended to be small-scale, method-oriented, and qualitative. However, recent years have witnessed a sea-change in this regard. Big-data empirical quantitative investigations are now coming to the forefront, enabled by enormous advances in storage capability and processing power. Diachronic corpora have grown beyond imagination, defying exploration by traditional manual qualitative methods, and language technology has become increasingly data-driven and semantics-oriented. These developments present a golden opportunity for the empirical study of semantic change over both long and short time spans. A major challenge presently is to integrate the hard-earned knowledge and expertise of traditional historical linguistics with cutting-edge methodology explored primarily in computational linguistics. The idea for the present volume came out of a concrete response to this challenge. The 1st International Workshop on Computational Approaches to Historical Language Change (LChange'19), at ACL 2019, brought together scholars from both fields. This volume offers a survey of this exciting new direction in the study of semantic change, a discussion of the many remaining challenges that we face in pursuing it, and considerably updated and extended versions of a selection of the contributions to the LChange'19 workshop, addressing both more theoretical problems — e.g., discovery of "laws of semantic change" — and practical applications, such as information retrieval in longitudinal text archives.

Word Embeddings: Reliability & Semantic Change

Word Embeddings: Reliability & Semantic Change
Author: J. Hellrich
Publisher: IOS Press
Total Pages: 190
Release: 2019-08-08
Genre: Computers
ISBN: 1614999953

Word embeddings are a form of distributional semantics increasingly popular for investigating lexical semantic change. However, typical training algorithms are probabilistic, limiting their reliability and the reproducibility of studies. Johannes Hellrich investigated this problem both empirically and theoretically and found some variants of SVD-based algorithms to be unaffected. Furthermore, he created the JeSemE website to make word embedding based diachronic research more accessible. It provides information on changes in word denotation and emotional connotation in five diachronic corpora. Finally, the author conducted two case studies on the applicability of these methods by investigating the historical understanding of electricity as well as words connected to Romanticism. They showed the high potential of distributional semantics for further applications in the digital humanities.

Computational Methods for Corpus Annotation and Analysis

Computational Methods for Corpus Annotation and Analysis
Author: Xiaofei Lu
Publisher: Springer
Total Pages: 192
Release: 2014-07-08
Genre: Language Arts & Disciplines
ISBN: 9401786453

In the past few decades the use of increasingly large text corpora has grown rapidly in language and linguistics research. This was enabled by remarkable strides in natural language processing (NLP) technology, technology that enables computers to automatically and efficiently process, annotate and analyze large amounts of spoken and written text in linguistically and/or pragmatically meaningful ways. It has become more desirable than ever before for language and linguistics researchers who use corpora in their research to gain an adequate understanding of the relevant NLP technology to take full advantage of its capabilities. This volume provides language and linguistics researchers with an accessible introduction to the state-of-the-art NLP technology that facilitates automatic annotation and analysis of large text corpora at both shallow and deep linguistic levels. The book covers a wide range of computational tools for lexical, syntactic, semantic, pragmatic and discourse analysis, together with detailed instructions on how to obtain, install and use each tool in different operating systems and platforms. The book illustrates how NLP technology has been applied in recent corpus-based language studies and suggests effective ways to better integrate such technology in future corpus linguistics research. This book provides language and linguistics researchers with a valuable reference for corpus annotation and analysis.

Polysemy

Polysemy
Author: Yael Ravin
Publisher: OUP Oxford
Total Pages: 242
Release: 2000-06-15
Genre: Language Arts & Disciplines
ISBN: 019158469X

This volume of newly commissioned essays examines current theoretical and computational work on polysemy, the term used in semantic analysis to describe words with more than one meaning or function, sometimes perhaps related (as in plain) and sometimes perhaps not (as in bank). Such words present few difficulties in everyday language, but pose central problems for linguists and lexicographers, especially for those involved in lexical semantics and in computational modelling. The contributors to this book–leading researchers in theoretical and computational linguistics–consider the implications of these problems for grammatical theory and how they may be addressed by computational means. The theoretical essays in the book examine polysemy as an aspect of a broader theory of word meaning. Three theoretical approaches are presented: the Classical (or Aristotelian), the Prototypical, and the Relational. Their authors describe the nature of polysemy, the criteria for detecting it, and its manifestations across languages. They examine the issues arising from the regularity of polysemy and the theoretical principles proposed to account for the interaction of lexical meaning with the semantics and syntax of the context in which it occurs. Finally they consider the formal representations of meaning in the lexicon, and their implications for dictionary construction. The computational essays are concerned with the challenge of polysemy to automatic sense disambiguation–how intended meaning for a word occurrence can be identified. The approaches presented include the exploitation of lexical information in machine-readable dictionaries, machine learning based on patterns of word co-occurrence, and hybrid approaches that combine the two. As a whole, the volume shows how on the one hand theoretical work provides the motivation and may suggest the basis for computational algorithms, while on the other computational results may validate, or reveal problems in, the principles set forth by theories.

Methods in Latin Computational Linguistics

Methods in Latin Computational Linguistics
Author: Barbara McGillivray
Publisher: BRILL
Total Pages: 246
Release: 2013-11-29
Genre: Language Arts & Disciplines
ISBN: 9004260129

In Methods in Latin Computational Linguistics, Barbara McGillivray presents some of the most significant methodological foundations of the emerging field of Latin Computational Linguistics. The reader will find an overview of the computational resources and tools available for Latin and three corpus case studies covering morpho-syntactic and lexical-semantic aspects of Latin verb valency, as well as quantitative diachronic explorations of the argument realization of Latin prefixed verbs. The computational models and the multivariate data analysis techniques employed are explained with a detailed but accessible language. Barbara McGillivray convincingly shows the challenges and opportunities of combining computational methods and historical language data, and contributes to driving the technological change that is affecting Historical Linguistics and the Humanities.

Foundations of Computational Linguistics

Foundations of Computational Linguistics
Author: Roland Hausser
Publisher: Springer Science & Business Media
Total Pages: 541
Release: 2013-03-09
Genre: Computers
ISBN: 3662039206

The central task of future-oriented computational linguistics is the development of cognitive machines which humans can freely speak to in their natural language. This will involve the development of a functional theory of language, an objective method of verification, and a wide range of practical applications. Natural communication requires not only verbal processing, but also non-verbal perception and action. Therefore, the content of this book is organized as a theory of language for the construction of talking robots with a focus on the mechanics of natural language communication in both the listener and the speaker.