TY - JOUR
T1 - On the von Neumann entropy of language networks
T2 - Applications to cross-linguistic comparisons
AU - Vera, Javier
AU - Fuentealba, Diego
AU - Lopez, Mario
AU - Ponce, Hector
AU - Zariquiey, Roberto
N1 - Publisher Copyright:
Copyright © 2022 EPLA.
PY - 2021/12
Y1 - 2021/12
N2 - Words are not isolated entities within a language. In this paper, we measure the number of choices transmitted in natural language by means of the von Neumann entropy of language networks. This quantity, introduced in Quantum Information accounts, provides a detailed characterization of network complexities. The simulations are based on a large parallel corpus of 362 languages across 55 linguistic families (focusing on the sub-sample of 85 languages from the Americas). With this, we constructed language networks as a simple way to describe word connectivity patterns for each language. We studied several aspects of the von Neumann entropy of language networks. First, we discovered large groups of languages with low average degree and high von Neumann entropy. The results suggested also that large von Neumann entropy is associated with word entropy (as a proxy for morphological complexity), and is inversely related to degree regularity. This means that there are pressures at play that keep a balance between word morphological complexity and patterns of connections between words. We suggested also a strong influence of functional words on low von Neumann entropy languages. Our approach is thus a simple network-based contribution to establish cross-linguistic language comparisons from textual data.
AB - Words are not isolated entities within a language. In this paper, we measure the number of choices transmitted in natural language by means of the von Neumann entropy of language networks. This quantity, introduced in Quantum Information accounts, provides a detailed characterization of network complexities. The simulations are based on a large parallel corpus of 362 languages across 55 linguistic families (focusing on the sub-sample of 85 languages from the Americas). With this, we constructed language networks as a simple way to describe word connectivity patterns for each language. We studied several aspects of the von Neumann entropy of language networks. First, we discovered large groups of languages with low average degree and high von Neumann entropy. The results suggested also that large von Neumann entropy is associated with word entropy (as a proxy for morphological complexity), and is inversely related to degree regularity. This means that there are pressures at play that keep a balance between word morphological complexity and patterns of connections between words. We suggested also a strong influence of functional words on low von Neumann entropy languages. Our approach is thus a simple network-based contribution to establish cross-linguistic language comparisons from textual data.
UR - http://www.scopus.com/inward/record.url?scp=85127385120&partnerID=8YFLogxK
U2 - 10.1209/0295-5075/ac39ee
DO - 10.1209/0295-5075/ac39ee
M3 - Article
AN - SCOPUS:85127385120
SN - 0295-5075
VL - 136
JO - EPL
JF - EPL
IS - 6
M1 - 68003
ER -