Next Article in Journal
Schrödinger’s Ballot: Quantum Information and the Violation of Arrow’s Impossibility Theorem
Next Article in Special Issue
A Method for Estimating the Entropy of Time Series Using Artificial Neural Networks
Previous Article in Journal
Passive Tracking of Multiple Underwater Targets in Incomplete Detection and Clutter Environment
Previous Article in Special Issue
Asymptotic Information-Theoretic Detection of Dynamical Organization in Complex Systems
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Using the Relative Entropy of Linguistic Complexity to Assess L2 Language Proficiency Development

1
Department of Linguistics, University of Tübingen, 72074 Tübingen, Germany
2
Institute of Natural Language Processing, University of Stuttgart, 70569 Stuttgart, Germany
*
Authors to whom correspondence should be addressed.
Submission received: 10 July 2021 / Revised: 12 August 2021 / Accepted: 17 August 2021 / Published: 20 August 2021
(This article belongs to the Special Issue Statistical Methods for Complex Systems)

Abstract

:
This study applies relative entropy in naturalistic large-scale corpus to calculate the difference among L2 (second language) learners at different levels. We chose lemma, token, POS-trigram, conjunction to represent lexicon and grammar to detect the patterns of language proficiency development among different L2 groups using relative entropy. The results show that information distribution discrimination regarding lexical and grammatical differences continues to increase from L2 learners at a lower level to those at a higher level. This result is consistent with the assumption that in the course of second language acquisition, L2 learners develop towards a more complex and diverse use of language. Meanwhile, this study uses the statistics method of time series to process the data on L2 differences yielded by traditional frequency-based methods processing the same L2 corpus to compare with the results of relative entropy. However, the results from the traditional methods rarely show regularity. As compared to the algorithms in traditional approaches, relative entropy performs much better in detecting L2 proficiency development. In this sense, we have developed an effective and practical algorithm for stably detecting and predicting the developments in L2 learners’ language proficiency.

1. Introduction

Measuring learners’ second language (L2) proficiency is an important issue not only as regards practical language teaching, but also with respect to research on L2 acquisition. To this end, in the following, both qualitative and quantitative methods have been employed, such as face-to-face interviews, standardized tests, and linguistic feature analysis and modeling. In recent years, a popular means of quantitatively measuring L2 proficiency has been to look at the linguistic features of the learners’ L2 production from the point of view of complexity [1,2], which is one aspect of the Complexity, Accuracy, and Fluency (CAF) framework for learner language analysis [3,4]. Complexity measures have been employed to gauge L2 proficiency and its development at multiple levels of linguistic representation such as lexis, morphosyntax, discourse, and psycholinguistics [5,6].
Linguistic complexity functions as a basic descriptor of L2 performance and as indicative of L2 proficiency and development in such research [1]. Hundreds of complexity measures have been developed and used in previous investigations [7,8,9]. In general terms, the calculation of these measures boils down to counting the number of linguistic components (e.g., words, dependent clauses, complex nominals, etc.) and the number and types of connections between these components [10]. Although research making use of these measures has helped to examine a number of important issues concerning L1 (the first language) and L2 acquisition, such as proficiency assessment, language development, and language teaching/learning [3,11,12], the methods for enumerating linguistic components and calculating their relative ratios make use of the available information. These studies contributed highly useful insights into the nature of L2 lexical/syntactic changes in general and of the development of L2 learners over time. Meanwhile, they also leave much room for further exploration.
There are several potential limitations regarding these measures and methods. First, most of the studies that employed them were based on the summation or ratio algorithm of frequency. Despite the variety of measurement units that have become available, we have to admit that in practice the summation or ratio algorithm of frequency for the operation of these units has not been upgraded: still based on simple addition, subtraction and ratios of frequencies. Although frequency is a very important and useful measure in quantitative research, it is merely the sum of all the reasons why a word could occur often in a corpus. Frequency as a measure does not consider the context. The lexical and syntactic complexity measures used in the past studies are actually based on this algorithm of frequency summation or ratio. However, the development of computational algorithms allows us to develop methods that make more precise measurements. In order to reduce the limit of frequency, many studies attempted to use a diversity of methods to explore linguistic changes. For instance, the n-gram algorithm can be used to calculate the bi-/tri-/four-gram transitional probability, which considers the larger context or the probability of the context.
Syntactic complexity, for example, employs measures such as dependent clause/T-unit (a T-unit is often a sentence), dependent clauses per clause, and complex nominals per clause to compute the ratios based on the frequencies of the respective components. The underlying algorithm used in these measures is the summation of frequency or the ratio of frequencies. However, these frequencies might not be comparable. This is because units (measures/metrics) with different meanings or usages are not exactly the same in different L2 corpora even though they might be categorized under the same label. A direct summation or ratio that is simply based on their frequency would completely ignore their contextual meanings or usages. For instance, “subordination” in English can be employed in a diversity of subordinate conjunctions that have different meanings and usages. That is to say, each subordinate conjunction (subordination) should be measured according to frequency by using different weights rather than by simply using frequency. However, the algorithm employed in past studies has ignored the distinctions between different types of subordination in meanings and usages. This can be illustrated using the following example. L2 learners at a lower level use a “that-clause” 10 times, while L2 learners at a higher level use various “time, contrast, reason-effect” adverbial clauses 10 times. An algorithm based on the summation or ratio of frequencies tends to treat the two cases as the same with regard to their use of subordination. This may be problematic because there is a variety of different kinds of subordination in the two groups. Similarly, the summation or ratio algorithm based on frequency that was used in the past studies also ignored the weights of the various lexical and grammatical units that were used to measure complexity and accuracy. This problem can be solved with a new algorithm that considers the weights of different units that were previously grouped under the same category. An alternative that is simpler and more effective is an algorithm that directly measures exactly the same linguistic units in two different groups of L2 learners.
One of the assumptions in L2 developments research has been that learners use a more elaborate and diverse language as they progress in their development. Such progress could then be seen, for instance, in the use of longer sentences, more embedding, and a greater range of vocabulary and morphology [13,14]. More recently, this assumption has been challenged by an approach to linguistic complexity that, on the one hand, considers such complexity to be a multidimensional phenomenon [1,11,12] and, on the other hand, specifies the relation between complexity and proficiency more accurately [4]. The conflicting results from these studies could have been brought about by the use of different fine-grained measures. However, studies based on fine-grained measures of the development of L2 language proficiency still compared different units but without considering their different weights. In addition to this, the algorithm of summation (or ratio) of frequency in complexity was still applied in these fine-grained studies in order to assess L2 development.
The present study proposes a practical and effective method for measuring the development of language proficiency in L2 learners. This practical and effective method should be capable of overcoming the aforementioned difficulties. The algorithm employed is that of relative entropy, which is used to estimate the information distribution discrimination between two groups of probabilities for the same set of events. Distinguishability is one of the central information-theoretic concepts in science. The distinguishability of language proficiency among L2 learners at different levels can certainly be measured by the information-theoretic method. Relative entropy is an effective algorithm for carrying out this task and it has already been widely applied in various fields of natural science, in the social sciences, and in the humanities [15,16,17,18,19]. In order to illustrate the effectiveness and stability of this new algorithm, we use the traditional frequency-based method and the new algorithm to process the same L2 corpus and compare which algorithm is better.
The current study uses relative entropy to detect lexical and grammatical developments in L2 learners and it is based on a large-scale L2 learners’ corpus (EFCAMDAT2). We will address the following two questions:
(1)
How distinct are the differences in language proficiency between L2 learners at a lower level and L2 learners at a higher level as compared to the differences between intermediate-level L2 learners and higher-level learners from the perspective of information gain?
(2)
Does the algorithm of relative entropy have advantages over the frequency-based algorithms for lexical and syntactic complexity in detecting development patterns of L2 language proficiency?

2. Background

2.1. Linguistic Complexity and the Development of Language Proficiency in L2

Linguistic complexity has been extensively studied as an indicator of linguistic performance, development, and proficiency in L2 learners. This means we first need to discuss what linguistic complexity is. A complexity metric quantifies how difficultly a linguistic expression is perceived. However, linguistic complexity has been evaluated using different measures in theoretical linguistics, in applied linguistics and in language cognition. In formal theoretical linguistics, the question of complexity differences among different languages does not arise because the complexity of individual languages is seen as determined by invariant universal mechanisms [20]. In contrast, cross-linguistic complexity differences have long been at the heart of functionalist and usage-based linguistics, particularly in relation to language typology [21,22,23,24,25]. Using information-theoretical measures to detect historical changes in language is also useful [26,27].
The research concepts and methods concerning linguistic complexity used in L2 have been reviewed by [28]. Complexity in L2 research has been measured either subjectively through rating scales or, more commonly, through the use of objective quantitative measures [29]. Crucially, L2 research also maintains the distinction between relative complexity (or difficulty) and absolute (or structural) complexity that is used in other disciplines [30,31]. In L2 research, absolute complexity has been associated with the length or size of linguistic units (words, phrases, clauses, sentences, T-units), with the range, variety, richness or diversity of items in a linguistic system or domain, and with properties that refer to the composition and hierarchic organization of linguistic units (e.g., embedding, subordination) [28]. The methods concerning L2 absolute complexity can be summarized using two terms, measures and algorithms, which were all discussed in the Introduction. Measures are linguistic units. In the last 20 years, the studies on lexical/syntactic complexity and development in L2 have played an important role in L2 research, gaining insights of L2 lexical/syntactic acquisition. However, as mentioned in the Introduction section, these advances in the last 20 years are basically finding new and diversified linguistic units to use as measurement in L2. The algorithm is the frequency of these units or the summation (ratio) of two types of frequencies concerning two types of linguistic units. Although new and diversified linguistic units are essential in assessing L2 lexical/syntactic complexity and development, the underlining algorithms also need to be upgraded.
The section of introduction pointed out the main problems with these methods: the algorithm is a little simple and it ignores the meanings and usage of the units. Recent studies have noted the weakness of the measures (linguistic units). In recent studies, it has been proposed to use more fine-grained measures to address such concerns [29]. Yet this weakness has not been completely overcome.
L2 complexity studies can be applied in order to compare the differences (mostly concerning language proficiency) between L2 learner groups or between L2 learners and L1 learners. These L2 complexity methods usually calculate a score using some given units within a group and they then compare the scores of this group to a number of other groups. These L2 complexity methods assess the differences between these groups concerning language proficiency, learning characters, language difficulty, and language development using the scores so obtained. The reasons for this are that the units measured by different studies are not really the same, and the ‘static’ comparison so discerned is not stable. Additionally, the number of linguistic units/measures used for the assessment is too large (see Appendix C). When many such units/measures are taken to evaluate complexity, this can cause some problems. For example, are these units symmetrical? Which unit is more important? Do too many different units result in a great increase in computational complexity and cost? The studies using these measures/approaches did not explain these difficulties or are simply unaware of these problems. In short, L2 complexity sets out to discern various differences through comparing the scores of given groups.
Some studies of L2 development have directly discerned the difference between two L2 learner groups. However, these studies, which compared the development of language proficiency between L2 learners or examined differences in language proficiency between L2 learners, focused on “static” data for each L2 level. Because we are concerned with the development of language proficiency, we will concentrate on the difference between the two L2 learner groups directly in a practical manner. Discerning the difference in language proficiency between two L2 learner groups can thus be treated as a “practical and effective” method.
The other possible area of improvement with respect to previous research on this issue is that many different studies have reported conflicting results concerning the growth of lexical and syntactic complexity in L2 learners. For instance, as the comparison of [1,32] shows, findings reported in different studies can be conflicting even though they were carried out in relatively similar contexts and on learners with a relatively similar proficiency level. However, the challenges that such discrepancies pose lead to more substantive questions. If complexity is achieved by different means at different points (i.e., different proficiency levels) in the developmental paths of L2 learners, then different areas of complexity may be relevant at one given proficiency level but irrelevant or at least less predictive of growth at a different given proficiency level. When these conflicting results concerning developmental patterns in L2 are brought together, the researcher is likely to feel confused and disappointed. The root cause of these conflicting results may be the simplicity of the algorithms, and we have already discussed their drawbacks and possible problems.
A new algorithm could solve these problems by considering the weights of different units that had been previously placed under the same category. However, it is potentially very difficult to assign different weights to fine-grained units. An alternative method that is simpler but operational is to directly measure exactly the same linguistic units in two different groups of L2 learners. Most studies on L2 complexity focused on lexical and syntactic measures, but they overestimated the interface between lexicon and syntax. Few recent studies have examined this area [33,34,35]. This study will include measures at the interface between morphology and syntax. In view of the direct detection of differences and the probability strength, relative entropy should be an operational algorithm for computing the discriminative information distribution through detecting the same set of units in two different groups.

2.2. Relative Entropy

In information theory, entropy [36] quantifies the amount of uncertainty involved in the value of a random variable or the outcome of a random process. “Relative entropy” or the Kullback–Leibler Divergence (KLD, also called “Bayesian surprise”) [37] is derived from entropy. It refers to the number of additional “bits” needed when a non-optimal encoding is used. Relative entropy is actually the expected discrimination information, and it is used to measure the loss and gain of information. Formally, given two probability distributions p(x) and q(x) over a discrete random variable X, the relative entropy given by D(p||q) is defined as follows:
D ( p q ) = x X p ( x ) l o g p ( x ) q ( x ) = x X p ( x ) ( l o g p ( x ) l o g q ( x ) )
D(p||q) (KLD) can also be understood as a measure of the information gained by revising one’s beliefs from the prior probability distribution q to the posterior probability distribution p. The KLD is therefore used to measure cognitive cost. P typically represents the “true” distribution of data or observations, while q typically represents ideal data or theoretical data. The order of p and q in (1) cannot be reversed. However, the KLD is applied to the same underlying set of events. Relative entropy has been widely applied to multiple subjects in science, in social studies, and in the humanities [15,16,17,18,38]. A recent strand of data-driven approaches in the analysis of diachronic change applies KLD measures. For instance, [39] applied the KLD and discovered that there is a correlation between the major intellectual periods in Darwin’s career as identified both by scholarship and his own self-commentary in his reading notes. The KLD was used to analyze the stylistic development of literary works over time in order to discern the patterns of stylistic influence in the evolution of literature [27].
We will illustrate how the KLD is applied to diachronic changes. For instance, the KLD method of a sliding window (e.g., 10 years) can be used to detect changes in units of language. Specifically, the KLD is used to compare preceding (pre-period) and subsequent (post-period) years in the sliding window. We can thus obtain Equation (2) based on Equation (1):
D ( p o s t p r e ) = i p ( u n i t i | p o s t ) ( l o g 2 p ( u n i t i | p o s t ) l o g 2 p ( u n i t i | p r e ) ) i p ( u n i t i | p o s t ) l o g 2 p ( u n i t i | p o s t ) i p ( u n i t i | p o s t ) l o g 2 p ( u n i t i | p r e )
Here, for instance, if the “1800s” are a “pre” period, the “1810s” are a “post” period. The following example is used to illustrate how to compute the KLD. Table 1 is the frequency of the lemma (A lexeme is the set of all forms that have the same meaning, while lemma refers to the particular form that is chosen by convention to represent the lexeme. In English, for example, run, runs, and running are forms of the same lexeme, but run is the lemma.) “take” from 1810 to 1840, which is extracted from the COHA (Corpus of Historical American English, https://www.english-corpora.org/coha/, accessed date: 18 August 2021). When we need to calculate the KLD of “take”, Equation (2) can be specified as follows if “post” and “pre” are the 1820s and the 1810s.
D(1820s||1810s) = ∑p(“take”|1820s)*(log2p(“take”|1820s) − log2p(“take”|1810s))
= ∑p(“take”|1820s)*log2p(“take”|1820s) − ∑p(“take”|1820s)*log2p(“take”|1810s)
The lemma “take” has the five variations, as shown in Table 1. We therefore can obtain the probability for each variation of this lemma in the 1820s:
  • p(“take”|1820s) = 556/1435 = 0.3875,
  • p(“took”|1820s) = 327/1435 = 0.2279,
  • p(“taken”|1820s) = 345/1435 = 0.2404,
  • p(“taking”|1820s) = 144/1435 = 0.1003,
  • p(“takes”|1820s) = 63/1435 = 0.0439.
p(“take”|1820s)*log2p(“take”|1820s)
= sum(0.3875*log2(0.3875) + 0.2279*log2(0.2279) + 0.2404*log2(0.2404) + 0.1003*log2(0.1003) + 0.0439*log2(0.0439)) = −2.041431
Similarly, we can obtain the probability for each variation in the 1810s:
  • p(“take”|1810s) = 728/1415 = 0.5145,
  • p(“took”|1810s) = 168/1415 = 0.1187,
  • p(“taken”|1810s) = 228/1415 = 0.1611,
  • p(“taking”|1810s) = 133/1415 = 0.094,
  • p(“took”|1810s) = 158/1415 = 0.1117.
p(“take”|1820s) * log2p(“take”|1810s) = −2.1864
With “take”, D(1820s||1810s) is 0.145 bit. This bit value indicates that the lemma ‘take’ needs an additional 0.145 bit in the 1820s to encode information as compared to the 1810s. This additional bit also suggests that 0.14506-bit more cost was elicited in language users in using this lemma in the 1820s than in the 1810s. The same method is adopted to calculate the KLD of “take” in the 1830s and 1840s, the KLD of the 1830s and 1820s as well as that between the 1840s and 1830s. Therefore, the KLD for the lemma “take” from 1810 to 1840 has the following values: 0.145, 0.0039, 0.0102. These values can be collected to examine how the relative entropy of “take” has changed.
When we talk about distribution information discrimination, the KLD is closely associated with the loss/gain of information. However, when emphasizing language users’ or readers’ cognitive experience, the KLD is related to cognitive cost. Relative entropy has been directly used or adopted to detect cognitive difference or cognitive cost. For instance, tasks that correspond to the encoding of large amounts of information (relative to a model) also have correspondingly higher cognitive costs [40]. The authors of [30] developed a formulation which connects cognitive cost to information cost. However, such information is not available to us because we only have access to the corpus of L2 learners.
Other information-theoretical metrics have been used to investigate L2 acquisition [35,41]. Previously, information-theoretical measures have been used to evaluate lexical/syntactic complexity [42,43]. For instance, the Kolmogorov complexity algorithm has been used to compute the complexity degree of each L2 learner group [42]. It turned out that more advanced learners use considerably more complex texts than beginner learners. However, Kolmogorov complexity can be used to assess text complexity or utterance complexity rather than grammatical complexity. For instance, the same author may produce different texts with greater or lesser degrees of complexity. Additionally, although the method belongs to information-theoretical metrics, this approach is a “static” approach, and it is hardly capable of capturing the precise differences between L2 learner groups. The authors of [35] employed the scores of the algorithm Pointwise Mutual Information (PMI) to assess collocational complexity of phraseology. Phraseology has been ignored in previous studies of L2 complexity. Although PMI is a measure in information theory that is used to detect associations between two sets of events, it cannot be applied in other types of linguistic phenomena.

3. Materials and Methods

3.1. Material

The EFCAMDAT2 (The EF-Cambridge Open Language Database) currently contains over 83 million words from 1 million assignments written by 174,000 learners across a wide range of levels (The Common European Framework of Reference for Languages, CEFR, the six levels within the CEFR are A1, A2, B1, B2, C1, and C2.). EFCAMDAT2 is naturalistic large-scale L2 corpus compared with the other L2 corpora. This text corpus includes information about learner errors, parts of speech, and grammatical relationships. Researchers can search for language patterns using a range of criteria, including learner nationality and level [44].
The amount of texts from C2 learners (C2 is the highest level in CEFR) according is relatively small as compared to that of the learners at other levels. For this reason, C2 learners will not be considered in the present study. Following [45], the current study took advantage of the information on L2 learner proficiency (A1, A2, B1, B2, and C1) to perform a number of cross-sectional comparisons. The texts in the EFCAMDAT2 came from learners categorized according to 16 proficiency levels that correspond to the six levels of the CEFR: A1 “beginner” (levels 1–3), A2 “elementary” (levels 4–6), B1 “intermediate” (levels 7–9), B2 “upper intermediate” (levels 10–12), C1 “advanced” (levels 13–15), and C2 “proficiency” (level 16). Table 2 summarizes the composition of the 15 different sub-corpora of the written essays section of the EFCAMDAT2 used in our analysis.

3.2. Method

(a)
Relative entropy and the discrimination of information distribution.
The algorithm of the KLD can also be used to detect and assess the development of L2 learners’ language proficiency, because the investigation of different levels of L2 learners proceeds in a fashion that is almost the same as the examination of the diachronic changes between several historical sub-corpora that was carried out in previous studies. The differences in language proficiency between L2 learners can be seen as resulting from the differences in the respective amount of time spent learning the language. In this sense, the underlying philosophy behind both investigations is the same.
This can be applied to examine the differences between L2 learners at the different levels, as shown in the following Equation (3). Here, “Levelh” refers to L2 learners at a higher level, but “Levell” L2 learners at a lower level.
D ( L e v e l h L e v e l l ) = i p ( u n i t i | L e v e l h ) ( l o g 2 p ( u n i t i | L e v e l h ) l o g 2 p ( u n i t i | L e v e l l ) ) i p ( u n i t i | L e v e l h ) l o g 2 p ( u n i t i | L e v e l h ) i p ( u n i t i | L e v e l h ) l o g 2 p ( u n i t i | L e v e l l )
The KLD can quantify cognitive cost by detecting distinctions in information between the two levels or periods for the same set of events. In Equation (3), a high KLD indicates linguistic novelty in comparison with the lower level. The KLD is also associated with a change in the learner’s reactions upon encountering the unexpected. That means the “Levell” is the distribution of linguistic phenomena that learners have encountered at a lower level and “Levelh” is the new distribution that learners will encounter at a higher level. More importantly, the algorithm of relative entropy examines the information differences between the same linguistic units encoded by two groups of L2 learners. This avoids the problem that characterized previous studies, namely ignoring the weights of different units and simply placing them under the same category.
Relative entropy in Equation (2) measures the average amount of additional bits per linguistic unit needed to encode the same linguistic unit distributed according to “L2 learners at a higher level (Levelh)” by using an encoding optimized for “L2 learners at a lower level (Levell)”. When applied to the comparison of sub-corpora of the EFCAMDAT2, the KLD serves as a strong indication of the degree of difference between two sub-corpora (representing two groups of L2 learners) measured in bits as well as of the linguistic units that are primarily associated with a difference. That is to say, the difference in the KLD indicates that linguistic units need high amounts of additional bits for encoding. We can find the KLD as an indicator of change after sliding over different groups of L2 learners’ lines in the EFCAMDATA2 and by comparing adjacent L2 learners’ groups.
Overall, the KLD detects the discrimination of information distribution for the same set of linguistic units between L2 learners at different levels. In contrast, past methods that focused on the assessment of L2 language proficiency set out to measure one given group of L2 learners within a specific level. This is the biggest difference between the KLD and the past methods. In order to enable cross-verification of the results, we introduced another entropy divergence equation to do the same job. This is the Jensen–Shannon Divergence (JSD) [46]. More details can be seen in Appendix A. The two computational models can be used for the purposes of cross-verification.
(b)
Language units (measures).
When we speak of a linguistic “unit”, the term here is confined to lexicon and grammar in their typicality. The “lexicon” has two forms: the first refers to words of all sorts, that is the “token”; the second is the “lemma”, which refers to the canonical form or dictionary form of a set of words. The grammatical forms will be represented by subordinate conjunctions and the POS trigram.
The KLD of lexicon represents the distinction between two groups of L2 learners in information concerning lexical learning. “Token” and “lemma” can ensure that the KLD method really does detect the lexical distinction in the information distribution between the two groups of L2 learners. With respect to the lexicon, all strings containing digits or symbols were removed in order to preserve pure texts for each sub-corpus. After 119 common stop words, one was removed from the lexicon in each sub-corpus, and the lemmas were filtered again by choosing these lexica with a length greater than 2. After the selection, the KLD was calculated using the same vocabulary in each pair of two sub-corpora (two L2 learner groups). As discussed in the Introduction, the algorithm used in previous studies ignored the weights of different linguistic units by placing them under the same category. The relative entropy algorithm computed the same linguistic units for two L2 groups. We used two forms, “token” and “lemma”, to carry out this procedure and finally we obtained the results on KLD tokens and KLD lemmas.
The first grammatical form is the “POS trigram”. The POS trigram refers to a bundle consisting of three words marked by a part of speech. e.g., “wake your dreams” is a trigram and the trigram’s POS is Verb-Pronoun-Noun (abbreviated as “VB-PN-NN”). Being distinct from the bundle of trigram, the POS trigram is an entity consisting of POS labels. According to [47], in practice it is more common to use trigram models in the field of natural language processing because a trigram model depends on the previous two words rather than the previous word. Additionally, according to [48], 3-word lexical bundles have a much higher frequency than 4-word or 5-word lexical bundles. The POS-trigram can play a role in detecting phraseological and grammatical information in texts [34,38,49]. Because of this, we use the trigram POS to present grammar and detect its changes in the sub-corpora of learners at different levels. In this way, the PoS trigram can represent syntactic and phrasal knowledge or information. The raw texts were annotated by Treetagger (https://www.cis.uni-muenchen.de/~schmid/tools/TreeTagger/, accessed on 18 August 2021).
In addition to the POS trigram, we used the other measure to detect grammatical differences between L2 learners at different levels. This measure is that of “subordinate conjunction”. Clause subordination has a complex syntactic structure, constituting the potential linguistic complexity, and using it places a higher cognitive demand on language users. Because of these features, clause subordination has been widely investigated in studies of language [50,51,52]. The two features of clause subordination make it very interesting to use to examine people with a limited linguistic competence such as L2 learners [8,53]. The frequency of subordinate conjunctions varies greatly with regard to L2 learners at different levels or with different backgrounds [54,55]. We did not remove any subordinated conjunctions in view of their small number. Each pair of two sub-corpora has the same repository of subordinate conjunctions for calculating the KLD. Note that all of frequencies of the four measures are standard ones, and all the frequencies are based on one million words.
Overall, we have four measures under the categories of lexicon and grammar. The processing of lemmas, tokens, POS-trigrams, and subordinate conjunctions by the KLD will be exactly applied to the JSD, such that the two computational models will take on the same probability distribution as vectors when they are employed to calculate the differences between L2 learners at different levels.
(c)
Traditional approaches to lexical/syntactic complexity and stationary time series.
As discussed previously, a number of approaches and tools for lexical and syntactic complexity for L2 based on the algorithm of frequency have been developed and used widely. In order to compare our relative entropy approach with those traditional approaches, we will use the lexical and syntactic complexity approaches [8,9] to process the five sub-corpora of EFCAM2. In this way, we can obtain lexical and grammatical complexities for each corpus representing different L2 levels. However, this is hard to directly compare them with relative entropy. We can obtain the subtraction between two L2 levels and the data on subtraction concerning different L2 levels will be useful in comparing the predictive and explanatory power of the existing methods and indices and the relative entropy approach. Such comparison could provide direct evidence for the assessment of the two approaches. The L2 Syntactic Complexity Analyzer (L2SCA) can analyze the syntactic complexity of written English language for L2 users using 14 different measures. The Lexical Complexity Analyzer (LCA) is able to analyze the lexical complexity of written English language for L2, using 25 measures (see Appendix C).
As discussed above, relative entropy measures information discrimination, so the underlying principle of relative entropy can be applied to examine the differences between L2 learners at the different levels. However, the index of traditional syntactic and lexical complexities cannot be compared to relative entropy. We can apply the principle of difference between L2 learners to calculate the syntactic and lexical gap between two L2 learners. Put it simply, for example, Difference of Syntactic Complexity (A2:A1) = syntactic complexity of A2—syntactic complexity of A1. In this way, we can obtain the difference values among different L2 levels. We will examine these difference values to explore whether these values exhibit certain patterns or not. In other words, we will examine whether the data composed by the difference between L2 levels really follows some patterns. The previous studies simply observed the data on a given L2 level. However, such an examination takes place from a static perspective. A ‘practical and effective’ perspective on L2 development should consider their differences. Overall, given that the data composed by the difference between L2 levels really does follow some patterns, this means the L2 developmental patterns can be recognized explicitly. The approach of difference between two L2 groups is actually one method of time series, which will be specified in the following.
The traditional method for judging the trend in L2 developments has suffered from weaknesses. We can look at language proficiency development in L2 from the perspective of a time series. The investigation of different levels of L2 learners proceeds in a fashion that is almost the same as the examination of the diachronic changes that was carried out in previous studies [56,57]. The differences in language proficiency between L2 learners can be seen as resulting from the differences in the respective amount of time spent learning the language. In this sense, the underlying philosophy behind language proficiency between L2 learners concerns the time series. Different L2 levels actually constitute a time series. That is to say, each L2 level can be observed. In time order, for example, A1 is recorded after L2 learners spend one year, and L2 learners continue to use the other one year to reach A2, and the third year to B1 and so on. Although not all L2 learners use the same amount of time to make continuous progress, many language learning programs for different levels are based on the same time interval, such as language learning in formal schools based on a semester system. As another example, much of the data collection for language proficiency in longitudinal studies is based on the same time interval [2,7]. In this sense, the data on different L2 levels is basically a sequence taken at successive equally spaced points (e.g., one year) in time. Thus, it is a sequence of discrete-time data, that is, time series. According to the time series statistics, a direct comparison between the values is somewhat unreliable.
Although many studies concerning L2 proficiency development have treated differences between L2 levels as time differences, the statistics of time series has seldom been employed to process the data on L2 difference. For this reason, it is promising to use the time series perspective to examine the data yielded by traditional methods. Borrowing the principle of relative entropy and the stationary and differencing method in time series [58,59], we used the “first difference” to examine the data on lexical/syntactic complexity of L2 by traditional algorithms. Differencing is performed by subtracting the previous observation from the current observation. It can be simply described as “Difference(t) = observation(t) − observation(t − 1)”. Here, t refers to time (time order). When the data on Difference (t, t + 1, t + 2, …, t + n) does not show any increase or decrease, this indicates that the original data are stationary. This means that there is no pattern depending on the time at which the series are observed.

4. Results

4.1. The Results from the KLD

The results of the KLD of lexicon and grammar regarding five sub-corpora of the EFCAMDAT2 can be seen in the following three tables. They are displayed from different perspectives.
Table 3 presents the KLD results which examine relative entropy from the perspective of L2 at a lower level. Note that in Table 3, the upward arrow represents an upward trend in the data of this group, while the downward arrow represents a downward trend in the data in that group. The upward arrow and the downward arrow used in the other tables have the same function. As for a group of relative entropy data, we have used linear regression model to test whether the values in this group have an increase trend or not. “Coef” is the abbreviation of coefficient, representing the coefficient of this linear model, and p-value helps judge whether this linear model is significant or not.)
Table 3 shows that the KLD between a given lower level (e.g., A1, A2, B1) and a higher level becomes larger when L2 learners’ level increases. However, the KLD of token starting from A2 level and that of token and lemma starting from the B1 level is an exception. We used linear regression to examine whether the data in each group has a significant increase or decrease. When coef is positive, it means that these data have an increase and a negative coef indicates a decrease. When the p value is greater than 0.05, this suggests that such an increase or decrease may be not significant. The linear regression model is just used for a general observation of the direction of the data, so that it can be compared with the results yielded by traditional methods. For example, after one group data under “POS-trigram”, “0.37, 0.58, 0.78, 1”, is examined by a linear regression, we find that the coefficient is 0.21 and p-value is smaller than 0.001. This indicates that an increase in this group is significant, so there is an upward arrow. Because there are few numbers within a group data, the regression models are only auxiliary (the p-value is only a reference and not decisive). It is actually easy to tell whether the numbers within a group are increasing or decreasing. These are applied in Table 3, Table 4 and Table 5, as well as Table A1, Table A2 and Table A3 of Appendix B. From the perspective of cognitive cost, the KLD tends to increase over different groups of L2 learners if the starting point is A1. Specifically, the upward trend is 75% (9/12) of all measurement units. This could suggest that L2 learners at a lower level find it more difficult to process these language units than L2 learners at a higher level. Overall, the KLD between L2 learners at a higher level and L2 learners at a lower level is much larger than that between L2 learners at a lower level and L2 learners at an intermediate level. This indicates that L2 learners at a lower level need larger amounts of additional bits for encoding these linguistic units in comparison to L2 learners at a higher level. We will now look more closely at the KLD between L2 learners from the perspective of the higher levels.
As Table 4 shows, the KLD between a higher level and a lower level is much larger than the KLD between a lower level and an intermediate level. The KLD drops when the level moves from the lower to higher. This indicates that the gap between the level of the learners’ degree of language proficiency is proportional to their KLD. Specifically, the downward trend is 100% (12/12) of all measurement units. This could suggest that the difference of language proficiency between L2 learners at a higher level and L2 learners at an intermediate level is smaller than that between learners at the intermediate level and learners at the lower level. It also indicates that L2 learners at a higher level need less additional bits for encoding these linguistic units in comparison to L2 learners at a lower level. This also suggests that L2 learners at a higher level do not have a greater degree of cognitive cost in processing these linguistic units than L2 learners at a lower level. The following examines the KLD of L2 learners at adjacent levels.
Table 5 shows that when language levels become higher, their KLDs tend to become smaller. This could indicate that learners at intermediate levels have fewer difficulties in learning lexicon and grammar in the course of improving their language proficiency. By contrast, learners at the lower level may need to make more effort in improving their language proficiency. Specifically, the upward trend is 75% (3/4) of all measurement units. The data on the KLD in the three tables are consistent in supporting the thesis that the KLD between higher levels is much smaller than that between lower levels or that between a lower level and an intermediate level.
We represented some of the data from the above tables in Figure 1 so as to present the changes of the KLD between L2 learners at various levels more clearly. In each plot of Figure 1, x-axis represents L2 proficiency level and y-axis stands for relative entropy value (bits). Given that the curve shows an upward or downward trend, it indicates that relative entropy can detect patterns of L2 proficiency growth. However, when a curve looks irregular, it means that relative entropy fails. Alternatively, when a curve has a different trend compared with the other curves in a plot, it suggests that this curve also fails to capture patterns of L2 development. Additionally, a flat curve indicates no clear growth. For example, in the right top two plots, there is one exception where the token has a clearly opposing trend to the other measures. The four plots on the top clearly show that relative entropy increases in the majority of cases given we view from A1 to a higher level. By contrast, there is a clear decrease of relative entropy in the bottom four plots in Figure 1 given we view from C1 to a lower level. As discussed above, the top four plots are consistent with the bottom four plots, that is, relative entropy is effective to detect the linear change of language proficiency among L2 at different levels. Figure 1 and the data from the tables above actually answer the first research question posed at the end of the Introduction.
To save space, we have placed the results of the JSD in Appendix A. The results from the JSD are almost the same as those of the KLD. Overall, the JSD data are consistent with those of the KLD.

4.2. The Results from Syntactic and Lexical Complexity

We used L2SCA and LCA to process the five sub-corpora. Due to the limit of the length, the original data are shown in Appendix C. Table A4 of Appendix C shows that the original data on syntactic complexity (measured by different metrics) for each L2 level and the gap between a given lower level (i.e., A1) and a higher level increases when L2 learners’ level of linguistic proficiency rises. Table A5 of Appendix C shows that the original data on lexical complexity (measured by different metrics) for each L2 level and the gap between a given lower level (i.e., A1) and a higher level increases when L2 learners’ level of linguistic proficiency rises. As discussed in the Methods section, it is hard to compare relative entropy with the data on syntactic or lexical complexity for each level. According to the time series perspective, each L2 level is treated as a date. The gap between two L2 levels is the difference between the two dates, which has been discussed in the section of Methods. The other benefit is that the gap between two levels represented by syntactic or lexical complexity can be used to compare relative entropy because they share the same concept. In order to make a better observation, we represented in Figure 2 the data contained in Table A4 and Table A5 of Appendix C.
Figure 2 is composed of four panels. The top two panels represent the syntactic complexity gap between a given lower level (i.e., A1) and a higher level. All syntactic complexity metrics should be represented in one plot. However, the values in the three metrics (i.e., MCL, MLS, MLT) are somewhat larger than the others, so these three are plotted in a different graph to allow the data to be easily seen. The two panels at the bottom represent the syntactic complexity gap between the two adjacent levels (e.g., A1–A2, A2–B1). For this reason, we have represented syntactic complexity from the perspective of adjacent levels in the same way.
The top panel has deviant groups: CP/C, CP/T, DC/C, DC/T, T/S (5/14). It means that 36% of syntactic measures do not show patterns from the perspective of the low level (A1). The bottom panel has deviant groups: C/S, C/T. CN/C, CN/T, CP/C, CP/T, CT/T, DC/C, DC/T, ML/C, T/S (11/14). This indicates that almost 80% of syntactic measures do not show patterns from the perspective of the adjacent level. None of the metrics exhibit patterns from the perspective of the higher level (C1), as shown in Table A4 of Appendix C. In comparison with Table 4 and Table 5, in Table A2 and Table A3 of Appendix B (KLD and JSD) the measures of traditional syntactic complexity do not perform better than relative entropy because the syntactic measures of relative entropy form more regular patterns than those of traditional syntactic complexity.
Figure 3 is also composed of four panels. The top two panels represent the lexical complexity gap between a given lower level (i.e., A1) and a higher level. All lexical complexity metrics should be represented in one plot. However, the values in the three metrics (i.e., cttr, cvs1, cvv1, rttr, uber) are somewhat larger than the others, so the three are plotted in a different graph so as to allow easy observation. The bottom two panels represent the lexical complexity gap between the two adjacent levels (e.g., A1–A2, A2–B1). For this reason, we have represented the lexical complexity from the perspective of adjacent levels in the same way.
The top panel has deviant groups: uber, ld, logttr, ls1, ls2, lv, msttr, nv, ttr, vs1, vs1, vv1, vv2 (13/16). This means that 81% of metrics do not show patterns from the perspective of the low level (A1). The bottom panel has deviant groups: cttr, rttr, uber, ld, logttr, ls1, ls2, lv, msttr, nv, ttr, vs1, vs1, vv1, vv2(15/16). It indicates that 94% of metrics do not show patterns from the perspective of the adjacent levels. The perspective of higher level (C1) reveals that no metrics exhibit any patterns, as shown in Table A5 of Appendix C. In comparison with Table 4 and Table 5, in Table A2 and Table A3 of Appendix B (KLD and JSD) the measures of traditional lexical complexity do not perform better than relative entropy. Although the lexical measures of relative entropy do not perform better than the syntactic measures of relative entropy, these lexical measures still exhibit more regular patterns than the traditional measures.
More importantly, the data in Figure 2 and Figure 3 can be taken as a Difference type from the perspective of time series because the Difference data comes out by subtracting from the data in one state with the data from the other state (i.e., “Difference(t) = observation(t) − observation(t − 1)”, t = time), as discussed in Section 3.2 (c). It turns out that the Difference data in Figure 2 and Figure 3 does not show any increase or decrease patterns. It indicates that the original data yielded by the traditional lexical/syntactic measures (as seen in Table A4 and Table A5 of Appendix C) is stationary. In other words, it does not matter when the original data is observed because the original data should look much the same at any point in time. In this sense, the data on lexical/syntactic complexity measured by the traditional methods at different L2 levels does not exhibit an increase or decrease trend from the perspective of time series. Instead, relative entropy is helpful in showing more regular patterns of language proficiency among the different L2 levels. Specifically, 75%, 100%, and 75% of metrics show patterns using KLD respectively, shown in Table 3, Table 4 and Table 5; 75%, 100%, and 100% of metrics using JSD respectively show patterns. By contrast, as analyzed above, under the framework of traditional methods, 20% and none of syntactic metrics show patterns, and 19% and 6% of lexical metrics show patterns. Overall, the algorithm of relative entropy is much more effective and stable in detecting language proficiency development in L2 than the traditional frequency-based methods.

5. Discussion

The following section discusses our findings and compares them with the findings reported in previous related studies. As discussed above, the strength of relative entropy lies in detecting the differences. We explain how our findings are to be interpreted and how they relate to the findings in the previous studies from three points of view. This will address the second research question.

5.1. Conflicting Results from Different Studies

In Section 2.1, we mentioned the conflicting results reported by various studies of L2 development. For instance, [53] proposed that global complexity (for example, the mean length of T-unit) can most likely capture overall changes in complexity in any data. They further predicted that subordination, for example, as measured by the mean number of clauses per T-unit, is the preferred means of understanding linguistic complexity used at intermediate levels. At the most advanced levels of proficiency, subordination would cease to be predictive, and they instead expect that phrasal elaboration would be the main area of growth. However, when we closely consider the syntactic areas of complexity, the findings across different studies can look disconcerting. For instance, [1] found the opposite pattern: at the upper-intermediate level, their sample showed phrasal elaboration changes, but subordination remained unchanged. Both phrasal and clausal complexity grow together with regard to lower-proficiency L2 writers. Such conflicting results can be frustrating for researchers [2].
As discussed in the Introduction and Background sections, the main reason for these conflicting growth assessments is that all the measures used therein were based on traditional measures, which only compute the scores for each L2 learner group. These traditional measures are greatly influenced by the corpus size and the diversity of texts. When different types of static data are collected, comparison reveals that their results conflict. For instance, using again the example of ‘subordination’ that was mentioned in the Introduction, L2 learners at the A1 level use a “that-clause” 10 times, while L2 learners at the B2 level use various “time, contrast, reason-effect” adverbial clauses 10 times. The algorithm based on the summation or ratio of frequencies tends to treat the two cases as the same in their use of subordination. The result of this is that L2 learners at the two levels seem to remain stable with regard to their use of ‘subordination’. Suppose that L2 learners at the B2 level use various “time, contrast, reason-effect” adverbial clauses 9 (lower than 10) times. It seems on this approach that L2 learners at the B2 level use less complex subordinate structures than learners at the A1 level. However, the real situation is the opposite to that given in the results from the algorithm of summation or the ratio of frequencies. The same mistake occurs when the algorithm for measuring lexical or phrasal complexity is applied because it does not examine the exact linguistic units when using these measures and thus cannot compute the diversity of given linguistic units. Clearly, these conflicting results do not necessarily challenge the assumption that as L2 development progresses, learners use more elaborate, complex, and diverse language as regards lexicon and grammar.

5.2. The Developmental Patterns of Language Proficiency in L2 Learners

The findings reported in Section 4.2 allow us to interpret these conflicting reports to some degree. Clearly, linear growth or fluctuating development was reported by various studies. However, the summation of frequency or the ratio of two types of frequencies means they cannot really be representative of L2 complexity. Consequently, the change in the summation scores or ratios cannot really represent the L2 development with respect to language proficiency.
As a matter of fact, it is more practical to describe the difference between the two groups of L2 learners rather than applying an indirect method. It is possible that L2 learners at a higher level could produce more complex structures with respect to some linguistic phenomena. However, L2 learners at a higher level could provide some simpler structures with respect to these linguistic phenomena. For instance, while beginning and intermediate L2 learners may prefer complexity through coordination and subordination, phrasal complexity may be favored at more advanced levels of L2 proficiency [2]. This means that L2 advanced learners could increase the degree of complexity in subordination. This is confirmed by [60] who found that the learners produce longer, more complex phrases (more modifiers per noun phrase) at the end of a course of study, but not more or even fewer, subordinated clauses (i.e., fewer verb phrases; there is no significant change in number of subject relative clauses or ‘‘that’’ verb complements). These findings are not consistent with a linear growth or development in complexity. However, the complexity measured in the algorithm of frequency could have changed to detect the diversity in this usage of subordination in L2 learners at different levels, as was explained in the Introduction. It is highly likely that the information distribution concerning subordination might show greater differences between L2 learners at a lower level and L2 learners at a higher level than L2 learners of an intermediate level and a higher level.
As shown in Table A4 and Table A5 of Appendix C as well as in Figure 2 and Figure 3, the traditional syntactic and lexical approaches do not really seem to capture the developmental patterns of language proficiency in L2 learners so well. Different metrics even show conflicting trends. In many cases, these metrics demonstrate irregular patterns. The Different data in time series also demonstrates that the data yielded by the traditional lexical and syntactic measures is one stationary type. In another word, these data can show few regular patterns in L2 language proficiency development.
When these metrics exhibit irregular patterns or even conflicting patterns, the possibility that the underlying algorithm (frequency, or ratio) behind these traditional measures is not really stable or effective, which has seldom been considered previously. Our data, which was yielded by relative entropy, reveals that the information distribution discrimination regarding lexical and grammatical differences continues to increase from L2 learners at a lower level to those at a higher level. This finding is consistent with the assumption that as language proficiency develops, L2 learners will acquire a more elaborate, complex, and diverse command of the language. In addition to this, the developmental pattern we found is consistent with the differences in language proficiency among these L2 learners. This means that our method can detect the development of language proficiency among L2 learners at different levels more robustly and stably. We have thus addressed the second research question.
Both the methods we developed and the findings could provide us inspiration for second language acquisition, language assessment, and language teaching programs. Firstly, we have developed a practical and effective method to detect the language proficiency development of L2 learners. Secondly, our findings suggest that L2 proficiency development may be towards a more complex and diverse use of language, which is consistent with our intuition. In another word, L2 learners may acquire a second language linearly. Thirdly, our finding can greatly help gain insight in L2 acquisition and language teaching programming, which will be discussed in greater detail. For example, at certain stages, the focus of instruction will change. Vocabulary instruction may be less of a focus when teaching L2 learners at a higher level, and more of a focus when teaching complex sentence structures. For those L2 learners, the focus may be on textual structure rather than sentence structure. However, the linear development in L2 lexicon and syntax may not be influenced by the fact that the teaching emphasis could have been put elsewhere (i.e., textual structure). The reason for this is that textual focus still helps L2 learners to improve their lexical and syntactic capabilities. However, some past studies thought that their syntactic capabilities would stop increasing when L2 learners are at the intermediate level. Accordingly, the curriculum should make changes to emphasize teaching sentence structure again for L2 learners at the intermediate level. However, our finding may suggest that an overemphasis on curriculum changes to cater to those past research findings based on the traditional methods may not achieve the results they are looking for. Further, the method of relative entropy can not only be applied in detecting English L2 proficiency development, but also in detecting L2 proficiency development in the other languages. Moreover, relative entropy can be applied in lexical and grammatical change patterns in across-genre and across-language studies. Although the past studies have applied relative entropy in across-genre and across-language studies, they mostly used lemma as their measurement units. The linguistic units we have employed (i.e., POS-trigam, conjunction) can be applied in these fields help reveal more findings.

5.3. Consistency with the Other Measures

Relative entropy thus has unique strengths in comparison with the previous methods when it comes to measuring the development of language proficiency in L2 learners. The present study shows that the relative entropy of lexicon is largely consistent with the relative entropy of grammar. In past studies, results that were calculated by the frequency algorithm based on fine-grained linguistic units were found to be inconsistent with those yielded by non-fine-grained linguistic units. We also discussed another type of inconsistency in previous studies, namely that between syntactic measures and lexical ones. However, such inconsistencies have not been interpreted satisfactorily. These inconsistencies indicate that these measures and their underlying algorithm are not necessarily a good predictor of L2 learners’ language proficiency and its development.
Many studies using the traditional algorithm actually make a post-hoc analysis. That is to say, after these studies have obtained their results, these are then compared with the language proficiency levels of L2 learners in an attempt to find out something that explains the connection between their data and these levels of language proficiency. By contrast, the change of distinctions in relative entropy is consistent with the discrepancy between the levels representing L2 learners’ language proficiency. Relative entropy can thus be treated as a measure for predicting L2 learners’ language proficiency. More importantly, all measures (metrics) in the traditional methods (see Appendix C) must be interpreted individually. That is to say, the summation or ratio of frequency cannot provide a consistent interpretation for these measures. For instance, the three measures, C/T (# of clauses/# of T-unit), CT/T (# of complex T-units/# of T-units), DC/C (# of dependent clauses/# of clauses), have to be interpreted differently. When these units of measurement have different interpretations, various factors (gender, prompt-task, nationality, etc.) will be incorporated to explain these differences and conflicting results, and all seem to have plausibility.
The other advantage of relative entropy is that it can be used to interpret cognitive distinctions among L2 learners. Relative entropy can be treated as a simpler version of cognitive cost although the formulation for evaluating cognitive cost in tasks is a bit more complex than the KLD. Cognitive cost/effort can be interpreted as a subjective feeling of exhaustion experienced when performing a cognitive task and its associated task-avoidance. The cognitive costs associated with different classes of tasks are known to be subjectively demanding. It has been showed that this informational perspective can provide a unitary perspective on various experimental findings. Furthermore, we have discussed how information costs could be translated into cognitive effort (i.e., the subjective feeling associated with performing costly tasks). According to Table 3, Table 4 and Table 5, the cognitive cost on the part of L2 learners at a lower level and L2 learners at an intermediate level is much larger than that between L2 learners at an intermediate level and L2 learners at a higher level. This could reflect a decrease in cognitive cost that takes place when L2 learners acquire a greater language proficiency. It could be that L2 learners at a lower level make greater cognitive efforts in improving their language proficiency. This finding may indicate that a L2 learner at a lower level may need more instructions and practices to improve their language capabilities. By contrast, L2 learners at a higher level will feel it is easier to improve their language proficiency.

6. Conclusions

The current study used a novel ‘practical and effective’ algorithm derived from information-theoretic metrics to discern the development of L2 learners’ acquisition of language proficiency and it was based on a large-scale L2 writing corpus. It turned out that the relative entropy of lexical and grammatical differences continues to increase from L2 learners at a lower level to those at a higher level. This finding is consistent with the assumption that L2 learners acquire a more complex and diverse language as they progress. It is also consistent with the different levels of L2 language proficiency. Another key finding of this study is that relative entropy is a better predictor of language proficiency than the algorithm based on frequency summation or ratio. The Different data in time series demonstrates that the data yielded by the traditional lexical and syntactic measures can show few regular patterns in L2 language proficiency development. This means that the traditional approach seems not to yield robust, and stable results in detecting developmental patterns in L2. By contrast, the algorithm of relative entropy reveals that the information distribution discrimination regarding lexical and grammatical differences continues to increase from L2 learners at a lower level to those at a higher level. Our findings also offer insights into the cognitive aspect of the development of L2 language proficiency, namely that L2 learners at a lower level have to make greater cognitive efforts to improve their language proficiency. All this indicates that as compared to the algorithm of frequency summation and ratio, the ‘practical and effective’ algorithm that uses relative entropy could be more effective and stable when it comes to detecting differences of linguistic complexity between L2 learners than the frequency-based algorithm that was previously used.
Overall, as far as it seems, at least we have found a possible correct direction. In spite of this, there is still much to be done to improve the current algorithm, for example, by considering more units of measurement (discourse connectives, stance words, etc.). Moreover, we will consider the (random) factors available in L2 corpus (gender, age, nationality, text length, etc.). It would be better to control for these factors to further explore how they take effect on relative entropy using generalized mixed-effects statistical models. Such explorations could make relative entropy detect L2 proficiency development in a more reliable and solid manner.

Author Contributions

Conceptualization, K.S. and R.W.; methodology, K.S. and R.W.; formal analysis, K.S. and R.W.; data curation, K.S. and R.W.; writing—original draft preparation, K.S.; review and editing, K.S. and R.W. All authors have read and agreed to the published version of the manuscript.

Funding

We acknowledge support by Deutsche Forschungsgemeinschaft and Open Access Publishing Fund of University of Tübingen.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A. JSD Algorithm

The usefulness of KL is simply that it is the fundamental building block of the JS divergences. An alternate approach is the JS divergence, which is another method of measuring the similarity between two probability distributions. It is a symmetric and smoothed version of the KL divergence, and it can be used as a distance metric. The JSD has also been widely applied in various fields [19,61,62]. The JS divergence is symmetrical in two distributions and the weights are each ½. Defining the quantity M = (P + Q)*(0.5), we can write the JS divergence as:
JS (p||q)= 1/2*KLD(p||(p + q)/2) + 1/2*KLD(q||(p + q)/2),
JS(Levelh||Levell)= 1/2*KLD(Levelh||(Levelh + Levell)/2) + 1/2KLD(Levell||(Levelh + Levell)/2).
In Equation (A1), the KLD is the KL divergences in the above equations. Similarly, we can obtain the JSD for the lemma “take” in Table 1: JSD = (1820s||1810s) = 0.0249; JSD(1830s||1820s) = 0.0007; JSD(1840s||1830s) = 0.0014.
This subsection presents the results calculated by the JSD. As discussed in the Method section, the strength of the JSD is that it has a symmetrical and smoothed function in comparison to the KLD. More importantly, the results from the JSD can be used to cross-verify the KLD in order to ensure the validity and consistency of its results. Similar to the KLD, the JSD is able to detect the discrimination of information distribution between L2 learners at different levels. KLD and JSD can cross-validate with each other.

Appendix B. The Results from the JSD

This subsection presents the results calculated by the JSD. As discussed in the Method section, the strength of the JSD is that it has a symmetrical and smoothed function in comparison to the KLD. More importantly, the results from the JSD can be used to cross-verify the KLD in order to ensure the validity and consistency of its results.
Table A1 shows that the JSD between a given lower level (e.g., A1, A2, B1) and a higher level increases when L2 learners’ level of linguistic proficiency rises. However, there is an exception in the JSD of token starting from A2 level and that of token and lemma starting from B1 level. There the JSD values remain stable. Specifically, the upward trend is 75% (9/12) of all measurement units. Overall, this could suggest that the JSD between L2 learners at a higher level and L2 learners at a lower level is much larger than that between L2 learners at a lower level and L2 learners at an intermediate level. This indicates that L2 learners at a lower level need a greater amount of additional bits for encoding these linguistic units in comparison with L2 learners at a higher level. It also suggests that L2 learners at a lower level have a greater degree of cognitive cost in processing these linguistic units than L2 learners at a higher level. In short, the JSD data in Table A1 is consistent with the KLD data in Table 3.
Table A1. JSD results from the perspective of L2 at a lower level.
Table A1. JSD results from the perspective of L2 at a lower level.
Cross-Proficiency Levels of L2JSD of GrammarJSD of Lexicon
POS-TrigramSub-Conj.TokenLemma
A1→A2A1→(A2, B1, B2, C1)0.05 Entropy 23 01080 i001
(coef = 0.03,
p = 0.003)
0.09 Entropy 23 01080 i001
(coef = 0.007,
p = 0.34)
0.17 Entropy 23 01080 i001
(coef = 0.003,
p = 0.23)
0.06 Entropy 23 01080 i001
(coef = 0.02,
p = 0.17)
A1→B10.080.1170.180.11
A1→B20.10.10.180.13
A1→C10.130.120.180.12
A2→B1A2→(B1, B2, C1)0.03 Entropy 23 01080 i001
(coef = 0.2,
p = 0.18)
0.04 Entropy 23 01080 i001
(coef = 0.02,
p < 0.001)
0.11 0.08 Entropy 23 01080 i001
(coef = 0.005,
p = 0.33)
A2→B20.040.060.120.09
A2→C10.070.080.10.09
B1→B2B1→(B2, C1)0.02 Entropy 23 01080 i0010.014 Entropy 23 01080 i0010.1 0.06
B1→C10.030.0170.090.06
B2→C1B2→(C1)0.02 0.009 0.07 0.05
As Table A2 shows, the JSD between a higher level and a lower level is much larger than that between a lower level and the intermediate levels. The JSD drops when the level moves from the lower to the higher. It indicates that the distinction between different groups is proportional to their JSD. Specifically, the downward trend is 100% (12/12) of all measurement units. This could mean that the difference in language proficiency between L2 learners at a higher level and learners at an intermediate level is smaller than that between learners at the intermediate level and learners at the lower level. The JSD data in Table A2 is completely consistent with the KLD data in Table 4. The following examines the JSD of adjacent levels.
Table A2. JSD results from the perspective of L2 at a higher level.
Table A2. JSD results from the perspective of L2 at a higher level.
Cross-Proficiency Levels of L2JSD of GrammarJSD of Lexicon
POS-TrigramSub-Conj.TokenLemma
A1→C1(A1, A2, B1, B2)→C10.13 Entropy 23 01080 i002
(coef = −0.04,
p = 0.043)
0.12 Entropy 23 01080 i002
(coef = −0.04,
p = 0.034)
0.18 Entropy 23 01080 i002
(coef = −0.034,
p = 0.09)
0.11 Entropy 23 01080 i002
(coef = −0.02,
p = 0.015)
A2→C10.070.0750.10.09
B1→C10.030.0170.090.06
B2→C10.020.0090.070.05
A1→B2(A1, A2, B1)→B20.1 Entropy 23 01080 i002
(coef = −0.04,
p = 0.26)
0.1 Entropy 23 01080 i002
(coef = −0.043,
p = 0.008)
0.18 Entropy 23 01080 i002
(coef = −0.04,
p = 0.18)
0.13 Entropy 23 01080 i002
(coef = −0.035,
p = 0.05)
A2→B20.030.0580.120.09
B1→B20.020.0140.10.06
A1→B1(A1, A2)→B10.08 Entropy 23 01080 i0020.12 Entropy 23 01080 i0020.18 Entropy 23 01080 i0020.1 Entropy 23 01080 i002
A2→B10.030.040.110.08
A1→A2(A1)→A20.05 0.09 0.17 0.06
Table A3 shows that when levels become higher, their JSDs tend to become smaller. It could be that learners at intermediate levels have fewer difficulties in learning lexicon and syntax in the course of improving their language proficiency. By contrast, learners at the lower level may need to make more effort in improving their language proficiency. Specifically, the upward trend is 100% (4/4) of all measurement units. The data on the JSD in the three tables are consistent in supporting the thesis that the JSD between higher levels is much smaller than that between lower levels or between a lower level and an intermediate level. Overall, the JSD data are completely consistent with that of the KLD. The visualization of JSD results is presented in Figure 1 of the body text.
Table A3. JSD results from the perspective of adjacent L2 levels.
Table A3. JSD results from the perspective of adjacent L2 levels.
Cross-Proficiency Levels of L2JSD of GrammarJSD of Lexicon
POS-TrigramSub-Conj.TokenLemma
A1→A2adjacent levels0.05 Entropy 23 01080 i002
(coef = −0.01,
p = 0.087)
0.09(coef = −0.027,
p = 0.06)
0.17(coef = −0.03,
p = 0.046)
0.06 Entropy 23 01080 i002
(coef = −0.005,
p = 0.48)
A2→B10.030.040.110.08
B1→B20.020.0150.10.06
B2→C10.020.0090.070.05

Appendix C. The Data on Syntactic and Lexical Complexity

These syntactic metrics in L2SCA are included as follows: Length of production unit: MLC (# of words/# of clauses), MLS (# of words/# of sentences), MLT (# of words/# of T-units). Amount of subordination: C/T (# of clauses/# of T-unit), CT/T (# of complex T-units/# of T-units), DC/C (# of dependent clauses/# of clauses), DC/T (# of dependent clauses/# of T-units). Amount of coordination: CP/C (# of coordinate phrases/# of clauses), CP/T (# of coordinate phrases/# of T-units), T/S (# of T-units/# of sentences). Degree of phrasal sophistication: CN/C (# of complex nominals/# of clauses), CN/T (# of complex nominals/# of T-units), VP/T (# of verb phrases/# of T-units). Overall sentence complexity: C/S (# of clauses/# of sentences).
These lexical metric in LCA are included as follows: LD(lexical diversity), LS1(Lexical Sophistication-I), LS2(Lexical Sophistication-II), VS1(Verb Sophistication-I), CVS1(Corrected VS1), VS2(Verb Sophistication-II), TTR(Type–Token Ratio), CTTR(Corrected TTR), Root TTR(Root TTR), LogTTR(Bilogarithmic TTR), Uber(Uber Index), LV(Lexical Word Variation). VV1(Verb Variation-I), SVV1(Squared VV1), CVV1(Corrected VV1), VVII(Verb Variation-II), NV(Noun Variation).
The following two tables are the original data on syntactic and lexical complexity by the traditional methods.
Table A4. Syntactic complexity for each L2 level in EFCAMDAT2 and the difference between different L2 levels.
Table A4. Syntactic complexity for each L2 level in EFCAMDAT2 and the difference between different L2 levels.
LevelsMLSMLTMLCC/SVP/TC/TDC/CDC/TT/SCT/TCP/TCP/CCN/TCN/C
A18.568.177.041.211.31.160.110.131.050.090.240.20.580.5
A211.5910.097.581.531.331.330.210.291.150.230.320.240.760.57
B113.2812.018.531.561.411.410.270.371.110.30.30.211.090.78
B214.8513.68.931.661.51.520.330.51.090.340.320.211.340.88
C116.3214.589.541.711.531.530.330.51.120.340.390.261.561.02
A1_C17.766.412.50.490.230.370.220.370.070.240.160.050.980.52
A2_C14.724.491.960.180.20.20.110.21−0.030.110.080.020.80.45
B1_C13.042.561.010.150.120.120.060.130.010.040.090.040.470.24
B2_C14.724.491.960.180.20.20.110.21−0.030.110.080.020.80.45
A2_A13.041.930.540.310.30.170.10.160.10.130.080.030.170.07
B1_A14.723.851.490.340.530.250.160.250.060.20.060.010.510.27
B2_A16.295.441.890.450.750.360.220.380.040.290.090.010.760.38
C1_A17.766.412.50.490.770.370.220.370.070.280.160.050.980.52
A2_A13.041.930.540.310.30.170.10.160.10.130.080.030.170.07
B1_A21.681.920.950.030.230.080.050.09−0.040.07−0.01−0.020.340.21
B2_B11.571.590.40.110.220.120.070.13−0.010.080.0200.250.1
C1_B21.470.980.610.050.0200−0.010.03−0.010.070.040.220.14
Table A5. Lexical complexity for each L2 level in EFCAMDAT2 and the difference between different L2 levels.
Table A5. Lexical complexity for each L2 level in EFCAMDAT2 and the difference between different L2 levels.
Levelsldls1ls2vs1vs2cvs1ttrmsttrcttrrttrlogttruberlvvv1svv1cvv1vv2nv
A10.540.560.880.0243.424.660.030.7317.3224.490.7422.290.050.0260.945.520.010.07
A20.530.520.860.0471.495.980.040.7517.5524.820.7522.320.080.04106.917.310.010.11
B10.520.470.930.02143.278.460.020.7919.8228.040.7223.450.030.02167.39.150.010.04
B20.510.480.960.01180.819.510.010.7920.6129.140.7123.980.020.01198.999.9700.03
C1_A10.520.490.930.02186.369.650.020.820.1528.50.7323.470.030.02218.0810.440.010.05
A2_A1−0.01−0.04−0.020.0228.071.320.010.020.230.330.010.030.030.0245.971.7900.04
B1_A1−0.02−0.090.05099.853.8−0.010.062.53.55−0.021.16−0.020106.363.630−0.03
B2_A1−0.03−0.080.08−0.01137.394.85−0.020.063.294.65−0.031.69−0.03−0.01138.054.45−0.01−0.04
C1_A1−0.02−0.070.050142.944.99−0.010.072.834.01−0.011.18−0.020157.144.920−0.02
A2_A1−0.01−0.04−0.020.0228.071.320.010.020.230.330.010.030.030.0245.971.7900.04
B1_A2−0.01−0.050.07−0.0271.782.48−0.020.042.273.22−0.031.13−0.05−0.0260.391.840−0.07
B2_B1−0.010.010.03−0.0137.541.05−0.0100.791.1−0.010.53−0.01−0.0131.690.82−0.01−0.01
C1_B20.010.01−0.030.015.550.140.010.01−0.46−0.640.02−0.510.010.0119.090.470.010.02

References

  1. Bulté, B.; Housen, A. Defining and operationalising L2 complexity. In Dimensions of L2 Performance and Proficiency: Complexity, Accuracy and Fluency in SLA; Housen, A., Kuiken, F., Vedder, I., Eds.; John Benjamins: Amsterdam, The Netherlands, 2012; pp. 23–46. [Google Scholar]
  2. Ortega, L.; Iberri-Shea, G. Longitudinal research in second language acquisition: Recent trends and future directions. Annu. Rev. Appl. Linguist. 2005, 25, 26–45. [Google Scholar] [CrossRef]
  3. Skehan, P. Modelling second language performance: Integrating complexity, accuracy, fluency and lexis. Appl. Linguist. 2009, 30, 510–532. [Google Scholar] [CrossRef]
  4. Pallotti, G. CAF: Defining, refining and differentiating constructs. Appl. Linguist. 2009, 30, 590–601. [Google Scholar] [CrossRef]
  5. Housen, A.; De Clercq, B.; Kuiken, F.; Vedder, I. Multiple approaches to complexity in second language research. Second Lang. Res. 2019, 35, 3–21. [Google Scholar] [CrossRef]
  6. Crossley, S. Linguistic features in writing quality and development: An overview. J. Writ. Res. 2020, 11, 415–443. [Google Scholar] [CrossRef]
  7. Crossley, S.A.; Skalicky, S. Examining lexical development in second language learners: An approximate replication of Salsbury, Crossley & McNamara (2011). Lang. Teach. 2019, 52, 385–405. [Google Scholar]
  8. Lu, X. Automatic analysis of syntactic complexity in second language writing. Int. J. Corpus Linguist. 2010, 15, 474–496. [Google Scholar] [CrossRef]
  9. Lu, X. The relationship of lexical richness to the quality of ESL learners’ oral narratives. Mod. Lang. J. 2012, 96, 190–208. [Google Scholar] [CrossRef]
  10. Miestamo, M. Implicational hierarchies and grammatical complexity. In Language Complexity as an Evolving Variable; Sampson, G., Gil, D., Trudgill, P., Eds.; Oxford University Press: Oxford, UK, 2009; pp. 80–97. [Google Scholar]
  11. Wagner, R.K.; Puranik, C.S.; Foorman, B.; Foster, E.; Wilson, L.G.; Tschinkel, E.; Kantor, P.T. Modeling the development of written language. Read. Writ. 2011, 24, 203–220. [Google Scholar] [CrossRef] [Green Version]
  12. De Clercq, B.; Housen, A. A cross-linguistic perspective on syntactic complexity in L2 development: Syntactic elaboration and diversity. Mod. Lang. J. 2017, 101, 315–334. [Google Scholar] [CrossRef]
  13. Ellis, R. The differential effects of three types of task planning on the fluency, complexity, and accuracy in L2 oral production. Appl. Linguist. 2009, 3, 474–509. [Google Scholar] [CrossRef]
  14. Housen, A.; Kuiken, F. Complexity, accuracy, and fluency in second language acquisition. Appl. Linguist. 2009, 30, 461–473. [Google Scholar] [CrossRef] [Green Version]
  15. Peters, J.; Mulling, K.; Altun, Y. Relative entropy policy search. In Proceedings of the Twenty-Fourth AAAI Conference on Artificial Intelligence, Atlanta, GA, USA, 11–15 July 2010; pp. 1607–1612. [Google Scholar]
  16. Vedral, V. The role of relative entropy in quantum information theory. Rev. Mod. Phys. 2002, 74, 197. [Google Scholar] [CrossRef] [Green Version]
  17. Sayood, K. Information theory and cognition: A review. Entropy 2018, 20, 706. [Google Scholar] [CrossRef] [Green Version]
  18. Joe, H. Relative entropy measures of multivariate dependence. J. Am. Stat. Assoc. 1989, 84, 157–164. [Google Scholar] [CrossRef]
  19. Klingenstein, S.; Hitchcock, T.; DeDeo, S. The civilizing process in London’s Old Bailey. Proc. Natl. Acad. Sci. USA 2014, 111, 9419–9424. [Google Scholar] [CrossRef] [Green Version]
  20. Trotzke, A.; Zwart, J. The complexity of narrow syntax: Minimalism, representational economy and simplest Merge. In Measuring Grammatical Complexity; Newmeyer, F., Preston, L., Eds.; Oxford University Press: Oxford, UK, 2014; pp. 128–477. [Google Scholar]
  21. Arnold, J.; Wasow, T.; Losongco, A.; Ginstrom, R. Heaviness vs. newness: The effects of structural complexity and discourse status on constituent ordering. Language 2000, 17, 28–55. [Google Scholar] [CrossRef]
  22. Dahl, O. The Growth and Maintenance of Linguistic Complexity; John Benjamins: Amsterdam, The Netherlands, 2004. [Google Scholar]
  23. Givón, T. The Genesis of Syntactic Complexity: Diachrony, Ontogeny, Neuro-Cognition, Evolution; John Benjamins: Amsterdam, The Netherlands, 2009. [Google Scholar]
  24. Kusters, W. Complexity in linguistic theory language learning and language change. In Language Complexity: Typology, Contact, Change; Miestamo, M., Sinnemäki, K., Karlsson, F., Eds.; John Benjamins: Amsterdam, The Netherlands, 2008; pp. 3–22. [Google Scholar]
  25. Szmrecsanyi, B.; Kortmann, B. Introduction: Linguistic complexity: Second language acquisition indigenization contact. In Linguistic Complexity: Second Language Acquisition, Indigenization, Contact; Kortmann, B., Szmrecsanyi, B., Eds.; De Gruyter: Berlin, Germany, 2012; pp. 6–34. [Google Scholar]
  26. Juola, P. The time course of language change. Comput. Humanit. 2003, 37, 77–96. [Google Scholar] [CrossRef]
  27. Hughes, J.M.; Foti, N.J.; Krakauer, D.C.; Rockmore, D.N. Quantitative patterns of stylistic influence in the evolution of literature. Proc. Natl. Acad. Sci. USA 2012, 109, 7682–7686. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  28. Unsworth, S. Comparing child L2 development with adult L2 development: How to measure L2 proficiency. In Current Trends in Child Second Language Acquisition; Gavruseva, E., Haznedar, B., Eds.; John Benjamins: Amsterdam, The Netherlands, 2008; pp. 301–336. [Google Scholar]
  29. DeKeyser, R.M. What makes learning second-language grammar difficult? A review of issues. Lang. Learn. 2005, 55, 1–25. [Google Scholar] [CrossRef]
  30. Alexandre, Z.; Oleg, S.; Giovanni, P. An information-theoretic perspective on the costs of cognition. Neuropsychologia 2019, 123, 5–18. [Google Scholar]
  31. Kuiken, F.; Vedder, I.; Housen, A.; De Clercq, B. Variation in syntactic complexity: Introduction. Int. J. Appl. Linguist. 2019, 29, 161–170. [Google Scholar] [CrossRef] [Green Version]
  32. Mazgutova, D.; Kormos, J. Syntactic and lexical development in an intensive English for Academic Purposes programme. J. Second Lang. Writ. 2015, 29, 3–15. [Google Scholar] [CrossRef] [Green Version]
  33. Ansarifar, A.; Shahriari, H.; Pishghadam, R. Phrasal complexity in academic writing: A comparison of abstracts written by graduate students and expert writers in applied linguistics. J. Engl. Acad. Purp. 2018, 31, 58–71. [Google Scholar] [CrossRef]
  34. Paquot, M. Phraseological competence: A missing component in university entrance language tests? Insights from a study of EFL learners’ use of statistical collocations. Lang. Assess. Q. 2018, 15, 29–43. [Google Scholar] [CrossRef]
  35. Paquot, M. The phraseological dimension in interlanguage complexity research. Second Lang. Res. 2019, 35, 121–145. [Google Scholar] [CrossRef]
  36. Shannon, C.E. A mathematical theory of communication. Bell Syst. Tech. J. 1948, 27, 379–423. [Google Scholar] [CrossRef] [Green Version]
  37. Kullback, S.; Leibler, R.A. On information and sufficiency. Ann. Math. Stat. 1951, 22, 79–86. [Google Scholar] [CrossRef]
  38. Sun, K.; Liu, H.; Xiong, W. The evolutionary pattern of language in scientific writings: A case study of Philosophical Transactions of Royal Society (1665–1869). Scientometrics 2021, 126, 1695–1724. [Google Scholar] [CrossRef]
  39. Murdock, J.; Allen, C.; DeDeo, S. Exploration and exploitation of Victorian science in Darwin’s reading notebooks. Cognition 2017, 159, 117–126. [Google Scholar] [CrossRef] [Green Version]
  40. Chen, L.; Wise, S. Neuronal activity in the supplementary eye field during acquisition of conditional oculomotor associations. J. Neurophysiol. 1995, 73, 1101–1121. [Google Scholar] [CrossRef] [Green Version]
  41. Ehret, K.; Szmrecsanyi, B. Compressing learner language: An information-theoretic measure of complexity in SLA production data. Second Lang. Res. 2019, 35, 23–45. [Google Scholar] [CrossRef] [Green Version]
  42. Thoiron, P. Diversity index and entropy as measures of lexical richness. Comput. Humanit. 1986, 20, 197–202. [Google Scholar] [CrossRef]
  43. Hale, J. Information-theoretical complexity metrics. Lang. Linguist. Compass 2016, 10, 397–412. [Google Scholar] [CrossRef]
  44. Huang, Y.; Murakami, A.; Alexopoulou, T.; Korhonen, A. Dependency parsing of learner English. Int. J. Corpus Linguist. 2018, 23, 28–54. [Google Scholar] [CrossRef]
  45. Römer, U. A corpus perspective on the development of verb constructions in second language learners. Int. J. Corpus Linguist. 2019, 24, 268–290. [Google Scholar] [CrossRef]
  46. Lin, J. Divergence measures based on the Shannon entropy. IEEE Trans. Inf. Theory 1991, 37, 145–151. [Google Scholar] [CrossRef] [Green Version]
  47. Jurafsky, D.; Martin, J.H. Speech and Language Processing; Prentice Hall: Upper Saddle River, NJ, USA, 2008. [Google Scholar]
  48. Biber, D.; Johansson, S.; Leech, G.; Conrad, S.; Finegan, E.; Quirk, R. Longman Grammar of Spoken and Written English; Longman: London, UK, 1999; pp. 994–995. [Google Scholar]
  49. Degaetano-Ortlieb, S.; Teich, E. Towards an optimal code for communication: The case of scientific English. Corpus Linguist. Linguist. Theory 2019. [Google Scholar] [CrossRef] [Green Version]
  50. Comrie, B. The acquisition of relative clauses in relation to language typology. Stud. Second Lang. Acquis. 2007, 29, 301–309. [Google Scholar] [CrossRef]
  51. Wells, J.B.; Christiansen, M.H.; Race, D.S.; Acheson, D.J.; MacDonald, M.C. Experience and sentence processing: Statistical learning and relative clause comprehension. Cogn. Psychol. 2009, 58, 250–271. [Google Scholar] [CrossRef] [Green Version]
  52. Kyle, K.; Crossley, S.A. Measuring syntactic complexity in L2 writing using fine-grained clausal and phrasal indices. Mod. Lang. J. 2018, 102, 333–349. [Google Scholar] [CrossRef]
  53. Norris, J.M.; Ortega, L. Towards an organic approach to investigating CAF in instructed SLA: The case of complexity. Appl. Linguist. 2009, 30, 555–578. [Google Scholar] [CrossRef]
  54. Lu, X.; Ai, H. Syntactic complexity in college-level English writing: Differences among writers with diverse L1 backgrounds. J. Second Lang. Writ. 2015, 29, 16–27. [Google Scholar] [CrossRef]
  55. Vyatkina, N. The development of second language writing complexity in groups and individuals: A longitudinal learner corpus study. Mod. Lang. J. 2012, 96, 576–598. [Google Scholar] [CrossRef] [Green Version]
  56. De Bot, K.; Lowie, W.; Verspoor, M. A dynamic systems theory approach to second language acquisition. Biling. Lang. Cogn. 2007, 10, 7–21. [Google Scholar] [CrossRef] [Green Version]
  57. Kempe, V.; Brooks, P.J. Linking adult second language learning and diachronic change: A cautionary note. Front. Psychol. 2018, 9, 480. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  58. Hamilton, J.D. Time Series Analysis; Princeton University Press: Princeton, NJ, USA, 2020. [Google Scholar]
  59. Hyndman, R.J.; Athanasopoulos, G. Forecasting: Principles and Practice. Available online: https://otexts.com/fpp2/ (accessed on 8 August 2021).
  60. Crossley, S.A.; McNamara, D.S. Does writing development equal writing quality? A computational investigation of syntactic complexity in L2 learners. J. Second Lang. Writ. 2014, 26, 66–79. [Google Scholar] [CrossRef]
  61. Grosse, I.; Bernaola-Galván, P.; Carpena, P.; Román-Roldán, R.; Oliver, J.; Stanley, H.E. Analysis of symbolic sequences using the Jensen-Shannon divergence. Phys. Rev. E 2002, 65, 041905. [Google Scholar] [CrossRef] [Green Version]
  62. Liese, F.; Vajda, I. On divergences and informations in statistics and information theory. IEEE Trans. Inf. Theory 2006, 52, 4394–4412. [Google Scholar] [CrossRef] [Green Version]
Figure 1. Relative entropy among L2 learners at cross-proficiency different levels (EFCAMDAT2). Note that JSD results are also visualized in this figure.
Figure 1. Relative entropy among L2 learners at cross-proficiency different levels (EFCAMDAT2). Note that JSD results are also visualized in this figure.
Entropy 23 01080 g001
Figure 2. The difference of syntactic complexity between different L2 levels. Here, each L2 level (A1, A2, B1, B2, C1) can be treated as time order (date). Here, x-axis is syntactic complexity measures/metrics, and y-axis is the difference of those complexity metrics across proficiency levels (discussed in the section of Methods). The left top plot shows that MLC, MLS, and MLT have a gradual increase, that is, B1_A1 is higher than A2_A1, and B2_B1 is higher than B1_A1, and C1_B2 is higher than CB2_B1. When a metric shows a regular increase, it indicates that this measure can detect patterns of L2 proficiency development. By contrast, in the right top plot, such a regular increase can only be found in 4 of the 12 metrics. In the bottom two plots, none of metrics shows a regular increase. Irregular changes suggest that these metrics cannot capture the patterns of L2 proficiency development.
Figure 2. The difference of syntactic complexity between different L2 levels. Here, each L2 level (A1, A2, B1, B2, C1) can be treated as time order (date). Here, x-axis is syntactic complexity measures/metrics, and y-axis is the difference of those complexity metrics across proficiency levels (discussed in the section of Methods). The left top plot shows that MLC, MLS, and MLT have a gradual increase, that is, B1_A1 is higher than A2_A1, and B2_B1 is higher than B1_A1, and C1_B2 is higher than CB2_B1. When a metric shows a regular increase, it indicates that this measure can detect patterns of L2 proficiency development. By contrast, in the right top plot, such a regular increase can only be found in 4 of the 12 metrics. In the bottom two plots, none of metrics shows a regular increase. Irregular changes suggest that these metrics cannot capture the patterns of L2 proficiency development.
Entropy 23 01080 g002
Figure 3. The difference of lexical complexity between different L2 levels.
Figure 3. The difference of lexical complexity between different L2 levels.
Entropy 23 01080 g003
Table 1. Frequencies of “take” and its tokens from 1810s to 1840s.
Table 1. Frequencies of “take” and its tokens from 1810s to 1840s.
Word1810s1820s1830s1840s
Take728556665529
Took168327351333
Taken228345344324
Taking133144164165
Takes158637686
Total1415143516001437
Table 2. Composition of the five sub-corpora of the essays section of the EFCAMDAT2 by language proficiency level.
Table 2. Composition of the five sub-corpora of the essays section of the EFCAMDAT2 by language proficiency level.
L2 Learners’ Proficiency LevelsTextsLearnersTokensLemmas
A1625,985103,74228.8 M27,065
A2307,99652,73424 M32,051
B1168,36132,85218.4 M26,276
B261,32913,9519.3 M21,312
C114,69828392.8 M16,464
Table 3. The KLD results from the perspective of L2 at a lower level.
Table 3. The KLD results from the perspective of L2 at a lower level.
Cross-Proficiency Levels of L2KLD of GrammarKLD of Lexicon
POS-TrigramSub-Conj.TokenLemma
A1→A2A1→(A2, B1, B2, C1)0.37 Entropy 23 01080 i001
(coef = 0.21,
p < 0.001)
0.56 Entropy 23 01080 i001
(coef = 0.07,
p = 0.31)
1.41 Entropy 23 01080 i001
(coef = 0.03,
p = 0.18)
0.52 Entropy 23 01080 i001
(coef = 0.15,
p = 0.2)
A1→B10.580.811.410.84
A1→B20.780.681.421.09
A1→C11.00.821.520.93
A2→B1A2→(B1, B2, C1)0.2 Entropy 23 01080 i001
(coef = 0.16,
p = 0.13)
0.24 Entropy 23 01080 i001
(coef = 0.12,
p = 0.016)
1.29 Entropy 23 01080 i002
(coef = −0.25,
p = 0.13)
0.71 Entropy 23 01080 i001
(coef = 0.025,
p = 0.66)
A2→B20.30.351.130.81
A2→C10.510.470.780.76
B1→B2B1→(B2, C1)0.13 0.085 1.1 0.62
B1→C10.210.110.870.54
B2→C1B2→(C1)0.16 0.052 0.52 0.51
Table 4. The KLD results from the perspective of L2 at a higher level.
Table 4. The KLD results from the perspective of L2 at a higher level.
Cross-Proficiency Levels of L2KLD of GrammarKLD of Lexicon
POS-TrigramSub-Conj.TokenLemma
A1→C1(A1, A2, B1, B2)→C11.0 Entropy 23 01080 i002
(coef = −0.28,
p = 0.056)
0.82 Entropy 23 01080 i002
(coef = −0.27,
p = 0.034)
1.52 Entropy 23 01080 i002
(coef = −0.29,
p = 0.12)
0.93 Entropy 23 01080 i002
(coef = −0.15,
p = 0.034)
A2→C10.510.470.780.76
B1→C10.210.110.870.54
B2→C10.160.050.520.51
A1→B2(A1, A2, B1)→B20.78 Entropy 23 01080 i002
(coef = −0.33,
p = 0.171)
0.68 Entropy 23 01080 i002
(coef = −0.3,
p = 0.04)
1.42 Entropy 23 01080 i002
(coef = −0.16,
p = 0.28)
1.09 Entropy 23 01080 i002
(coef = −0.24,
p = 0.07)
A2→B20.30.351.130.81
B1→B20.130.0851.10.62
A1→B1(A1, A2)→B10.58 Entropy 23 01080 i0020.81 Entropy 23 01080 i0021.41 Entropy 23 01080 i0020.84 Entropy 23 01080 i002
A2→B10.20.241.280.71
A1→A2(A1)→A20.37 0.56 1.41 0.52
Table 5. The KLD results from the perspective of adjacent L2 levels.
Table 5. The KLD results from the perspective of adjacent L2 levels.
Cross-Proficiency Levels of L2KLD of GrammarKLD of Lexicon
POS-TrigramSub-Conj.TokenLemma
A1→A2adjacent levels0.37 Entropy 23 01080 i002
(coef = −0.07,
p = 0.16)
0.56 Entropy 23 01080 i002
(coef = −0.16,
p = 0.086)
1.41 Entropy 23 01080 i002
(coef = −0.28,
p = 0.063)
0.52
A2→B10.20.241.280.71
B1→B20.130.0851.10.62
B2→C10.160.0810.520.51
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Sun, K.; Wang, R. Using the Relative Entropy of Linguistic Complexity to Assess L2 Language Proficiency Development. Entropy 2021, 23, 1080. https://0-doi-org.brum.beds.ac.uk/10.3390/e23081080

AMA Style

Sun K, Wang R. Using the Relative Entropy of Linguistic Complexity to Assess L2 Language Proficiency Development. Entropy. 2021; 23(8):1080. https://0-doi-org.brum.beds.ac.uk/10.3390/e23081080

Chicago/Turabian Style

Sun, Kun, and Rong Wang. 2021. "Using the Relative Entropy of Linguistic Complexity to Assess L2 Language Proficiency Development" Entropy 23, no. 8: 1080. https://0-doi-org.brum.beds.ac.uk/10.3390/e23081080

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop