13
views
0
recommends
+1 Recommend
0 collections
    0
    shares
      • Record: found
      • Abstract: found
      • Article: found
      Is Open Access

      Morphosyntactic but not lexical corpus-based probabilities can substitute for cloze probabilities in reading experiments

      research-article

      Read this article at

      Bookmark
          There is no author summary for this article yet. Authors can add summaries to their articles on ScienceOpen to make them more accessible to a non-specialist audience.

          Abstract

          During reading or listening, people can generate predictions about the lexical and morphosyntactic properties of upcoming input based on available context. Psycholinguistic experiments that study predictability or control for it conventionally rely on a human-based approach and estimate predictability via the cloze task. Our study investigated an alternative corpus-based approach for estimating predictability via language predictability models. We obtained cloze and corpus-based probabilities for all words in 144 Russian sentences, correlated the two measures, and found a strong correlation between them. Importantly, we estimated how much variance in eye movements registered while reading the same sentences was explained by each of the two probabilities and whether the two probabilities explain the same variance. Along with lexical predictability (the activation of a particular word form), we analyzed morphosyntactic predictability (the activation of morphological features of words) and its effect on reading times over and above lexical predictability. We found that for predicting reading times, cloze and corpus-based measures of both lexical and morphosyntactic predictability explained the same amount of variance. However, cloze and corpus-based lexical probabilities both independently contributed to a better model fit, whereas for morphosyntactic probabilities, the contributions of cloze and corpus-based measures were interchangeable. Therefore, morphosyntactic but not lexical corpus-based probabilities can substitute for cloze probabilities in reading experiments. Our results also indicate that in languages with rich inflectional morphology, such as Russian, when people engage in prediction, they are much more successful in predicting isolated morphosyntactic features than predicting the particular lexeme and its full morphosyntactic markup.

          Related collections

          Most cited references40

          • Record: found
          • Abstract: found
          • Article: not found

          Long Short-Term Memory

          Learning to store information over extended time intervals by recurrent backpropagation takes a very long time, mostly because of insufficient, decaying error backflow. We briefly review Hochreiter's (1991) analysis of this problem, then address it by introducing a novel, efficient, gradient-based method called long short-term memory (LSTM). Truncating the gradient where this does not do harm, LSTM can learn to bridge minimal time lags in excess of 1000 discrete-time steps by enforcing constant error flow through constant error carousels within special units. Multiplicative gate units learn to open and close access to the constant error flow. LSTM is local in space and time; its computational complexity per time step and weight is O(1). Our experiments with artificial data involve local, distributed, real-valued, and noisy pattern representations. In comparisons with real-time recurrent learning, back propagation through time, recurrent cascade correlation, Elman nets, and neural sequence chunking, LSTM leads to many more successful runs, and learns much faster. LSTM also solves complex, artificial long-time-lag tasks that have never been solved by previous recurrent network algorithms.
            Bookmark
            • Record: found
            • Abstract: found
            • Article: not found

            What do we mean by prediction in language comprehension?

            We consider several key aspects of prediction in language comprehension: its computational nature, the representational level(s) at which we predict, whether we use higher level representations to predictively pre-activate lower level representations, and whether we 'commit' in any way to our predictions, beyond pre-activation. We argue that the bulk of behavioral and neural evidence suggests that we predict probabilistically and at multiple levels and grains of representation. We also argue that we can, in principle, use higher level inferences to predictively pre-activate information at multiple lower representational levels. We also suggest that the degree and level of predictive pre-activation might be a function of the expected utility of prediction, which, in turn, may depend on comprehenders' goals and their estimates of the relative reliability of their prior knowledge and the bottom-up input. Finally, we argue that all these properties of language understanding can be naturally explained and productively explored within a multi-representational hierarchical actively generative architecture whose goal is to infer the message intended by the producer, and in which predictions play a crucial role in explaining the bottom-up input.
              Bookmark
              • Record: found
              • Abstract: found
              • Article: not found
              Is Open Access

              The effect of word predictability on reading time is logarithmic.

              It is well known that real-time human language processing is highly incremental and context-driven, and that the strength of a comprehender's expectation for each word encountered is a key determinant of the difficulty of integrating that word into the preceding context. In reading, this differential difficulty is largely manifested in the amount of time taken to read each word. While numerous studies over the past thirty years have shown expectation-based effects on reading times driven by lexical, syntactic, semantic, pragmatic, and other information sources, there has been little progress in establishing the quantitative relationship between expectation (or prediction) and reading times. Here, by combining a state-of-the-art computational language model, two large behavioral data-sets, and non-parametric statistical techniques, we establish for the first time the quantitative form of this relationship, finding that it is logarithmic over six orders of magnitude in estimated predictability. This result is problematic for a number of established models of eye movement control in reading, but lends partial support to an optimal perceptual discrimination account of word recognition. We also present a novel model in which language processing is highly incremental well below the level of the individual word, and show that it predicts both the shape and time-course of this effect. At a more general level, this result provides challenges for both anticipatory processing and semantic integration accounts of lexical predictability effects. And finally, this result provides evidence that comprehenders are highly sensitive to relative differences in predictability - even for differences between highly unpredictable words - and thus helps bring theoretical unity to our understanding of the role of prediction at multiple levels of linguistic structure in real-time language comprehension. Copyright © 2013 The Authors. Published by Elsevier B.V. All rights reserved.
                Bookmark

                Author and article information

                Contributors
                Role: ConceptualizationRole: MethodologyRole: SupervisionRole: Writing – original draftRole: Writing – review & editing
                Role: Data curationRole: Formal analysisRole: MethodologyRole: ResourcesRole: SoftwareRole: VisualizationRole: Writing – review & editing
                Role: Formal analysisRole: InvestigationRole: MethodologyRole: SoftwareRole: VisualizationRole: Writing – review & editing
                Role: Editor
                Journal
                PLoS One
                PLoS One
                plos
                plosone
                PLoS ONE
                Public Library of Science (San Francisco, CA USA )
                1932-6203
                28 January 2021
                2021
                : 16
                : 1
                : e0246133
                Affiliations
                [1 ] Center for Language and Brain, HSE University, Moscow, Russia
                [2 ] Vinogradov Institute of the Russian Language, Moscow, Russia
                [3 ] Independent Researcher, Moscow, Russia
                [4 ] University of Potsdam, Potsdam, Germany
                University of Trento, ITALY
                Author notes

                Competing Interests: The authors have read the journal's policy and the authors of this manuscript have the following competing interests: KL is a paid employee of Scrapinghub. This does not alter our adherence to PLOS ONE policies on sharing data and materials. There are no patents, products in development or marketed products associated with this research to declare

                Author information
                https://orcid.org/0000-0002-2283-1644
                Article
                PONE-D-20-19003
                10.1371/journal.pone.0246133
                7842903
                33508029
                c635e898-35eb-40eb-8a01-81c5e702d9ba
                © 2021 Lopukhina et al

                This is an open access article distributed under the terms of the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited.

                History
                : 20 June 2020
                : 13 January 2021
                Page count
                Figures: 7, Tables: 8, Pages: 26
                Funding
                Funded by: Center for Language and Brain NRU Higher School of Economics
                Award ID: RF Government Grant, Ag. № 14.641.31.0004
                Award Recipient :
                Funded by: Center for Language and Brain NRU Higher School of Economics
                Award ID: RF Government Grant, Ag. № 14.641.31.0004
                Award Recipient :
                Funded by: Scrapinghub
                Award Recipient :
                The contribution of ALop and ALau was supported by the Center for Language and Brain NRU Higher School of Economics, RF Government Grant, Ag. № 14.641.31.0004. Scrapinghub provided support in the form of a salary to KL. The specific roles of this author are articulated in the ‘author contributions’ section. The funder had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.
                Categories
                Research Article
                Social Sciences
                Linguistics
                Linguistic Morphology
                Social Sciences
                Linguistics
                Semantics
                Biology and Life Sciences
                Neuroscience
                Cognitive Science
                Cognitive Psychology
                Language
                Biology and Life Sciences
                Psychology
                Cognitive Psychology
                Language
                Social Sciences
                Psychology
                Cognitive Psychology
                Language
                Biology and Life Sciences
                Physiology
                Sensory Physiology
                Visual System
                Eye Movements
                Biology and Life Sciences
                Neuroscience
                Sensory Systems
                Visual System
                Eye Movements
                Biology and Life Sciences
                Neuroscience
                Cognitive Science
                Cognitive Psychology
                Attention
                Biology and Life Sciences
                Psychology
                Cognitive Psychology
                Attention
                Social Sciences
                Psychology
                Cognitive Psychology
                Attention
                Social Sciences
                Linguistics
                Psycholinguistics
                Research and Analysis Methods
                Mathematical and Statistical Techniques
                Statistical Methods
                Forecasting
                Physical Sciences
                Mathematics
                Statistics
                Statistical Methods
                Forecasting
                Physical Sciences
                Mathematics
                Probability Theory
                Probability Distribution
                Custom metadata
                All data to this article can be found online at DOI 10.17605/OSF.IO/YQ46B.

                Uncategorized
                Uncategorized

                Comments

                Comment on this article