Linguistic Modeling and its Interfaces
Oberseminar, Detmar Meurers, Winter Semester 2016
This series features presentations and discussions of current issues in linguistic modeling and its interfaces. This includes linguistic modeling in computational linguistics, language acquisition research, Intelligent Computer-Assisted Language Learning, and education, as well as theoretical linguistic research with a focus on the interfaces of syntax and information structure. It is open to anyone interested in this interdisciplinary enterprise.
A list of talks in previous semesters can be found here:
Abstract While the formal pragmatic concepts in information structure, such as the focus of an
utterance, are precisely defined in theoretical linguistics and potentially very useful in conceptual
and practical terms, it has turned out to be difficult to reliably annotate such notions in
corpus data (Ritz et al., 2008; Calhoun et al., 2010). We present a large-scale focus
annotation effort designed to overcome this problem. Our annotation study is based
on the tasked-based corpus CREG (Ott et al., 2012), which consists of answers to
explicitly given reading comprehension questions. We compare focus annotation by
trained annotators with a crowd-sourcing setup making use of untrained native speakers.
Given the task context and an annotation process incrementally making the question
form and answer type explicit, the trained annotators reach substantial agreement for
focus annotation. Interestingly, the crowd-sourcing setup also supports high-quality
annotation – for specific subtypes of data. Finally, we turn to the question whether the
relevance of focus annotation can be extrinsically evaluated. We show that automatic
short-answer assessment significantly improves for focus annotated data. The focus
annotated CREG corpus is freely available and constitutes the largest such resource for
German.
Abstract Despite the promise of research conducted at the intersection of Computer- Assisted
Language Learning (CALL), Natural Language Processing (NLP) and Second Language
Acquisition (SLA), few studies have explored the potential benefits of using Intelligent
Computer-Assisted Language Learning (ICALL) systems to deepen our understanding of the
process and products of second language (L2) learning. The strategic use of technology offers
researchers novel methodological opportunities to examine how incremental changes in L2
development occur during treatment, as well as how the longitudinal impacts of experimental
interventions on L2 learning outcomes occur on a case-by-case basis. Drawing on the pilot results
from a project examining the effects of automatic input enhancement on L2 learners’ development,
this article explores how the use of technology offers additional methodological and analytical
choices for the investigation of the process and outcomes of L2 development, illustrating
the opportunities to study what learners do during visually enhanced instructional
activities.
Abstract Combining fair, informative questions into a balanced exam requires skill and experience.
What are the properties of a good test, and how can we ensure they are met? I will discuss the
”naive” approach to test creation and the use of probabilistic test theory and their respective
strengths: The naive approach (write questions, assign points, determine absolute point threshold
for passing) gives test-takers maximum transparency of expected performance, while item
norming according to probabilistic test theory more easily allows the automated creation
of tests from large item banks or user-adaptive testing. I will briefly present ongoing
research that uses insights from both methods to better understand questions (and
answers).
Abstract Informed by research on readability and language acquisition, computational linguists
have developed sophisticated tools for the analysis of linguistic complexity. While some tools are
starting to become accessible on the web, there still is a disconnect between the features that can
in principle be identified based on state-of-the-art computational linguistic analysis, and the
analyses a second language acquisition researcher, teacher, or textbook writer can readily obtain
and visualize for their own collection of texts. This short paper presents a web-based tool
development that aims to meet this challenge. The Common Text Analysis Platform (CTAP) is
designed to support fully configurable linguistic feature extraction for a wide range of complexity
analyses. It features a user-friendly interface, modularized and reusable analysis component
integration, and flexible corpus and feature management. Building on the Unstructured
Information Management framework (UIMA), CTAP readily supports integration of
state-of-the-art NLP and complexity feature extraction maintaining modularization and reusability.
CTAP thereby aims at providing a common platform for complexity analysis, encouraging research
collaboration and sharing of feature extraction components—to jointly advance the
state-of-the-art in complexity analysis in a form that readily supports real-life use by ordinary
users.
Abstract Ever since Learner Corpus Research (LCR) came into being, it has been viewed as having
great potential to impact upon the related fields of Second Language Acquisition (SLA) and
Foreign Language Teaching (FLT). Due to disparate research methodologies, however, the links
between the three fields have proven to be somewhat weak. While SLA studies have made
extensive use of experimental data and assessed the impact of a variety of learner variables,
in particular of the cognitive and affective types, LCR studies have often remained
monofactorial, relying on no more than a single predictor in their study of learner
language and omitting cognitive and affective variables from their designs. Likewise, even
though studies of FLT materials have been undertaken, results have not usually been
linked to the output produced by the learners taught via the materials in question.
In our study, we analyse learner language, experimental data, individual differences
and teaching materials. By assessing the impact of Content and Language Integrated
Learning (CLIL) on the acquisition of the English passive by German learners at the
secondary level, we show that, while monofactorial approaches may yield interesting results,
multifactorial designs, i.e. approaches taking into account a variety of predictors and different
types of data, necessarily provide greater insight into the processes at work in language
acquisition.
Abstract Many students are challenged with the demand of writing cohesive explanations. To
support students in writing cohesive explanations, we developed CohVis, a computer-based
feedback tool that visualizes cohesion deficits of students’ expository texts in a concept map.
CohVis uses natural language processing technologies and state-of-the-art visualization methods to
automatically provide students with formative graphical feedback about the cohesion of their texts.
In this talk, I will first provide an overview about the methodologies, implemented in CohVis.
Additionally, I present several studies in which we investigated the effectiveness of such feedback as
well as the underlying cognitive processes.
Abstract In this talk, I will discuss specific LFG proposals for the analysis of resultative and benefactive constructions. Some authors suggested phrasal analyses, that is, analyses that assume a certain fixed phrase structure that licenses these constructions. The proposals under consideration are interesting since they use a resource sensitive semantics (glue semantics) that makes it possible to use templates that remap arguments to different grammatical functions. This makes it possible to use inheritance techniques to inherit constraints that account for passivization. So, although arguments are added in phrase structure configurations passive can be accounted for lexically and interacts appropriately with arguments that are introduced configurationally.
I will show that such approaches nevertheless have shortcomings namely the interaction with
derivational morphology and the interaction of the constructions under consideration with other
phenomena like passivization and extraction. I will show that such phrasal approaches do
not capture cross-linguistic generalizations. I develop a lexical rule-based approach
in the framework of HPSG, which is also implemented in grammars of English and
German.
The semantics-based part of the original paper is rather complex and cannot be discussed in
full detail during the talk. Those who are really interested in these issues may read
the paper by Ashudeh et. al. 2014 to facilitate discussion during and after the talk:
http://web.stanford.edu/group/cslipublications/cslipublications/LFG/19/papers/lfg14asudehetal.pdf
Part of the material that will be covered in the talk is here:
http://hpsg.fu-berlin.de/~stefan/Pub/phrasal-lfg-headlex2016.html
Chris Biemann (Univ. Hamburg): Adaptive Natural Language Processing in the Cognitive Computing Paradigm
Abstract In the past decades of NLP, there has been a steady shift away from rule-based,
linguistically motivated modeling towards statistical learning and the induction of unsupervised
feature representations. However, natural language components used in today’s NLP pipelines are
still static in the sense that their statistical model or rule-base is created once, then subsequently
applied without further change. In this talk, I will motivate an adaptive approach to
natural language processing, where NLP components get smarter through usage over
time, following a ’cognitive computing’ approach to natural language processing. With
the help of recent research prototypes, three stages of data-driven adaptation will be
illustrated: feature/resource induction, induction of processing components and continuous
data-driven learning. Finally, I will discuss challenges in the evaluation of adaptive NLP
components.
Barbara Geist (Univ. Leipzig): Wie Schüler über Schreibungen sprechen – Erste
Ergebnisse aus dem Korpus ‘Rechtschreibgespräche’
Abstract Reziprokes Lernen, bei dem SchülerInnen sich untereinander Zusammenhänge erklären, zählt zu den Unterrichtsmaßnahmen mit besonders hohen Lerneffekten. Rechtschreibgespräche fordern die Kinder genau dazu auf und stellen sie vor die Herausforderung, Schreibungen systematisch und im kooperativen Austausch zu untersuchen und ihre Gedanken zu explizieren. Obwohl in praxisorientierten Zeitschriften zahlreiche Veröffentlichungen zur Initiierung und Gestaltung von Rechtschreibgesprächen vorliegen, gibt es bislang noch keine wissenschaftlichen Untersuchungen, die die gegenstandsbezogenen Lehr-Lern-Prozesse in diesem didaktischen Setting genauer in den Blick nehmen. In der Forschung zur Unterrichtskommunikation wurde bislang vorrangig (lehrerzentrierter) Unterricht im Klassenverband untersucht, selten Schüler-Schüler-Interaktionen in Kleingruppen (jedoch Riegler 2010). Außeracht gelassen wurde bislang in Studien zur Unterrichtskommunikation der sprachliche Hintergrund der Schüler (Deutsch als Erst- oder Zweitsprache).
In dem Projekt wird erforscht, wie Schüler mit Deutsch als Erst- und Zweitsprache in Kleingruppen und im Klassenverband über Schreibungen sprechen. Aus orthografiedidaktischer Sicht ist von Interesse, welche Schreibungen die Schüler als schwierig klassifizieren und welche Erklärungen die Schüler (gemeinsam) formulieren. Aus der Perspektive der Spracherwerbsforschung und dem Kompetenzbereich “Sprechen und Zuhören” (KMK 2004) wird analysiert, welche sprachlichen Strukturen die Schüler verwenden, um sich über den vorgegebenen fachlichen Inhalt auszutauschen. Zeigen sich Unterschiede für Parameter wie z.B. Äußerungslänge zwischen Schülern verschiedener Klassenstufen und/oder zwischen Schülern mit Deutsch als Erst- und Zweitsprache? Zukünftig ist neben qualitativen Analysen der Lehr-Lernprozesse geplant linguistische Korrelate von “Bildungssprache” zu analysieren (z.B. syntaktisch: Verbstellung, Umfang des Mittelfelds).
Das Design der Studie und Ergebnisse einer kontrastiven Fallanalyse eines Gesprächs von
Schülerinnen mit Deutsch als Erst- und Zweitsprache sowie ausschließlich von Schülerinnen mit
Deutsch als Zweitsprache werden vor- und zur Diskussion gestellt.
Abstract Standard supervised learning is learning from gold standard examples. This scenario is infeasible if examples have rich structure (parses, translations) and are required in large amounts. We present an approach to perform structured prediction by learning from interaction, following an online learning protocol where on each of a sequence of iterations, the learner receives an input, predicts an output structure, and receives partial feedback in form of a task loss evaluation of the predicted structure. Under this type of “bandit feedback”, the learner does not know what the correct prediction looks like, nor what would have happened if it had predicted differently. We present new learning objectives and algorithms for this interactive scenario, focusing on convergence speed and ease of elicitability of feedback, and several simulation experiments for different NLP tasks (see recent papers at ACL 2016 (http://www.cl.uni-heidelberg.de/~riezler/publications/papers/ACL2016.pdf) and NIPS 2016 (http://www.cl.uni-heidelberg.de/~riezler/publications/papers/NIPS2016.pdf)). Furthermore, we give an outlook on lifting this scenario to neural networks and how to perform offline learning for Bandit Structured Prediction.
_________________________________________________________________________________
Last updated: January 31, 2017