Alexander Koller


2024

pdf bib
Strengthening Structural Inductive Biases by Pre-training to Perform Syntactic Transformations
Matthias Lindemann | Alexander Koller | Ivan Titov
Proceedings of the 2024 Conference on Empirical Methods in Natural Language Processing

Models need appropriate inductive biases to effectively learn from small amounts of data and generalize systematically outside of the training distribution. While Transformers are highly versatile and powerful, they can still benefit from enhanced structural inductive biases for seq2seq tasks, especially those involving syntactic transformations, such as converting active to passive voice or semantic parsing. In this paper, we propose to strengthen the structural inductive bias of a Transformer by intermediate pre-training to perform synthetically generated syntactic transformations of dependency trees given a description of the transformation. Our experiments confirm that this helps with few-shot learning of syntactic tasks such as chunking, and also improves structural generalization for semantic parsing. Our analysis shows that the intermediate pre-training leads to attention heads that keep track of which syntactic transformation needs to be applied to which token, and that the model can leverage these attention heads on downstream tasks.

pdf bib
Scope-enhanced Compositional Semantic Parsing for DRT
Xiulin Yang | Jonas Groschwitz | Alexander Koller | Johan Bos
Proceedings of the 2024 Conference on Empirical Methods in Natural Language Processing

Discourse Representation Theory (DRT) distinguishes itself from other semantic representation frameworks by its ability to model complex semantic and discourse phenomena through structural nesting and variable binding. While seq2seq models hold the state of the art on DRT parsing, their accuracy degrades with the complexity of the sentence, and they sometimes struggle to produce well-formed DRT representations. We introduce the AMS parser, a compositional, neurosymbolic semantic parser for DRT. It rests on a novel mechanism for predicting quantifier scope. We show that the AMS parser reliably produces well-formed outputs and performs well on DRT parsing, especially on complex sentences.

pdf bib
ADaPT: As-Needed Decomposition and Planning with Language Models
Archiki Prasad | Alexander Koller | Mareike Hartmann | Peter Clark | Ashish Sabharwal | Mohit Bansal | Tushar Khot
Findings of the Association for Computational Linguistics: NAACL 2024

Large Language Models (LLMs) are increasingly being used for interactive decision-making tasks requiring planning and adapting to the environment. Recent works employ LLMs-as-agents in broadly two ways: iteratively determining the next action (iterative executors) or generating plans and executing sub-tasks using LLMs (plan-and-execute). However, these methods struggle with task complexity, as the inability to execute any sub-task may lead to task failure. To address these shortcomings, we introduce As-Needed Decomposition and Planning for complex Tasks (ADaPT), an approach that explicitly plans and decomposes complex sub-tasks as-needed, i.e., when the LLM is unable to execute them. ADaPT recursively decomposes sub-tasks to adapt to both task complexity and LLM capability. Our results demonstrate that ADaPT substantially outperforms established strong baselines, achieving success rates up to 28.3% higher in ALFWorld, 27% in WebShop, and 33% in TextCraft – a novel compositional dataset that we introduce. Through extensive analysis, we illustrate the importance of multilevel decomposition and establish that ADaPT dynamically adjusts to the capabilities of the executor LLM as well as to task complexity.

pdf bib
Predicting generalization performance with correctness discriminators
Yuekun Yao | Alexander Koller
Findings of the Association for Computational Linguistics: EMNLP 2024

The ability to predict an NLP model’s accuracy on unseen, potentially out-of-distribution data is a prerequisite for trustworthiness. We present a novel model that establishes upper and lower bounds on the accuracy, without requiring gold labels for the unseen data. We achieve this by training a *discriminator* which predicts whether the output of a given sequence-to-sequence model is correct or not. We show across a variety of tagging, parsing, and semantic parsing tasks that the gold accuracy is reliably between the predicted upper and lower bounds, and that these bounds are remarkably close together.

pdf bib
A Dialogue Game for Eliciting Balanced Collaboration
Isidora Jeknic | David Schlangen | Alexander Koller
Proceedings of the 25th Annual Meeting of the Special Interest Group on Discourse and Dialogue

Collaboration is an integral part of human dialogue. Typical task-oriented dialogue games assign asymmetric roles to the participants, which limits their ability to elicit naturalistic role-taking in collaboration and its negotiation. We present a novel and simple online setup that favors balanced collaboration: a two-player 2D object placement game in which the players must negotiate the goal state themselves. We show empirically that human players exhibit a variety of role distributions, and that balanced collaboration improves task performance. We also present an LLM-based baseline agent which demonstrates that automatic playing of our game is an interesting challenge for artificial systems.

pdf bib
Simple and effective data augmentation for compositional generalization
Yuekun Yao | Alexander Koller
Proceedings of the 2024 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies (Volume 1: Long Papers)

Compositional generalization, the ability to predict complex meanings from training on simpler sentences, poses challenges for powerful pretrained seq2seq models. In this paper, we show that data augmentation methods that sample MRs and backtranslate them can be effective for compositional generalization, but only if we sample from the right distribution. Remarkably, sampling from a uniform distribution performs almost as well as sampling from the test distribution, and greatly outperforms earlier methods that sampled from the training distribution.We further conduct experiments to investigate the reason why this happens and where the benefit of such data augmentation methods come from.

pdf bib
SIP: Injecting a Structural Inductive Bias into a Seq2Seq Model by Simulation
Matthias Lindemann | Alexander Koller | Ivan Titov
Proceedings of the 62nd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)

Strong inductive biases enable learning from little data and help generalization outside the training distribution. Popular neural architectures such as Transformers lack strong structural inductive biases for seq2seq NLP tasks on their own. Consequently, they struggle with systematic generalization beyond the training distribution, e.g. with extrapolating to longer inputs, even when pre-trained on large amounts of text.We show how a structural inductive bias can be efficiently injected into a seq2seq model by pre-training it to simulate structural transformations on synthetic data. Specifically, we inject an inductive bias towards Finite State Transducers (FSTs) into a Transformer by pre-training it to simulate FSTs given their descriptions. Our experiments show that our method imparts the desired inductive bias, resulting in improved systematic generalization and better few-shot learning for FST-like tasks. Our analysis shows that fine-tuned models accurately capture the state dynamics of the unseen underlying FSTs, suggesting that the simulation process is internalized by the fine-tuned model.

pdf bib
A Corpus of German Abstract Meaning Representation (DeAMR)
Christoph Otto | Jonas Groschwitz | Alexander Koller | Xiulin Yang | Lucia Donatelli
Proceedings of the 2024 Joint International Conference on Computational Linguistics, Language Resources and Evaluation (LREC-COLING 2024)

We present the first comprehensive set of guidelines for German Abstract Meaning Representation (Deutsche AMR, DeAMR) along with an annotated corpus of 400 DeAMR. Taking English AMR (EnAMR) as our starting point, we propose significant adaptations to faithfully represent the structure and semantics of German, focusing particularly on verb frames, compound words, and modality. We validate our annotation through inter-annotator agreement and further evaluate our corpus with a comparison of structural divergences between EnAMR and DeAMR on parallel sentences, replicating previous work that finds both cases of cross-lingual structural alignment and cases of meaningful linguistic divergence. Finally, we fine-tune state-of-the-art multi-lingual and cross-lingual AMR parsers on our corpus and find that, while our small corpus is insufficient to produce quality output, there is a need to continue develop and evaluate against gold non-English AMR data.

2023

pdf bib
What’s the Meaning of Superhuman Performance in Today’s NLU?
Simone Tedeschi | Johan Bos | Thierry Declerck | Jan Hajič | Daniel Hershcovich | Eduard Hovy | Alexander Koller | Simon Krek | Steven Schockaert | Rico Sennrich | Ekaterina Shutova | Roberto Navigli
Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)

In the last five years, there has been a significant focus in Natural Language Processing (NLP) on developing larger Pretrained Language Models (PLMs) and introducing benchmarks such as SuperGLUE and SQuAD to measure their abilities in language understanding, reasoning, and reading comprehension. These PLMs have achieved impressive results on these benchmarks, even surpassing human performance in some cases. This has led to claims of superhuman capabilities and the provocative idea that certain tasks have been solved. In this position paper, we take a critical look at these claims and ask whether PLMs truly have superhuman abilities and what the current benchmarks are really evaluating. We show that these benchmarks have serious limitations affecting the comparison between humans and PLMs and provide recommendations for fairer and more transparent benchmarks.

pdf bib
Compositional Generalization without Trees using Multiset Tagging and Latent Permutations
Matthias Lindemann | Alexander Koller | Ivan Titov
Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)

Seq2seq models have been shown to struggle with compositional generalization in semantic parsing, i.e. generalizing to unseen compositions of phenomena that the model handles correctly in isolation. We phrase semantic parsing as a two-step process: we first tag each input token with a multiset of output tokens. Then we arrange the tokens into an output sequence using a new way of parameterizing and predicting permutations. We formulate predicting a permutation as solving a regularized linear program and we backpropagate through the solver. In contrast to prior work, our approach does not place a priori restrictions on possible permutations, making it very expressive. Our model outperforms pretrained seq2seq models and prior work on realistic semantic parsing tasks that require generalization to longer examples. We also outperform non-tree-based models on structural generalization on the COGS benchmark. For the first time, we show that a model without an inductive bias provided by trees achieves high accuracy on generalization to deeper recursion depth.

pdf bib
We’re Afraid Language Models Aren’t Modeling Ambiguity
Alisa Liu | Zhaofeng Wu | Julian Michael | Alane Suhr | Peter West | Alexander Koller | Swabha Swayamdipta | Noah Smith | Yejin Choi
Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing

Ambiguity is an intrinsic feature of natural language. Managing ambiguity is a key part of human language understanding, allowing us to anticipate misunderstanding as communicators and revise our interpretations as listeners. As language models are increasingly employed as dialogue interfaces and writing aids, handling ambiguous language is critical to their success. We capture ambiguity in a sentence through its effect on entailment relations with another sentence, and collect AmbiEnt, a linguist-annotated benchmark of 1,645 examples with diverse kinds of ambiguity. We design a suite of tests based on AmbiEnt, presenting the first evaluation of pretrained LMs to recognize ambiguity and disentangle possible meanings. We find that the task remains extremely challenging, including for GPT-4, whose generated disambiguations are considered correct only 32% of the time in crowdworker evaluation, compared to 90% for disambiguations in our dataset. Finally, to illustrate the value of ambiguity-sensitive tools, we show that a multilabel NLI model can flag political claims in the wild that are misleading due to ambiguity. We encourage the field to rediscover the importance of ambiguity for NLP.

pdf bib
SLOG: A Structural Generalization Benchmark for Semantic Parsing
Bingzhi Li | Lucia Donatelli | Alexander Koller | Tal Linzen | Yuekun Yao | Najoung Kim
Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing

The goal of compositional generalization benchmarks is to evaluate how well models generalize to new complex linguistic expressions. Existing benchmarks often focus on lexical generalization, the interpretation of novel lexical items in syntactic structures familiar from training; structural generalization tasks, where a model needs to interpret syntactic structures that are themselves unfamiliar from training, are often underrepresented, resulting in overly optimistic perceptions of how well models can generalize. We introduce SLOG, a semantic parsing dataset that extends COGS (Kim and Linzen, 2020) with 17 structural generalization cases. In our experiments, the generalization accuracy of Transformer models, including pretrained ones, only reaches 40.6%, while a structure-aware parser only achieves 70.8%. These results are far from the near-perfect accuracy existing models achieve on COGS, demonstrating the role of SLOG in foregrounding the large discrepancy between models’ lexical and structural generalization capacities.

pdf bib
Compositional Generalisation with Structured Reordering and Fertility Layers
Matthias Lindemann | Alexander Koller | Ivan Titov
Proceedings of the 17th Conference of the European Chapter of the Association for Computational Linguistics

Seq2seq models have been shown to struggle with compositional generalisation, i.e. generalising to new and potentially more complex structures than seen during training. Taking inspiration from grammar-based models that excel at compositional generalisation, we present a flexible end-to-end differentiable neural model that composes two structural operations: a fertility step, which we introduce in this work, and a reordering step based on previous work (Wang et al., 2021). To ensure differentiability, we use the expected value of each step, which we compute using dynamic programming. Our model outperforms seq2seq models by a wide margin on challenging compositional splits of realistic semantic parsing tasks that require generalisation to longer examples. It also compares favourably to other models targeting compositional generalisation.

pdf bib
From Sentence to Action: Splitting AMR Graphs for Recipe Instructions
Katharina Stein | Lucia Donatelli | Alexander Koller
Proceedings of the Fourth International Workshop on Designing Meaning Representations

Accurately interpreting the relationships between actions in a recipe text is essential to successful recipe completion. We explore using Abstract Meaning Representation (AMR) to represent recipe instructions, abstracting away from syntax and sentence structure that may order recipe actions in arbitrary ways. We present an algorithm to split sentence-level AMRs into action-level AMRs for individual cooking steps. Our approach provides an automatic way to derive fine-grained AMR representations of actions in cooking recipes and can be a useful tool for downstream, instructional tasks.

2022

pdf bib
Structural generalization is hard for sequence-to-sequence models
Yuekun Yao | Alexander Koller
Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing

Sequence-to-sequence (seq2seq) models have been successful across many NLP tasks,including ones that require predicting linguistic structure. However, recent work on compositional generalization has shown that seq2seq models achieve very low accuracy in generalizing to linguistic structures that were not seen in training. We present new evidence that this is a general limitation of seq2seq models that is present not just in semantic parsing, but also in syntactic parsing and in text-to-text tasks, and that this limitation can often be overcome by neurosymbolic models that have linguistic knowledge built in. We further report on some experiments that give initial answers on the reasons for these limitations.

pdf bib
Compositional generalization with a broad-coverage semantic parser
Pia Weißenhorn | Lucia Donatelli | Alexander Koller
Proceedings of the 11th Joint Conference on Lexical and Computational Semantics

We show how the AM parser, a compositional semantic parser (Groschwitz et al., 2018) can solve compositional generalization on the COGS dataset. It is the first semantic parser that achieves high accuracy on both naturally occurring language and the synthetic COGS dataset. We discuss implications for corpus and model design for learning human-like generalization. Our results suggest that compositional generalization can be best achieved by building compositionality into semantic parsers.

pdf bib
Zero-shot Script Parsing
Fangzhou Zhai | Vera Demberg | Alexander Koller
Proceedings of the 29th International Conference on Computational Linguistics

Script knowledge is useful to a variety of NLP tasks. However, existing resources only cover a small number of activities, limiting its practical usefulness. In this work, we propose a zero-shot learning approach to script parsing, the task of tagging texts with scenario-specific event and participant types, which enables us to acquire script knowledge without domain-specific annotations. We (1) learn representations of potential event and participant mentions by promoting cluster consistency according to the annotated data; (2) perform clustering on the event / participant candidates from unannotated texts that belongs to an unseen scenario. The model achieves 68.1/74.4 average F1 for event / participant parsing, respectively, outperforming a previous CRF model that, in contrast, has access to scenario-specific supervision. We also evaluate the model by testing on a different corpus, where it achieved 55.5/54.0 average F1 for event / participant parsing.

2021

pdf bib
Script Parsing with Hierarchical Sequence Modelling
Fangzhou Zhai | Iza Škrjanec | Alexander Koller
Proceedings of *SEM 2021: The Tenth Joint Conference on Lexical and Computational Semantics

Scripts capture commonsense knowledge about everyday activities and their participants. Script knowledge proved useful in a number of NLP tasks, such as referent prediction, discourse classification, and story generation. A crucial step for the exploitation of script knowledge is script parsing, the task of tagging a text with the events and participants from a certain activity. This task is challenging: it requires information both about the ways events and participants are usually uttered in surface language as well as the order in which they occur in the world. We show how to do accurate script parsing with a hierarchical sequence model and transfer learning. Our model improves the state of the art of event parsing by over 16 points F-score and, for the first time, accurately tags script participants.

pdf bib
Learning compositional structures for semantic graph parsing
Jonas Groschwitz | Meaghan Fowlie | Alexander Koller
Proceedings of the 5th Workshop on Structured Prediction for NLP (SPNLP 2021)

AM dependency parsing is a method for neural semantic graph parsing that exploits the principle of compositionality. While AM dependency parsers have been shown to be fast and accurate across several graphbanks, they require explicit annotations of the compositional tree structures for training. In the past, these were obtained using complex graphbank-specific heuristics written by experts. Here we show how they can instead be trained directly on the graphs with a neural latent-variable model, drastically reducing the amount and complexity of manual heuristics. We demonstrate that our model picks up on several linguistic phenomena on its own and achieves comparable accuracy to supervised training, greatly facilitating the use of AM dependency parsing for new sembanks.

pdf bib
Aligning Actions Across Recipe Graphs
Lucia Donatelli | Theresa Schmidt | Debanjali Biswas | Arne Köhn | Fangzhou Zhai | Alexander Koller
Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing

Recipe texts are an idiosyncratic form of instructional language that pose unique challenges for automatic understanding. One challenge is that a cooking step in one recipe can be explained in another recipe in different words, at a different level of abstraction, or not at all. Previous work has annotated correspondences between recipe instructions at the sentence level, often glossing over important correspondences between cooking steps across recipes. We present a novel and fully-parsed English recipe corpus, ARA (Aligned Recipe Actions), which annotates correspondences between individual actions across similar recipes with the goal of capturing information implicit for accurate recipe understanding. We represent this information in the form of recipe graphs, and we train a neural model for predicting correspondences on ARA. We find that substantial gains in accuracy can be obtained by taking fine-grained structural information about the recipes into account.

2020

pdf bib
Predicting Coreference in Abstract Meaning Representations
Tatiana Anikina | Alexander Koller | Michael Roth
Proceedings of the Third Workshop on Computational Models of Reference, Anaphora and Coreference

This work addresses coreference resolution in Abstract Meaning Representation (AMR) graphs, a popular formalism for semantic parsing. We evaluate several current coreference resolution techniques on a recently published AMR coreference corpus, establishing baselines for future work. We also demonstrate that coreference resolution can improve the accuracy of a state-of-the-art semantic parser on this corpus.

pdf bib
Story Generation with Rich Details
Fangzhou Zhai | Vera Demberg | Alexander Koller
Proceedings of the 28th International Conference on Computational Linguistics

Automatically generated stories need to be not only coherent, but also interesting. Apart from realizing a story line, the text also needs to include rich details to engage the readers. We propose a model that features two different generation components: an outliner, which proceeds the main story line to realize global coherence; a detailer, which supplies relevant details to the story in a locally coherent manner. Human evaluations show our model substantially improves the informativeness of generated text while retaining its coherence, outperforming various baselines.

pdf bib
Generating Instructions at Different Levels of Abstraction
Arne Köhn | Julia Wichlacz | Álvaro Torralba | Daniel Höller | Jörg Hoffmann | Alexander Koller
Proceedings of the 28th International Conference on Computational Linguistics

When generating technical instructions, it is often convenient to describe complex objects in the world at different levels of abstraction. A novice user might need an object explained piece by piece, while for an expert, talking about the complex object (e.g. a wall or railing) directly may be more succinct and efficient. We show how to generate building instructions at different levels of abstraction in Minecraft. We introduce the use of hierarchical planning to this end, a method from AI planning which can capture the structure of complex objects neatly. A crowdsourcing evaluation shows that the choice of abstraction level matters to users, and that an abstraction strategy which balances low-level and high-level object descriptions compares favorably to ones which don’t.

pdf bib
Normalizing Compositional Structures Across Graphbanks
Lucia Donatelli | Jonas Groschwitz | Matthias Lindemann | Alexander Koller | Pia Weißenhorn
Proceedings of the 28th International Conference on Computational Linguistics

The emergence of a variety of graph-based meaning representations (MRs) has sparked an important conversation about how to adequately represent semantic structure. MRs exhibit structural differences that reflect different theoretical and design considerations, presenting challenges to uniform linguistic analysis and cross-framework semantic parsing. Here, we ask the question of which design differences between MRs are meaningful and semantically-rooted, and which are superficial. We present a methodology for normalizing discrepancies between MRs at the compositional level (Lindemann et al., 2019), finding that we can normalize the majority of divergent phenomena using linguistically-grounded rules. Our work significantly increases the match in compositional structure between MRs and improves multi-task learning (MTL) in a low-resource setting, serving as a proof of concept for future broad-scale cross-MR normalization.

pdf bib
Climbing towards NLU: On Meaning, Form, and Understanding in the Age of Data
Emily M. Bender | Alexander Koller
Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics

The success of the large neural language models on many NLP tasks is exciting. However, we find that these successes sometimes lead to hype in which these models are being described as “understanding” language or capturing “meaning”. In this position paper, we argue that a system trained only on form has a priori no way to learn meaning. In keeping with the ACL 2020 theme of “Taking Stock of Where We’ve Been and Where We’re Going”, we argue that a clear understanding of the distinction between form and meaning will help guide the field towards better science around natural language understanding.

pdf bib
MC-Saar-Instruct: a Platform for Minecraft Instruction Giving Agents
Arne Köhn | Julia Wichlacz | Christine Schäfer | Álvaro Torralba | Joerg Hoffmann | Alexander Koller
Proceedings of the 21th Annual Meeting of the Special Interest Group on Discourse and Dialogue

We present a comprehensive platform to run human-computer experiments where an agent instructs a human in Minecraft, a 3D blocksworld environment. This platform enables comparisons between different agents by matching users to agents. It performs extensive logging and takes care of all boilerplate, allowing to easily incorporate new agents to evaluate them. Our environment is prepared to evaluate any kind of instruction giving system, recording the interaction and all actions of the user. We provide example architects, a Wizard-of-Oz architect and set-up scripts to automatically download, build and start the platform.

pdf bib
Fast semantic parsing with well-typedness guarantees
Matthias Lindemann | Jonas Groschwitz | Alexander Koller
Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP)

AM dependency parsing is a linguistically principled method for neural semantic parsing with high accuracy across multiple graphbanks. It relies on a type system that models semantic valency but makes existing parsers slow. We describe an A* parser and a transition-based parser for AM dependency parsing which guarantee well-typedness and improve parsing speed by up to 3 orders of magnitude, while maintaining or improving accuracy.

2019

pdf bib
Semantic Expressive Capacity with Bounded Memory
Antoine Venant | Alexander Koller
Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics

We investigate the capacity of mechanisms for compositional semantic parsing to describe relations between sentences and semantic representations. We prove that in order to represent certain relations, mechanisms which are syntactically projective must be able to remember an unbounded number of locations in the semantic representations, where nonprojective mechanisms need not. This is the first result of this kind, and has consequences both for grammar-based and for neural systems.

pdf bib
Compositional Semantic Parsing across Graphbanks
Matthias Lindemann | Jonas Groschwitz | Alexander Koller
Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics

Most semantic parsers that map sentences to graph-based meaning representations are hand-designed for specific graphbanks. We present a compositional neural semantic parser which achieves, for the first time, competitive accuracies across a diverse range of graphbanks. Incorporating BERT embeddings and multi-task learning improves the accuracy further, setting new states of the art on DM, PAS, PSD, AMR 2015 and EDS.

pdf bib
Graph-Based Meaning Representations: Design and Processing
Alexander Koller | Stephan Oepen | Weiwei Sun
Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics: Tutorial Abstracts

This tutorial is on representing and processing sentence meaning in the form of labeled directed graphs. The tutorial will (a) briefly review relevant background in formal and linguistic semantics; (b) semi-formally define a unified abstract view on different flavors of semantic graphs and associated terminology; (c) survey common frameworks for graph-based meaning representation and available graph banks; and (d) offer a technical overview of a representative selection of different parsing approaches.

pdf bib
Saarland at MRP 2019: Compositional parsing across all graphbanks
Lucia Donatelli | Meaghan Fowlie | Jonas Groschwitz | Alexander Koller | Matthias Lindemann | Mario Mina | Pia Weißenhorn
Proceedings of the Shared Task on Cross-Framework Meaning Representation Parsing at the 2019 Conference on Natural Language Learning

We describe the Saarland University submission to the shared task on Cross-Framework Meaning Representation Parsing (MRP) at the 2019 Conference on Computational Natural Language Learning (CoNLL).

pdf bib
Talking about what is not there: Generating indefinite referring expressions in Minecraft
Arne Köhn | Alexander Koller
Proceedings of the 12th International Conference on Natural Language Generation

When generating technical instructions, it is often necessary to describe an object that does not exist yet. For example, an NLG system which explains how to build a house needs to generate sentences like “build *a wall of height five to your left*” and “now build *a wall on the other side*.” Generating (indefinite) referring expressions to objects that do not exist yet is fundamentally different from generating the usual definite referring expressions, because the new object must be distinguished from an infinite set of possible alternatives. We formalize this problem and present an algorithm for generating such expressions, in the context of generating building instructions within the Minecraft video game.

2018

pdf bib
AMR dependency parsing with a typed semantic algebra
Jonas Groschwitz | Matthias Lindemann | Meaghan Fowlie | Mark Johnson | Alexander Koller
Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)

We present a semantic parser for Abstract Meaning Representations which learns to parse strings into tree representations of the compositional structure of an AMR graph. This allows us to use standard neural techniques for supertagging and dependency tree parsing, constrained by a linguistically principled type system. We present two approximative decoding algorithms, which achieve state-of-the-art accuracy and outperform strong baselines.

pdf bib
Generalized chart constraints for efficient PCFG and TAG parsing
Stefan Grünewald | Sophie Henning | Alexander Koller
Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers)

Chart constraints, which specify at which string positions a constituent may begin or end, have been shown to speed up chart parsers for PCFGs. We generalize chart constraints to more expressive grammar formalisms and describe a neural tagger which predicts chart constraints at very high precision. Our constraints accelerate both PCFG and TAG parsing, and combine effectively with other pruning techniques (coarse-to-fine and supertagging) for an overall speedup of two orders of magnitude, while improving accuracy.

pdf bib
Discovering User Groups for Natural Language Generation
Nikos Engonopoulos | Christoph Teichmann | Alexander Koller
Proceedings of the 19th Annual SIGdial Meeting on Discourse and Dialogue

We present a model which predicts how individual users of a dialog system understand and produce utterances based on user groups. In contrast to previous work, these user groups are not specified beforehand, but learned in training. We evaluate on two referring expression (RE) generation tasks; our experiments show that our model can identify user groups and learn how to most effectively talk to them, and can dynamically assign unseen users to the correct groups as they interact with the system.

2017

pdf bib
Proceedings of the 15th Conference of the European Chapter of the Association for Computational Linguistics: Volume 1, Long Papers
Mirella Lapata | Phil Blunsom | Alexander Koller
Proceedings of the 15th Conference of the European Chapter of the Association for Computational Linguistics: Volume 1, Long Papers

pdf bib
Proceedings of the 15th Conference of the European Chapter of the Association for Computational Linguistics: Volume 2, Short Papers
Mirella Lapata | Phil Blunsom | Alexander Koller
Proceedings of the 15th Conference of the European Chapter of the Association for Computational Linguistics: Volume 2, Short Papers

pdf bib
Alto: Rapid Prototyping for Parsing and Translation
Johannes Gontrum | Jonas Groschwitz | Alexander Koller | Christoph Teichmann
Proceedings of the Software Demonstrations of the 15th Conference of the European Chapter of the Association for Computational Linguistics

We present Alto, a rapid prototyping tool for new grammar formalisms. Alto implements generic but efficient algorithms for parsing, translation, and training for a range of monolingual and synchronous grammar formalisms. It can easily be extended to new formalisms, which makes all of these algorithms immediately available for the new formalism.

pdf bib
Integrated sentence generation using charts
Alexander Koller | Nikos Engonopoulos
Proceedings of the 10th International Conference on Natural Language Generation

Integrating surface realization and the generation of referring expressions into a single algorithm can improve the quality of the generated sentences. Existing algorithms for doing this, such as SPUD and CRISP, are search-based and can be slow or incomplete. We offer a chart-based algorithm for integrated sentence generation and demonstrate its runtime efficiency.

pdf bib
A Feature Structure Algebra for FTAG
Alexander Koller
Proceedings of the 13th International Workshop on Tree Adjoining Grammars and Related Formalisms

pdf bib
Parsing Minimalist Languages with Interpreted Regular Tree Grammars
Meaghan Fowlie | Alexander Koller
Proceedings of the 13th International Workshop on Tree Adjoining Grammars and Related Formalisms

pdf bib
Coarse-To-Fine Parsing for Expressive Grammar Formalisms
Christoph Teichmann | Alexander Koller | Jonas Groschwitz
Proceedings of the 15th International Conference on Parsing Technologies

We generalize coarse-to-fine parsing to grammar formalisms that are more expressive than PCFGs and/or describe languages of trees or graphs. We evaluate our algorithm on PCFG, PTAG, and graph parsing. While we achieve the expected performance gains on PCFGs, coarse-to-fine does not help for PTAG and can even slow down parsing for graphs. We discuss the implications of this finding.

pdf bib
A constrained graph algebra for semantic parsing with AMRs
Jonas Groschwitz | Meaghan Fowlie | Mark Johnson | Alexander Koller
Proceedings of the 12th International Conference on Computational Semantics (IWCS) — Long papers

pdf bib
Generating Contrastive Referring Expressions
Martín Villalba | Christoph Teichmann | Alexander Koller
Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)

The referring expressions (REs) produced by a natural language generation (NLG) system can be misunderstood by the hearer, even when they are semantically correct. In an interactive setting, the NLG system can try to recognize such misunderstandings and correct them. We present an algorithm for generating corrective REs that use contrastive focus (“no, the BLUE button”) to emphasize the information the hearer most likely misunderstood. We show empirically that these contrastive REs are preferred over REs without contrast marking.

2016

pdf bib
Adaptive Importance Sampling from Finite State Automata
Christoph Teichmann | Kasimir Wansing | Alexander Koller
Proceedings of the SIGFSM Workshop on Statistical NLP and Weighted Automata

pdf bib
Proceedings of the 12th International Workshop on Tree Adjoining Grammars and Related Formalisms (TAG+12)
David Chiang | Alexander Koller
Proceedings of the 12th International Workshop on Tree Adjoining Grammars and Related Formalisms (TAG+12)

pdf bib
Efficient techniques for parsing with tree automata
Jonas Groschwitz | Alexander Koller | Mark Johnson
Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)

2015

pdf bib
Semantic Dependency Graph Parsing Using Tree Approximations
Željko Agić | Alexander Koller | Stephan Oepen
Proceedings of the 11th International Conference on Computational Semantics

pdf bib
Semantic construction with graph grammars
Alexander Koller
Proceedings of the 11th International Conference on Computational Semantics

pdf bib
Proceedings of the 16th Annual Meeting of the Special Interest Group on Discourse and Dialogue
Alexander Koller | Gabriel Skantze | Filip Jurcicek | Masahiro Araki | Carolyn Penstein Rose
Proceedings of the 16th Annual Meeting of the Special Interest Group on Discourse and Dialogue

pdf bib
Graph parsing with s-graph grammars
Jonas Groschwitz | Alexander Koller | Christoph Teichmann
Proceedings of the 53rd Annual Meeting of the Association for Computational Linguistics and the 7th International Joint Conference on Natural Language Processing (Volume 1: Long Papers)

pdf bib
The Impact of Listener Gaze on Predicting Reference Resolution
Nikolina Koleva | Martín Villalba | Maria Staudte | Alexander Koller
Proceedings of the 53rd Annual Meeting of the Association for Computational Linguistics and the 7th International Joint Conference on Natural Language Processing (Volume 2: Short Papers)

pdf bib
Lexicalization and Generative Power in CCG
Marco Kuhlmann | Alexander Koller | Giorgio Satta
Computational Linguistics, Volume 41, Issue 2 - June 2015

2014

pdf bib
Potsdam: Semantic Dependency Parsing by Bidirectional Graph-Tree Transformations and Syntactic Parsing
Željko Agić | Alexander Koller
Proceedings of the 8th International Workshop on Semantic Evaluation (SemEval 2014)

pdf bib
Generating effective referring expressions using charts
Nikolaos Engonopoulos | Alexander Koller
Proceedings of the INLG and SIGDIAL 2014 Joint Session

2013

pdf bib
Predicting the Resolution of Referring Expressions from User Behavior
Nikos Engonopoulos | Martín Villalba | Ivan Titov | Alexander Koller
Proceedings of the 2013 Conference on Empirical Methods in Natural Language Processing

pdf bib
General binarization for parsing and translation
Matthias Büchse | Alexander Koller | Heiko Vogler
Proceedings of the 51st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)

pdf bib
Incremental, Predictive Parsing with Psycholinguistically Motivated Tree-Adjoining Grammar
Vera Demberg | Frank Keller | Alexander Koller
Computational Linguistics, Volume 39, Issue 4 - December 2013

bib
Proceedings of the 10th International Conference on Computational Semantics (IWCS 2013) – Long Papers
Alexander Koller | Katrin Erk
Proceedings of the 10th International Conference on Computational Semantics (IWCS 2013) – Long Papers

bib
Proceedings of the 10th International Conference on Computational Semantics (IWCS 2013) – Short Papers
Alexander Koller | Katrin Erk
Proceedings of the 10th International Conference on Computational Semantics (IWCS 2013) – Short Papers

2012

pdf bib
Enhancing Referential Success by Tracking Hearer Gaze
Alexander Koller | Konstantina Garoufi | Maria Staudte | Matthew Crocker
Proceedings of the 13th Annual Meeting of the Special Interest Group on Discourse and Dialogue

pdf bib
Decomposing TAG Algorithms Using Simple Algebraizations
Alexander Koller | Marco Kuhlmann
Proceedings of the 11th International Workshop on Tree Adjoining Grammars and Related Formalisms (TAG+11)

pdf bib
Generation of landmark-based navigation instructions from open-source data
Markus Dräger | Alexander Koller
Proceedings of the 13th Conference of the European Chapter of the Association for Computational Linguistics

2011

pdf bib
Combining symbolic and corpus-based approaches for the generation of successful referring expressions
Konstantina Garoufi | Alexander Koller
Proceedings of the 13th European Workshop on Natural Language Generation

pdf bib
Generation Challenges 2011 Preface
Anja Belz | Albert Gatt | Alexander Koller | Kristina Striegnitz
Proceedings of the 13th European Workshop on Natural Language Generation

pdf bib
Report on the Second Second Challenge on Generating Instructions in Virtual Environments (GIVE-2.5)
Kristina Striegnitz | Alexandre Denis | Andrew Gargett | Konstantina Garoufi | Alexander Koller | Mariët Theune
Proceedings of the 13th European Workshop on Natural Language Generation

pdf bib
The Potsdam NLG systems at the GIVE-2.5 Challenge
Konstantina Garoufi | Alexander Koller
Proceedings of the 13th European Workshop on Natural Language Generation

pdf bib
A Generalized View on Parsing and Translation
Alexander Koller | Marco Kuhlmann
Proceedings of the 12th International Conference on Parsing Technologies

pdf bib
Learning Script Participants from Unlabeled Data
Michaela Regneri | Alexander Koller | Josef Ruppenhofer | Manfred Pinkal
Proceedings of the International Conference Recent Advances in Natural Language Processing 2011

2010

pdf bib
Computing Weakest Readings
Alexander Koller | Stefan Thater
Proceedings of the 48th Annual Meeting of the Association for Computational Linguistics

pdf bib
The Importance of Rule Restrictions in CCG
Marco Kuhlmann | Alexander Koller | Giorgio Satta
Proceedings of the 48th Annual Meeting of the Association for Computational Linguistics

pdf bib
Learning Script Knowledge with Web Experiments
Michaela Regneri | Alexander Koller | Manfred Pinkal
Proceedings of the 48th Annual Meeting of the Association for Computational Linguistics

pdf bib
Automated Planning for Situated Natural Language Generation
Konstantina Garoufi | Alexander Koller
Proceedings of the 48th Annual Meeting of the Association for Computational Linguistics

pdf bib
Generation Challenges 2010 Preface
Anja Belz | Albert Gatt | Alexander Koller
Proceedings of the 6th International Natural Language Generation Conference

pdf bib
Report on the Second NLG Challenge on Generating Instructions in Virtual Environments (GIVE-2)
Alexander Koller | Kristina Striegnitz | Andrew Gargett | Donna Byron | Justine Cassell | Robert Dale | Johanna Moore | Jon Oberlander
Proceedings of the 6th International Natural Language Generation Conference

pdf bib
Sentence Generation as Planning with Probabilistic LTAG
Daniel Bauer | Alexander Koller
Proceedings of the 10th International Workshop on Tree Adjoining Grammar and Related Frameworks (TAG+10)

pdf bib
The GIVE-2 Corpus of Giving Instructions in Virtual Environments
Andrew Gargett | Konstantina Garoufi | Alexander Koller | Kristina Striegnitz
Proceedings of the Seventh International Conference on Language Resources and Evaluation (LREC'10)

We present the GIVE-2 Corpus, a new corpus of human instruction giving. The corpus was collected by asking one person in each pair of subjects to guide the other person towards completing a task in a virtual 3D environment with typed instructions. This is the same setting as that of the recent GIVE Challenge, and thus the corpus can serve as a source of data and as a point of comparison for NLG systems that participate in the GIVE Challenge. The instruction-giving data we collect is multilingual (45 German and 63 English dialogues), and can easily be extended to further languages by using our software, which we have made available. We analyze the corpus to study the effects of learning by repeated participation in the task and the effects of the participants' spatial navigation abilities. Finally, we present a novel annotation scheme for situated referring expressions and compare the referring expressions in the German and English data.

2009

pdf bib
A Logic of Semantic Representations for Shallow Parsing
Alexander Koller | Alex Lascarides
Proceedings of the 12th Conference of the European Chapter of the ACL (EACL 2009)

pdf bib
Dependency Trees and the Strong Generative Capacity of CCG
Alexander Koller | Marco Kuhlmann
Proceedings of the 12th Conference of the European Chapter of the ACL (EACL 2009)

pdf bib
The Software Architecture for the First Challenge on Generating Instructions in Virtual Environments
Alexander Koller | Donna Byron | Justine Cassell | Robert Dale | Johanna Moore | Jon Oberlander | Kristina Striegnitz
Proceedings of the Demonstrations Session at EACL 2009

pdf bib
Report on the First NLG Challenge on Generating Instructions in Virtual Environments (GIVE)
Donna Byron | Alexander Koller | Kristina Striegnitz | Justine Cassell | Robert Dale | Johanna Moore | Jon Oberlander
Proceedings of the 12th European Workshop on Natural Language Generation (ENLG 2009)

pdf bib
Validating the web-based evaluation of NLG systems
Alexander Koller | Kristina Striegnitz | Donna Byron | Justine Cassell | Robert Dale | Sara Dalzel-Job | Johanna Moore | Jon Oberlander
Proceedings of the ACL-IJCNLP 2009 Conference Short Papers

2008

pdf bib
Referring Expressions as Formulas of Description Logic
Carlos Areces | Alexander Koller | Kristina Striegnitz
Proceedings of the Fifth International Natural Language Generation Conference

pdf bib
Regular Tree Grammars as a Formalism for Scope Underspecification
Alexander Koller | Michaela Regneri | Stefan Thater
Proceedings of ACL-08: HLT

pdf bib
Efficient Processing of Underspecified Discourse Representations
Michaela Regneri | Markus Egg | Alexander Koller
Proceedings of ACL-08: HLT, Short Papers

2007

pdf bib
Sentence generation as a planning problem
Alexander Koller | Matthew Stone
Proceedings of the 45th Annual Meeting of the Association of Computational Linguistics

2006

pdf bib
An Improved Redundancy Elimination Algorithm for Underspecified Representations
Alexander Koller | Stefan Thater
Proceedings of the 21st International Conference on Computational Linguistics and 44th Annual Meeting of the Association for Computational Linguistics

pdf bib
Proceedings of the Fifth International Workshop on Inference in Computational Semantics (ICoS-5)
Johan Bos | Alexander Koller
Proceedings of the Fifth International Workshop on Inference in Computational Semantics (ICoS-5)

pdf bib
Towards a redundancy elimination algorithm for underspecified descriptions
Alexander Koller | Stefan Thater
Proceedings of the Fifth International Workshop on Inference in Computational Semantics (ICoS-5)

2005

pdf bib
The Evolution of Dominance Constraint Solvers
Alexander Koller | Stefan Thater
Proceedings of Workshop on Software

pdf bib
Efficient Solving and Exploration of Scope Ambiguities
Alexander Koller | Stefan Thater
Proceedings of the ACL Interactive Poster and Demonstration Sessions

2004

pdf bib
Minimal Recursion Semantics as Dominance Constraints: Translation, Evaluation, and Analysis
Ruth Fuchss | Alexander Koller | Joachim Niehren | Stefan Thater
Proceedings of the 42nd Annual Meeting of the Association for Computational Linguistics (ACL-04)

pdf bib
Computing Locally Coherent Discourses
Ernst Althaus | Nikiforos Karamanis | Alexander Koller
Proceedings of the 42nd Annual Meeting of the Association for Computational Linguistics (ACL-04)

pdf bib
A Relational Syntax-Semantics Interface Based on Dependency Grammar
Ralph Debusmann | Denys Duchier | Alexander Koller | Marco Kuhlmann | Gert Smolka | Stefan Thater
COLING 2004: Proceedings of the 20th International Conference on Computational Linguistics

pdf bib
Talking robots with Lego MindStorms
Alexander Koller | Geert-Jan Kruijff
COLING 2004: Proceedings of the 20th International Conference on Computational Linguistics

2003

pdf bib
Underspecification formalisms: Hole semantics as dominance constraints
Alexander Koller | Joachim Niehren | Stefan Thater
10th Conference of the European Chapter of the Association for Computational Linguistics

2002

pdf bib
Natural Language and Inference in a Computer Game
Malte Gabsdil | Alexander Koller | Kristina Striegnitz
COLING 2002: The 19th International Conference on Computational Linguistics

pdf bib
Generation as Dependency Parsing
Alexander Koller | Kristina Striegnitz
Proceedings of the 40th Annual Meeting of the Association for Computational Linguistics

2001

pdf bib
Underspecified Beta Reduction
Manuel Bodirsky | Katrin Erk | Alexander Koller | Joachim Niehren
Proceedings of the 39th Annual Meeting of the Association for Computational Linguistics

2000

pdf bib
A Polynomial-Time Fragment of Dominance Constraints
Alexander Koller | Kurt Mehlhorn | Joachim Niehren
Proceedings of the 38th Annual Meeting of the Association for Computational Linguistics

pdf bib
On Underspecified Processing of Dynamic Semantics
Alexander Koller | Joachim Niehren
COLING 2000 Volume 1: The 18th International Conference on Computational Linguistics

Search
Co-authors