Dataset Viewer
Auto-converted to Parquet Duplicate
title
stringlengths
23
150
authors
stringlengths
12
456
abstract
stringlengths
262
1.75k
url
stringlengths
42
44
detail_url
stringclasses
1 value
abs
stringclasses
1 value
OpenReview
stringclasses
1 value
Download PDF
stringlengths
46
48
tags
stringclasses
1 value
Knowledge Router: Learning Disentangled Representations for Knowledge Graphs
Shuai Zhang, Xi Rao, Yi Tay, Ce Zhang
The design of expressive representations of entities and relations in a knowledge graph is an important endeavor. While many of the existing approaches have primarily focused on learning from relational patterns and structural information, the intrinsic complexity of KG entities has been more or less overlooked. More c...
https://aclanthology.org/2021.naacl-main.1
https://aclanthology.org/2021.naacl-main.1.pdf
NAACL 2021
Distantly Supervised Relation Extraction with Sentence Reconstruction and Knowledge Base Priors
Fenia Christopoulou, Makoto Miwa, Sophia Ananiadou
We propose a multi-task, probabilistic approach to facilitate distantly supervised relation extraction by bringing closer the representations of sentences that contain the same Knowledge Base pairs. To achieve this, we bias the latent space of sentences via a Variational Autoencoder (VAE) that is trained jointly with a...
https://aclanthology.org/2021.naacl-main.2
https://aclanthology.org/2021.naacl-main.2.pdf
NAACL 2021
Cross-Task Instance Representation Interactions and Label Dependencies for Joint Information Extraction with Graph Convolutional Networks
Minh Van Nguyen, Viet Dac Lai, Thien Huu Nguyen
Existing works on information extraction (IE) have mainly solved the four main tasks separately (entity mention recognition, relation extraction, event trigger detection, and argument extraction), thus failing to benefit from inter-dependencies between tasks. This paper presents a novel deep learning model to simultane...
https://aclanthology.org/2021.naacl-main.3
https://aclanthology.org/2021.naacl-main.3.pdf
NAACL 2021
Abstract Meaning Representation Guided Graph Encoding and Decoding for Joint Information Extraction
Zixuan Zhang, Heng Ji
The tasks of Rich Semantic Parsing, such as Abstract Meaning Representation (AMR), share similar goals with Information Extraction (IE) to convert natural language texts into structured semantic representations. To take advantage of such similarity, we propose a novel AMR-guided framework for joint information extracti...
https://aclanthology.org/2021.naacl-main.4
https://aclanthology.org/2021.naacl-main.4.pdf
NAACL 2021
A Frustratingly Easy Approach for Entity and Relation Extraction
Zexuan Zhong, Danqi Chen
End-to-end relation extraction aims to identify named entities and extract relations between them. Most recent work models these two subtasks jointly, either by casting them in one structured prediction framework, or performing multi-task learning through shared representations. In this work, we present a simple pipeli...
https://aclanthology.org/2021.naacl-main.5
https://aclanthology.org/2021.naacl-main.5.pdf
NAACL 2021
Event Time Extraction and Propagation via Graph Attention Networks
Haoyang Wen, Yanru Qu, Heng Ji, Qiang Ning, Jiawei Han, Avi Sil, Hanghang Tong, Dan Roth
Grounding events into a precise timeline is important for natural language understanding but has received limited attention in recent work. This problem is challenging due to the inherent ambiguity of language and the requirement for information propagation over inter-related events. This paper first formulates this pr...
https://aclanthology.org/2021.naacl-main.6
https://aclanthology.org/2021.naacl-main.6.pdf
NAACL 2021
Probing Word Translations in the Transformer and Trading Decoder for Encoder Layers
Hongfei Xu, Josef van Genabith, Qiuhui Liu, Deyi Xiong
Due to its effectiveness and performance, the Transformer translation model has attracted wide attention, most recently in terms of probing-based approaches. Previous work focuses on using or probing source linguistic features in the encoder. To date, the way word translation evolves in Transformer layers has not yet b...
https://aclanthology.org/2021.naacl-main.7
https://aclanthology.org/2021.naacl-main.7.pdf
NAACL 2021
Mediators in Determining what Processing BERT Performs First
Aviv Slobodkin, Leshem Choshen, Omri Abend
Probing neural models for the ability to perform downstream tasks using their activation patterns is often used to localize what parts of the network specialize in performing what tasks. However, little work addressed potential mediating factors in such comparisons. As a test-case mediating factor, we consider the pred...
https://aclanthology.org/2021.naacl-main.8
https://aclanthology.org/2021.naacl-main.8.pdf
NAACL 2021
Automatic Generation of Contrast Sets from Scene Graphs: Probing the Compositional Consistency of GQA
Yonatan Bitton, Gabriel Stanovsky, Roy Schwartz, Michael Elhadad
Recent works have shown that supervised models often exploit data artifacts to achieve good test scores while their performance severely degrades on samples outside their training distribution. Contrast sets (Gardneret al., 2020) quantify this phenomenon by perturbing test samples in a minimal way such that the output ...
https://aclanthology.org/2021.naacl-main.9
https://aclanthology.org/2021.naacl-main.9.pdf
NAACL 2021
Multilingual Language Models Predict Human Reading Behavior
Nora Hollenstein, Federico Pirovano, Ce Zhang, Lena Jäger, Lisa Beinborn
We analyze if large language models are able to predict patterns of human reading behavior. We compare the performance of language-specific and multilingual pretrained transformer models to predict reading time measures reflecting natural human sentence processing on Dutch, English, German, and Russian texts. This resu...
https://aclanthology.org/2021.naacl-main.10
https://aclanthology.org/2021.naacl-main.10.pdf
NAACL 2021
Do Syntactic Probes Probe Syntax? Experiments with Jabberwocky Probing
Rowan Hall Maudslay, Ryan Cotterell
Analysing whether neural language models encode linguistic information has become popular in NLP. One method of doing so, which is frequently cited to support the claim that models like BERT encode syntax, is called probing; probes are small supervised models trained to extract linguistic information from another model...
https://aclanthology.org/2021.naacl-main.11
https://aclanthology.org/2021.naacl-main.11.pdf
NAACL 2021
A Non-Linear Structural Probe
Jennifer C. White, Tiago Pimentel, Naomi Saphra, Ryan Cotterell
Probes are models devised to investigate the encoding of knowledge—e.g. syntactic structure—in contextual representations. Probes are often designed for simplicity, which has led to restrictions on probe design that may not allow for the full exploitation of the structure of encoded information; one such restriction is...
https://aclanthology.org/2021.naacl-main.12
https://aclanthology.org/2021.naacl-main.12.pdf
NAACL 2021
Concealed Data Poisoning Attacks on NLP Models
Eric Wallace, Tony Zhao, Shi Feng, Sameer Singh
Adversarial attacks alter NLP model predictions by perturbing test-time inputs. However, it is much less understood whether, and how, predictions can be manipulated with small, concealed changes to the training data. In this work, we develop a new data poisoning attack that allows an adversary to control model predicti...
https://aclanthology.org/2021.naacl-main.13
https://aclanthology.org/2021.naacl-main.13.pdf
NAACL 2021
Backtranslation Feedback Improves User Confidence in MT, Not Quality
Vilém Zouhar, Michal Novák, Matúš Žilinec, Ondřej Bojar, Mateo Obregón, Robin L. Hill, Frédéric Blain, Marina Fomicheva, Lucia Specia, Lisa Yankovskaya
Translating text into a language unknown to the text’s author, dubbed outbound translation, is a modern need for which the user experience has significant room for improvement, beyond the basic machine translation facility. We demonstrate this by showing three ways in which user confidence in the outbound translation, ...
https://aclanthology.org/2021.naacl-main.14
https://aclanthology.org/2021.naacl-main.14.pdf
NAACL 2021
Data Filtering using Cross-Lingual Word Embeddings
Christian Herold, Jan Rosendahl, Joris Vanvinckenroye, Hermann Ney
Data filtering for machine translation (MT) describes the task of selecting a subset of a given, possibly noisy corpus with the aim to maximize the performance of an MT system trained on this selected data. Over the years, many different filtering approaches have been proposed. However, varying task definitions and dat...
https://aclanthology.org/2021.naacl-main.15
https://aclanthology.org/2021.naacl-main.15.pdf
NAACL 2021
Improving the Lexical Ability of Pretrained Language Models for Unsupervised Neural Machine Translation
Alexandra Chronopoulou, Dario Stojanovski, Alexander Fraser
Successful methods for unsupervised neural machine translation (UNMT) employ cross-lingual pretraining via self-supervision, often in the form of a masked language modeling or a sequence generation task, which requires the model to align the lexical- and high-level representations of the two languages. While cross-ling...
https://aclanthology.org/2021.naacl-main.16
https://aclanthology.org/2021.naacl-main.16.pdf
NAACL 2021
Neural Machine Translation without Embeddings
Uri Shaham, Omer Levy
Many NLP models operate over sequences of subword tokens produced by hand-crafted tokenization rules and heuristic subword induction algorithms. A simple universal alternative is to represent every computerized text as a sequence of bytes via UTF-8, obviating the need for an embedding layer since there are fewer token ...
https://aclanthology.org/2021.naacl-main.17
https://aclanthology.org/2021.naacl-main.17.pdf
NAACL 2021
Counterfactual Data Augmentation for Neural Machine Translation
Qi Liu, Matt Kusner, Phil Blunsom
We propose a data augmentation method for neural machine translation. It works by interpreting language models and phrasal alignment causally. Specifically, it creates augmented parallel translation corpora by generating (path-specific) counterfactual aligned phrases. We generate these by sampling new source phrases fr...
https://aclanthology.org/2021.naacl-main.18
https://aclanthology.org/2021.naacl-main.18.pdf
NAACL 2021
Cultural and Geographical Influences on Image Translatability of Words across Languages
Nikzad Khani, Isidora Tourni, Mohammad Sadegh Rasooli, Chris Callison-Burch, Derry Tanti Wijaya
Neural Machine Translation (NMT) models have been observed to produce poor translations when there are few/no parallel sentences to train the models. In the absence of parallel data, several approaches have turned to the use of images to learn translations. Since images of words, e.g., horse may be unchanged across lan...
https://aclanthology.org/2021.naacl-main.19
https://aclanthology.org/2021.naacl-main.19.pdf
NAACL 2021
Multilingual BERT Post-Pretraining Alignment
Lin Pan, Chung-Wei Hang, Haode Qi, Abhishek Shah, Saloni Potdar, Mo Yu
We propose a simple method to align multilingual contextual embeddings as a post-pretraining step for improved cross-lingual transferability of the pretrained language models. Using parallel data, our method aligns embeddings on the word level through the recently proposed Translation Language Modeling objective as wel...
https://aclanthology.org/2021.naacl-main.20
https://aclanthology.org/2021.naacl-main.20.pdf
NAACL 2021
A Million Tweets Are Worth a Few Points: Tuning Transformers for Customer Service Tasks
Amir Hadifar, Sofie Labat, Veronique Hoste, Chris Develder, Thomas Demeester
In online domain-specific customer service applications, many companies struggle to deploy advanced NLP models successfully, due to the limited availability of and noise in their datasets. While prior research demonstrated the potential of migrating large open-domain pretrained models for domain-specific tasks, the app...
https://aclanthology.org/2021.naacl-main.21
https://aclanthology.org/2021.naacl-main.21.pdf
NAACL 2021
Paragraph-level Rationale Extraction through Regularization: A case study on European Court of Human Rights Cases
Ilias Chalkidis, Manos Fergadiotis, Dimitrios Tsarapatsanis, Nikolaos Aletras, Ion Androutsopoulos, Prodromos Malakasiotis
Interpretability or explainability is an emerging research field in NLP. From a user-centric point of view, the goal is to build models that provide proper justification for their decisions, similar to those of humans, by requiring the models to satisfy additional constraints. To this end, we introduce a new applicatio...
https://aclanthology.org/2021.naacl-main.22
https://aclanthology.org/2021.naacl-main.22.pdf
NAACL 2021
Answering Product-Questions by Utilizing Questions from Other Contextually Similar Products
Ohad Rozen, David Carmel, Avihai Mejer, Vitaly Mirkis, Yftah Ziser
Predicting the answer to a product-related question is an emerging field of research that recently attracted a lot of attention. Answering subjective and opinion-based questions is most challenging due to the dependency on customer generated content. Previous works mostly focused on review-aware answer prediction; howe...
https://aclanthology.org/2021.naacl-main.23
https://aclanthology.org/2021.naacl-main.23.pdf
NAACL 2021
EnSidNet: Enhanced Hybrid Siamese-Deep Network for grouping clinical trials into drug-development pathways
Lucia Pagani
Siamese Neural Networks have been widely used to perform similarity classification in multi-class settings. Their architecture can be used to group the clinical trials belonging to the same drug-development pathway along the several clinical trial phases. Here we present an approach for the unmet need of drug-developme...
https://aclanthology.org/2021.naacl-main.24
https://aclanthology.org/2021.naacl-main.24.pdf
NAACL 2021
DATE: Detecting Anomalies in Text via Self-Supervision of Transformers
Andrei Manolache, Florin Brad, Elena Burceanu
Leveraging deep learning models for Anomaly Detection (AD) has seen widespread use in recent years due to superior performances over traditional methods. Recent deep methods for anomalies in images learn better features of normality in an end-to-end self-supervised setting. These methods train a model to discriminate b...
https://aclanthology.org/2021.naacl-main.25
https://aclanthology.org/2021.naacl-main.25.pdf
NAACL 2021
A Simple Approach for Handling Out-of-Vocabulary Identifiers in Deep Learning for Source Code
Nadezhda Chirkova, Sergey Troshin
There is an emerging interest in the application of natural language processing models to source code processing tasks. One of the major problems in applying deep learning to software engineering is that source code often contains a lot of rare identifiers, resulting in huge vocabularies. We propose a simple, yet effec...
https://aclanthology.org/2021.naacl-main.26
https://aclanthology.org/2021.naacl-main.26.pdf
NAACL 2021
Fast and Scalable Dialogue State Tracking with Explicit Modular Decomposition
Dingmin Wang, Chenghua Lin, Qi Liu, Kam-Fai Wong
We present a fast and scalable architecture called Explicit Modular Decomposition (EMD), in which we incorporate both classification-based and extraction-based methods and design four modules (for clas- sification and sequence labelling) to jointly extract dialogue states. Experimental results based on the MultiWoz 2.0...
https://aclanthology.org/2021.naacl-main.27
https://aclanthology.org/2021.naacl-main.27.pdf
NAACL 2021
Augmented SBERT: Data Augmentation Method for Improving Bi-Encoders for Pairwise Sentence Scoring Tasks
Nandan Thakur, Nils Reimers, Johannes Daxenberger, Iryna Gurevych
There are two approaches for pairwise sentence scoring: Cross-encoders, which perform full-attention over the input pair, and Bi-encoders, which map each input independently to a dense vector space. While cross-encoders often achieve higher performance, they are too slow for many practical use cases. Bi-encoders, on th...
https://aclanthology.org/2021.naacl-main.28
https://aclanthology.org/2021.naacl-main.28.pdf
NAACL 2021
SmBoP: Semi-autoregressive Bottom-up Semantic Parsing
Ohad Rubin, Jonathan Berant
The de-facto standard decoding method for semantic parsing in recent years has been to autoregressively decode the abstract syntax tree of the target program using a top-down depth-first traversal. In this work, we propose an alternative approach: a Semi-autoregressive Bottom-up Parser (SmBoP) that constructs at decodi...
https://aclanthology.org/2021.naacl-main.29
https://aclanthology.org/2021.naacl-main.29.pdf
NAACL 2021
SGL: Speaking the Graph Languages of Semantic Parsing via Multilingual Translation
Luigi Procopio, Rocco Tripodi, Roberto Navigli
{'url': 'https://github.com/SapienzaNLP/sgl', '#text': 'Graph-based semantic parsing aims to represent textual meaning through directed graphs. As one of the most promising general-purpose meaning representations, these structures and their parsing have gained a significant interest momentum during recent years, with s...
https://aclanthology.org/2021.naacl-main.30
https://aclanthology.org/2021.naacl-main.30.pdf
NAACL 2021
Unifying Cross-Lingual Semantic Role Labeling with Heterogeneous Linguistic Resources
Simone Conia, Andrea Bacciu, Roberto Navigli
{'url': 'https://github.com/SapienzaNLP/unify-srl', '#text': 'While cross-lingual techniques are finding increasing success in a wide range of Natural Language Processing tasks, their application to Semantic Role Labeling (SRL) has been strongly limited by the fact that each language adopts its own linguistic formalism...
https://aclanthology.org/2021.naacl-main.31
https://aclanthology.org/2021.naacl-main.31.pdf
NAACL 2021
Fool Me Twice: Entailment from Wikipedia Gamification
Julian Eisenschlos, Bhuwan Dhingra, Jannis Bulian, Benjamin Börschinger, Jordan Boyd-Graber
We release FoolMeTwice (FM2 for short), a large dataset of challenging entailment pairs collected through a fun multi-player game. Gamification encourages adversarial examples, drastically lowering the number of examples that can be solved using “shortcuts” compared to other popular entailment datasets. Players are pre...
https://aclanthology.org/2021.naacl-main.32
https://aclanthology.org/2021.naacl-main.32.pdf
NAACL 2021
Meta-Learning for Domain Generalization in Semantic Parsing
Bailin Wang, Mirella Lapata, Ivan Titov
The importance of building semantic parsers which can be applied to new domains and generate programs unseen at training has long been acknowledged, and datasets testing out-of-domain performance are becoming increasingly available. However, little or no attention has been devoted to learning algorithms or objectives w...
https://aclanthology.org/2021.naacl-main.33
https://aclanthology.org/2021.naacl-main.33.pdf
NAACL 2021
Aspect-Controlled Neural Argument Generation
Benjamin Schiller, Johannes Daxenberger, Iryna Gurevych
We rely on arguments in our daily lives to deliver our opinions and base them on evidence, making them more convincing in turn. However, finding and formulating arguments can be challenging. In this work, we present the Arg-CTRL - a language model for argument generation that can be controlled to generate sentence-leve...
https://aclanthology.org/2021.naacl-main.34
https://aclanthology.org/2021.naacl-main.34.pdf
NAACL 2021
Text Generation from Discourse Representation Structures
Jiangming Liu, Shay B. Cohen, Mirella Lapata
We propose neural models to generate text from formal meaning representations based on Discourse Representation Structures (DRSs). DRSs are document-level representations which encode rich semantic detail pertaining to rhetorical relations, presupposition, and co-reference within and across sentences. We formalize the ...
https://aclanthology.org/2021.naacl-main.35
https://aclanthology.org/2021.naacl-main.35.pdf
NAACL 2021
APo-VAE: Text Generation in Hyperbolic Space
Shuyang Dai, Zhe Gan, Yu Cheng, Chenyang Tao, Lawrence Carin, Jingjing Liu
Natural language often exhibits inherent hierarchical structure ingrained with complex syntax and semantics. However, most state-of-the-art deep generative models learn embeddings only in Euclidean vector space, without accounting for this structural property of language. In this paper, we investigate text generation i...
https://aclanthology.org/2021.naacl-main.36
https://aclanthology.org/2021.naacl-main.36.pdf
NAACL 2021
DART: Open-Domain Structured Data Record to Text Generation
Linyong Nan, Dragomir Radev, Rui Zhang, Amrit Rau, Abhinand Sivaprasad, Chiachun Hsieh, Xiangru Tang, Aadit Vyas, Neha Verma, Pranav Krishna, Yangxiaokang Liu, Nadia Irwanto, Jessica Pan, Faiaz Rahman, Ahmad Zaidi, Mutethia Mutuma, Yasin Tarabar, Ankit Gupta, Tao Yu, Yi Chern Tan, Xi Victoria Lin, Caiming Xiong, Richar...
{'url': 'https://github.com/Yale-LILY/dart', '#text': 'We present DART, an open domain structured DAta Record to Text generation dataset with over 82k instances (DARTs). Data-to-text annotations can be a costly process, especially when dealing with tables which are the major source of structured data and contain nontri...
https://aclanthology.org/2021.naacl-main.37
https://aclanthology.org/2021.naacl-main.37.pdf
NAACL 2021
When Being Unseen from mBERT is just the Beginning: Handling New Languages With Multilingual Language Models
Benjamin Muller, Antonios Anastasopoulos, Benoît Sagot, Djamé Seddah
Transfer learning based on pretraining language models on a large amount of raw data has become a new norm to reach state-of-the-art performance in NLP. Still, it remains unclear how this approach should be applied for unseen languages that are not covered by any available large-scale multilingual language model and fo...
https://aclanthology.org/2021.naacl-main.38
https://aclanthology.org/2021.naacl-main.38.pdf
NAACL 2021
Multi-Adversarial Learning for Cross-Lingual Word Embeddings
Haozhou Wang, James Henderson, Paola Merlo
Generative adversarial networks (GANs) have succeeded in inducing cross-lingual word embeddings - maps of matching words across languages - without supervision. Despite these successes, GANs’ performance for the difficult case of distant languages is still not satisfactory. These limitations have been explained by GANs...
https://aclanthology.org/2021.naacl-main.39
https://aclanthology.org/2021.naacl-main.39.pdf
NAACL 2021
Multi-view Subword Regularization
Xinyi Wang, Sebastian Ruder, Graham Neubig
Multilingual pretrained representations generally rely on subword segmentation algorithms to create a shared multilingual vocabulary. However, standard heuristic algorithms often lead to sub-optimal segmentation, especially for languages with limited amounts of data. In this paper, we take two major steps towards allev...
https://aclanthology.org/2021.naacl-main.40
https://aclanthology.org/2021.naacl-main.40.pdf
NAACL 2021
mT5: A Massively Multilingual Pre-trained Text-to-Text Transformer
Linting Xue, Noah Constant, Adam Roberts, Mihir Kale, Rami Al-Rfou, Aditya Siddhant, Aditya Barua, Colin Raffel
The recent “Text-to-Text Transfer Transformer” (T5) leveraged a unified text-to-text format and scale to attain state-of-the-art results on a wide variety of English-language NLP tasks. In this paper, we introduce mT5, a multilingual variant of T5 that was pre-trained on a new Common Crawl-based dataset covering 101 la...
https://aclanthology.org/2021.naacl-main.41
https://aclanthology.org/2021.naacl-main.41.pdf
NAACL 2021
MetaXL: Meta Representation Transformation for Low-resource Cross-lingual Learning
Mengzhou Xia, Guoqing Zheng, Subhabrata Mukherjee, Milad Shokouhi, Graham Neubig, Ahmed Hassan Awadallah
The combination of multilingual pre-trained representations and cross-lingual transfer learning is one of the most effective methods for building functional NLP systems for low-resource languages. However, for extremely low-resource languages without large-scale monolingual corpora for pre-training or sufficient annota...
https://aclanthology.org/2021.naacl-main.42
https://aclanthology.org/2021.naacl-main.42.pdf
NAACL 2021
Open Domain Question Answering over Tables via Dense Retrieval
Jonathan Herzig, Thomas Müller, Syrine Krichene, Julian Eisenschlos
Recent advances in open-domain QA have led to strong models based on dense retrieval, but only focused on retrieving textual passages. In this work, we tackle open-domain QA over tables for the first time, and show that retrieval can be improved by a retriever designed to handle tabular context. We present an effective...
https://aclanthology.org/2021.naacl-main.43
https://aclanthology.org/2021.naacl-main.43.pdf
NAACL 2021
Open-Domain Question Answering Goes Conversational via Question Rewriting
Raviteja Anantha, Svitlana Vakulenko, Zhucheng Tu, Shayne Longpre, Stephen Pulman, Srinivas Chappidi
We introduce a new dataset for Question Rewriting in Conversational Context (QReCC), which contains 14K conversations with 80K question-answer pairs. The task in QReCC is to find answers to conversational questions within a collection of 10M web pages (split into 54M passages). Answers to questions in the same conversa...
https://aclanthology.org/2021.naacl-main.44
https://aclanthology.org/2021.naacl-main.44.pdf
NAACL 2021
QA-GNN: Reasoning with Language Models and Knowledge Graphs for Question Answering
Michihiro Yasunaga, Hongyu Ren, Antoine Bosselut, Percy Liang, Jure Leskovec
The problem of answering questions using knowledge from pre-trained language models (LMs) and knowledge graphs (KGs) presents two challenges: given a QA context (question and answer choice), methods need to (i) identify relevant knowledge from large KGs, and (ii) perform joint reasoning over the QA context and KG. Here...
https://aclanthology.org/2021.naacl-main.45
https://aclanthology.org/2021.naacl-main.45.pdf
NAACL 2021
XOR QA: Cross-lingual Open-Retrieval Question Answering
Akari Asai, Jungo Kasai, Jonathan Clark, Kenton Lee, Eunsol Choi, Hannaneh Hajishirzi
{'url': 'https://nlp.cs.washington.edu/xorqa/', '#text': 'Multilingual question answering tasks typically assume that answers exist in the same language as the question. Yet in practice, many languages face both information scarcity—where languages have few reference articles—and information asymmetry—where questions r...
https://aclanthology.org/2021.naacl-main.46
https://aclanthology.org/2021.naacl-main.46.pdf
NAACL 2021
SPARTA: Efficient Open-Domain Question Answering via Sparse Transformer Matching Retrieval
Tiancheng Zhao, Xiaopeng Lu, Kyusong Lee
We introduce SPARTA, a novel neural retrieval method that shows great promise in performance, generalization, and interpretability for open-domain question answering. Unlike many neural ranking methods that use dense vector nearest neighbor search, SPARTA learns a sparse representation that can be efficiently implement...
https://aclanthology.org/2021.naacl-main.47
https://aclanthology.org/2021.naacl-main.47.pdf
NAACL 2021
Implicitly Abusive Language – What does it actually look like and why are we not getting there?
Michael Wiegand, Josef Ruppenhofer, Elisabeth Eder
Abusive language detection is an emerging field in natural language processing which has received a large amount of attention recently. Still the success of automatic detection is limited. Particularly, the detection of implicitly abusive language, i.e. abusive language that is not conveyed by abusive words (e.g. dumba...
https://aclanthology.org/2021.naacl-main.48
https://aclanthology.org/2021.naacl-main.48.pdf
NAACL 2021
The Importance of Modeling Social Factors of Language: Theory and Practice
Dirk Hovy, Diyi Yang
Natural language processing (NLP) applications are now more powerful and ubiquitous than ever before. With rapidly developing (neural) models and ever-more available data, current NLP models have access to more information than any human speaker during their life. Still, it would be hard to argue that NLP models have r...
https://aclanthology.org/2021.naacl-main.49
https://aclanthology.org/2021.naacl-main.49.pdf
NAACL 2021
On learning and representing social meaning in NLP: a sociolinguistic perspective
Dong Nguyen, Laura Rosseel, Jack Grieve
The field of NLP has made substantial progress in building meaning representations. However, an important aspect of linguistic meaning, social meaning, has been largely overlooked. We introduce the concept of social meaning to NLP and discuss how insights from sociolinguistics can inform work on representation learning...
https://aclanthology.org/2021.naacl-main.50
https://aclanthology.org/2021.naacl-main.50.pdf
NAACL 2021
Preregistering NLP research
Emiel van Miltenburg, Chris van der Lee, Emiel Krahmer
Preregistration refers to the practice of specifying what you are going to do, and what you expect to find in your study, before carrying out the study. This practice is increasingly common in medicine and psychology, but is rarely discussed in NLP. This paper discusses preregistration in more detail, explores how NLP ...
https://aclanthology.org/2021.naacl-main.51
https://aclanthology.org/2021.naacl-main.51.pdf
NAACL 2021
Get Your Vitamin C! Robust Fact Verification with Contrastive Evidence
Tal Schuster, Adam Fisch, Regina Barzilay
Typical fact verification models use retrieved written evidence to verify claims. Evidence sources, however, often change over time as more information is gathered and revised. In order to adapt, models must be sensitive to subtle differences in supporting evidence. We present VitaminC, a benchmark infused with challen...
https://aclanthology.org/2021.naacl-main.52
https://aclanthology.org/2021.naacl-main.52.pdf
NAACL 2021
Representing Numbers in NLP: a Survey and a Vision
Avijit Thawani, Jay Pujara, Filip Ilievski, Pedro Szekely
NLP systems rarely give special consideration to numbers found in text. This starkly contrasts with the consensus in neuroscience that, in the brain, numbers are represented differently from words. We arrange recent NLP work on numeracy into a comprehensive taxonomy of tasks and methods. We break down the subjective no...
https://aclanthology.org/2021.naacl-main.53
https://aclanthology.org/2021.naacl-main.53.pdf
NAACL 2021
Extending Multi-Document Summarization Evaluation to the Interactive Setting
Ori Shapira, Ramakanth Pasunuru, Hadar Ronen, Mohit Bansal, Yael Amsterdamer, Ido Dagan
Allowing users to interact with multi-document summarizers is a promising direction towards improving and customizing summary results. Different ideas for interactive summarization have been proposed in previous work but these solutions are highly divergent and incomparable. In this paper, we develop an end-to-end eval...
https://aclanthology.org/2021.naacl-main.54
https://aclanthology.org/2021.naacl-main.54.pdf
NAACL 2021
Identifying Helpful Sentences in Product Reviews
Iftah Gamzu, Hila Gonen, Gilad Kutiel, Ran Levy, Eugene Agichtein
In recent years online shopping has gained momentum and became an important venue for customers wishing to save time and simplify their shopping process. A key advantage of shopping online is the ability to read what other customers are saying about products of interest. In this work, we aim to maintain this advantage ...
https://aclanthology.org/2021.naacl-main.55
https://aclanthology.org/2021.naacl-main.55.pdf
NAACL 2021
Noisy Self-Knowledge Distillation for Text Summarization
Yang Liu, Sheng Shen, Mirella Lapata
In this paper we apply self-knowledge distillation to text summarization which we argue can alleviate problems with maximum-likelihood training on single reference and noisy datasets. Instead of relying on one-hot annotation labels, our student summarization model is trained with guidance from a teacher which generates...
https://aclanthology.org/2021.naacl-main.56
https://aclanthology.org/2021.naacl-main.56.pdf
NAACL 2021
Improving Zero and Few-Shot Abstractive Summarization with Intermediate Fine-tuning and Data Augmentation
Alexander Fabbri, Simeng Han, Haoyuan Li, Haoran Li, Marjan Ghazvininejad, Shafiq Joty, Dragomir Radev, Yashar Mehdad
Models pretrained with self-supervised objectives on large text corpora achieve state-of-the-art performance on English text summarization tasks. However, these models are typically fine-tuned on hundreds of thousands of data points, an infeasible requirement when applying summarization to new, niche domains. In this w...
https://aclanthology.org/2021.naacl-main.57
https://aclanthology.org/2021.naacl-main.57.pdf
NAACL 2021
Enhancing Factual Consistency of Abstractive Summarization
Chenguang Zhu, William Hinthorn, Ruochen Xu, Qingkai Zeng, Michael Zeng, Xuedong Huang, Meng Jiang
Automatic abstractive summaries are found to often distort or fabricate facts in the article. This inconsistency between summary and original text has seriously impacted its applicability. We propose a fact-aware summarization model FASum to extract and integrate factual relations into the summary generation process vi...
https://aclanthology.org/2021.naacl-main.58
https://aclanthology.org/2021.naacl-main.58.pdf
NAACL 2021
Few-shot Intent Classification and Slot Filling with Retrieved Examples
Dian Yu, Luheng He, Yuan Zhang, Xinya Du, Panupong Pasupat, Qi Li
Few-shot learning arises in important practical scenarios, such as when a natural language understanding system needs to learn new semantic labels for an emerging, resource-scarce domain. In this paper, we explore retrieval-based methods for intent classification and slot filling tasks in few-shot settings. Retrieval-b...
https://aclanthology.org/2021.naacl-main.59
https://aclanthology.org/2021.naacl-main.59.pdf
NAACL 2021
“Nice Try, Kiddo”: Investigating Ad Hominems in Dialogue Responses
Emily Sheng, Kai-Wei Chang, Prem Natarajan, Nanyun Peng
Ad hominem attacks are those that target some feature of a person’s character instead of the position the person is maintaining. These attacks are harmful because they propagate implicit biases and diminish a person’s credibility. Since dialogue systems respond directly to user input, it is important to study ad homine...
https://aclanthology.org/2021.naacl-main.60
https://aclanthology.org/2021.naacl-main.60.pdf
NAACL 2021
Human-like informative conversations: Better acknowledgements using conditional mutual information
Ashwin Paranjape, Christopher Manning
This work aims to build a dialogue agent that can weave new factual content into conversations as naturally as humans. We draw insights from linguistic principles of conversational analysis and annotate human-human conversations from the Switchboard Dialog Act Corpus to examine humans strategies for acknowledgement, tr...
https://aclanthology.org/2021.naacl-main.61
https://aclanthology.org/2021.naacl-main.61.pdf
NAACL 2021
A Comparative Study on Schema-Guided Dialogue State Tracking
Jie Cao, Yi Zhang
Frame-based state representation is widely used in modern task-oriented dialog systems to model user intentions and slot values. However, a fixed design of domain ontology makes it difficult to extend to new services and APIs. Recent work proposed to use natural language descriptions to define the domain ontology inste...
https://aclanthology.org/2021.naacl-main.62
https://aclanthology.org/2021.naacl-main.62.pdf
NAACL 2021
Spoken Language Understanding for Task-oriented Dialogue Systems with Augmented Memory Networks
Jie Wu, Ian Harris, Hongzhi Zhao
Spoken language understanding, usually including intent detection and slot filling, is a core component to build a spoken dialog system. Recent research shows promising results by jointly learning of those two tasks based on the fact that slot filling and intent detection are sharing semantic knowledge. Furthermore, at...
https://aclanthology.org/2021.naacl-main.63
https://aclanthology.org/2021.naacl-main.63.pdf
NAACL 2021
How to Motivate Your Dragon: Teaching Goal-Driven Agents to Speak and Act in Fantasy Worlds
Prithviraj Ammanabrolu, Jack Urbanek, Margaret Li, Arthur Szlam, Tim Rocktäschel, Jason Weston
We seek to create agents that both act and communicate with other agents in pursuit of a goal. Towards this end, we extend LIGHT (Urbanek et al. 2019)—a large-scale crowd-sourced fantasy text-game—with a dataset of quests. These contain natural language motivations paired with in-game goals and human demonstrations; co...
https://aclanthology.org/2021.naacl-main.64
https://aclanthology.org/2021.naacl-main.64.pdf
NAACL 2021
Linking Entities to Unseen Knowledge Bases with Arbitrary Schemas
Yogarshi Vyas, Miguel Ballesteros
In entity linking, mentions of named entities in raw text are disambiguated against a knowledge base (KB). This work focuses on linking to unseen KBs that do not have training data and whose schema is unknown during training. Our approach relies on methods to flexibly convert entities with several attribute-value pairs...
https://aclanthology.org/2021.naacl-main.65
https://aclanthology.org/2021.naacl-main.65.pdf
NAACL 2021
Self-Training with Weak Supervision
Giannis Karamanolakis, Subhabrata Mukherjee, Guoqing Zheng, Ahmed Hassan Awadallah
State-of-the-art deep neural networks require large-scale labeled training data that is often expensive to obtain or not available for many tasks. Weak supervision in the form of domain-specific rules has been shown to be useful in such settings to automatically generate weakly labeled training data. However, learning ...
https://aclanthology.org/2021.naacl-main.66
https://aclanthology.org/2021.naacl-main.66.pdf
NAACL 2021
Neural Language Modeling for Contextualized Temporal Graph Generation
Aman Madaan, Yiming Yang
This paper presents the first study on using large-scale pre-trained language models for automated generation of an event-level temporal graph for a document. Despite the huge success of neural pre-training methods in NLP tasks, its potential for temporal reasoning over event graphs has not been sufficiently explored. ...
https://aclanthology.org/2021.naacl-main.67
https://aclanthology.org/2021.naacl-main.67.pdf
NAACL 2021
Probabilistic Box Embeddings for Uncertain Knowledge Graph Reasoning
Xuelu Chen, Michael Boratko, Muhao Chen, Shib Sankar Dasgupta, Xiang Lorraine Li, Andrew McCallum
Knowledge bases often consist of facts which are harvested from a variety of sources, many of which are noisy and some of which conflict, resulting in a level of uncertainty for each triple. Knowledge bases are also often incomplete, prompting the use of embedding methods to generalize from known facts, however, existi...
https://aclanthology.org/2021.naacl-main.68
https://aclanthology.org/2021.naacl-main.68.pdf
NAACL 2021
Document-Level Event Argument Extraction by Conditional Generation
Sha Li, Heng Ji, Jiawei Han
Event extraction has long been treated as a sentence-level task in the IE community. We argue that this setting does not match human informative seeking behavior and leads to incomplete and uninformative extraction results. We propose a document-level neural event argument extraction model by formulating the task as co...
https://aclanthology.org/2021.naacl-main.69
https://aclanthology.org/2021.naacl-main.69.pdf
NAACL 2021
Template Filling with Generative Transformers
Xinya Du, Alexander Rush, Claire Cardie
Template filling is generally tackled by a pipeline of two separate supervised systems – one for role-filler extraction and another for template/event recognition. Since pipelines consider events in isolation, they can suffer from error propagation. We introduce a framework based on end-to-end generative transformers f...
https://aclanthology.org/2021.naacl-main.70
https://aclanthology.org/2021.naacl-main.70.pdf
NAACL 2021
Towards Interpreting and Mitigating Shortcut Learning Behavior of NLU models
Mengnan Du, Varun Manjunatha, Rajiv Jain, Ruchi Deshpande, Franck Dernoncourt, Jiuxiang Gu, Tong Sun, Xia Hu
Recent studies indicate that NLU models are prone to rely on shortcut features for prediction, without achieving true language understanding. As a result, these models fail to generalize to real-world out-of-distribution data. In this work, we show that the words in the NLU training set can be modeled as a long-tailed ...
https://aclanthology.org/2021.naacl-main.71
https://aclanthology.org/2021.naacl-main.71.pdf
NAACL 2021
On Attention Redundancy: A Comprehensive Study
Yuchen Bian, Jiaji Huang, Xingyu Cai, Jiahong Yuan, Kenneth Church
Multi-layer multi-head self-attention mechanism is widely applied in modern neural language models. Attention redundancy has been observed among attention heads but has not been deeply studied in the literature. Using BERT-base model as an example, this paper provides a comprehensive study on attention redundancy which...
https://aclanthology.org/2021.naacl-main.72
https://aclanthology.org/2021.naacl-main.72.pdf
NAACL 2021
Does BERT Pretrained on Clinical Notes Reveal Sensitive Data?
Eric Lehman, Sarthak Jain, Karl Pichotta, Yoav Goldberg, Byron Wallace
{'url': 'https://github.com/elehman16/exposing_patient_data_release', '#text': 'Large Transformers pretrained over clinical notes from Electronic Health Records (EHR) have afforded substantial gains in performance on predictive clinical tasks. The cost of training such models (and the necessity of data access to do so)...
https://aclanthology.org/2021.naacl-main.73
https://aclanthology.org/2021.naacl-main.73.pdf
NAACL 2021
Low-Complexity Probing via Finding Subnetworks
Steven Cao, Victor Sanh, Alexander Rush
The dominant approach in probing neural networks for linguistic properties is to train a new shallow multi-layer perceptron (MLP) on top of the model’s internal representations. This approach can detect properties encoded in the model, but at the cost of adding new parameters that may learn the task directly. We instea...
https://aclanthology.org/2021.naacl-main.74
https://aclanthology.org/2021.naacl-main.74.pdf
NAACL 2021
An Empirical Comparison of Instance Attribution Methods for NLP
Pouya Pezeshkpour, Sarthak Jain, Byron Wallace, Sameer Singh
{'url': 'https://github.com/successar/instance_attributions_NLP', '#text': 'Widespread adoption of deep models has motivated a pressing need for approaches to interpret network outputs and to facilitate model debugging. Instance attribution methods constitute one means of accomplishing these goals by retrieving trainin...
https://aclanthology.org/2021.naacl-main.75
https://aclanthology.org/2021.naacl-main.75.pdf
NAACL 2021
Generalization in Instruction Following Systems
Soham Dan, Michael Zhou, Dan Roth
Understanding and executing natural language instructions in a grounded domain is one of the hallmarks of artificial intelligence. In this paper, we focus on instruction understanding in the blocks world domain and investigate the language understanding abilities of two top-performing systems for the task. We aim to un...
https://aclanthology.org/2021.naacl-main.76
https://aclanthology.org/2021.naacl-main.76.pdf
NAACL 2021
LightningDOT: Pre-training Visual-Semantic Embeddings for Real-Time Image-Text Retrieval
Siqi Sun, Yen-Chun Chen, Linjie Li, Shuohang Wang, Yuwei Fang, Jingjing Liu
Multimodal pre-training has propelled great advancement in vision-and-language research. These large-scale pre-trained models, although successful, fatefully suffer from slow inference speed due to enormous computational cost mainly from cross-modal attention in Transformer architecture. When applied to real-life appli...
https://aclanthology.org/2021.naacl-main.77
https://aclanthology.org/2021.naacl-main.77.pdf
NAACL 2021
Measuring Social Biases in Grounded Vision and Language Embeddings
Candace Ross, Boris Katz, Andrei Barbu
We generalize the notion of measuring social biases in word embeddings to visually grounded word embeddings. Biases are present in grounded embeddings, and indeed seem to be equally or more significant than for ungrounded embeddings. This is despite the fact that vision and language can suffer from different biases, wh...
https://aclanthology.org/2021.naacl-main.78
https://aclanthology.org/2021.naacl-main.78.pdf
NAACL 2021
MTAG: Modal-Temporal Attention Graph for Unaligned Human Multimodal Language Sequences
Jianing Yang, Yongxin Wang, Ruitao Yi, Yuying Zhu, Azaan Rehman, Amir Zadeh, Soujanya Poria, Louis-Philippe Morency
Human communication is multimodal in nature; it is through multiple modalities such as language, voice, and facial expressions, that opinions and emotions are expressed. Data in this domain exhibits complex multi-relational and temporal interactions. Learning from this data is a fundamentally challenging research probl...
https://aclanthology.org/2021.naacl-main.79
https://aclanthology.org/2021.naacl-main.79.pdf
NAACL 2021
Grounding Open-Domain Instructions to Automate Web Support Tasks
Nancy Xu, Sam Masling, Michael Du, Giovanni Campagna, Larry Heck, James Landay, Monica Lam
Grounding natural language instructions on the web to perform previously unseen tasks enables accessibility and automation. We introduce a task and dataset to train AI agents from open-domain, step-by-step instructions originally written for people. We build RUSS (Rapid Universal Support Service) to tackle this problem...
https://aclanthology.org/2021.naacl-main.80
https://aclanthology.org/2021.naacl-main.80.pdf
NAACL 2021
Modular Networks for Compositional Instruction Following
Rodolfo Corona, Daniel Fried, Coline Devin, Dan Klein, Trevor Darrell
Standard architectures used in instruction following often struggle on novel compositions of subgoals (e.g. navigating to landmarks or picking up objects) observed during training. We propose a modular architecture for following natural language instructions that describe sequences of diverse subgoals. In our approach,...
https://aclanthology.org/2021.naacl-main.81
https://aclanthology.org/2021.naacl-main.81.pdf
NAACL 2021
Improving Cross-Modal Alignment in Vision Language Navigation via Syntactic Information
Jialu Li, Hao Tan, Mohit Bansal
Vision language navigation is the task that requires an agent to navigate through a 3D environment based on natural language instructions. One key challenge in this task is to ground instructions with the current visual information that the agent perceives. Most of the existing work employs soft attention over individu...
https://aclanthology.org/2021.naacl-main.82
https://aclanthology.org/2021.naacl-main.82.pdf
NAACL 2021
Improving Pretrained Models for Zero-shot Multi-label Text Classification through Reinforced Label Hierarchy Reasoning
Hui Liu, Danqing Zhang, Bing Yin, Xiaodan Zhu
Exploiting label hierarchies has become a promising approach to tackling the zero-shot multi-label text classification (ZS-MTC) problem. Conventional methods aim to learn a matching model between text and labels, using a graph encoder to incorporate label hierarchies to obtain effective label representations (Rios and ...
https://aclanthology.org/2021.naacl-main.83
https://aclanthology.org/2021.naacl-main.83.pdf
NAACL 2021
Fine-Tuning Pre-trained Language Model with Weak Supervision: A Contrastive-Regularized Self-Training Approach
Yue Yu, Simiao Zuo, Haoming Jiang, Wendi Ren, Tuo Zhao, Chao Zhang
{'url': 'https://github.com/yueyu1030/COSINE', '#text': 'Fine-tuned pre-trained language models (LMs) have achieved enormous success in many natural language processing (NLP) tasks, but they still require excessive labeled data in the fine-tuning stage. We study the problem of fine-tuning pre-trained LMs using only wea...
https://aclanthology.org/2021.naacl-main.84
https://aclanthology.org/2021.naacl-main.84.pdf
NAACL 2021
Posterior Differential Regularization with f-divergence for Improving Model Robustness
Hao Cheng, Xiaodong Liu, Lis Pereira, Yaoliang Yu, Jianfeng Gao
We address the problem of enhancing model robustness through regularization. Specifically, we focus on methods that regularize the model posterior difference between clean and noisy inputs. Theoretically, we provide a connection of two recent methods, Jacobian Regularization and Virtual Adversarial Training, under this...
https://aclanthology.org/2021.naacl-main.85
https://aclanthology.org/2021.naacl-main.85.pdf
NAACL 2021
Understanding Hard Negatives in Noise Contrastive Estimation
Wenzheng Zhang, Karl Stratos
The choice of negative examples is important in noise contrastive estimation. Recent works find that hard negatives—highest-scoring incorrect examples under the model—are effective in practice, but they are used without a formal justification. We develop analytical tools to understand the role of hard negatives. Specif...
https://aclanthology.org/2021.naacl-main.86
https://aclanthology.org/2021.naacl-main.86.pdf
NAACL 2021
Certified Robustness to Word Substitution Attack with Differential Privacy
Wenjie Wang, Pengfei Tang, Jian Lou, Li Xiong
The robustness and security of natural language processing (NLP) models are significantly important in real-world applications. In the context of text classification tasks, adversarial examples can be designed by substituting words with synonyms under certain semantic and syntactic constraints, such that a well-trained...
https://aclanthology.org/2021.naacl-main.87
https://aclanthology.org/2021.naacl-main.87.pdf
NAACL 2021
DReCa: A General Task Augmentation Strategy for Few-Shot Natural Language Inference
Shikhar Murty, Tatsunori B. Hashimoto, Christopher Manning
Meta-learning promises few-shot learners that can adapt to new distributions by repurposing knowledge acquired from previous training. However, we believe meta-learning has not yet succeeded in NLP due to the lack of a well-defined task distribution, leading to attempts that treat datasets as tasks. Such an ad hoc task...
https://aclanthology.org/2021.naacl-main.88
https://aclanthology.org/2021.naacl-main.88.pdf
NAACL 2021
Harnessing Multilinguality in Unsupervised Machine Translation for Rare Languages
Xavier Garcia, Aditya Siddhant, Orhan Firat, Ankur Parikh
Unsupervised translation has reached impressive performance on resource-rich language pairs such as English-French and English-German. However, early studies have shown that in more realistic settings involving low-resource, rare languages, unsupervised translation performs poorly, achieving less than 3.0 BLEU. In this...
https://aclanthology.org/2021.naacl-main.89
https://aclanthology.org/2021.naacl-main.89.pdf
NAACL 2021
Macro-Average: Rare Types Are Important Too
Thamme Gowda, Weiqiu You, Constantine Lignos, Jonathan May
While traditional corpus-level evaluation metrics for machine translation (MT) correlate well with fluency, they struggle to reflect adequacy. Model-based MT metrics trained on segment-level human judgments have emerged as an attractive replacement due to strong correlation results. These models, however, require poten...
https://aclanthology.org/2021.naacl-main.90
https://aclanthology.org/2021.naacl-main.90.pdf
NAACL 2021
Assessing Reference-Free Peer Evaluation for Machine Translation
Sweta Agrawal, George Foster, Markus Freitag, Colin Cherry
Reference-free evaluation has the potential to make machine translation evaluation substantially more scalable, allowing us to pivot easily to new languages or domains. It has been recently shown that the probabilities given by a large, multilingual model can achieve state of the art results when used as a reference-fr...
https://aclanthology.org/2021.naacl-main.91
https://aclanthology.org/2021.naacl-main.91.pdf
NAACL 2021
The Curious Case of Hallucinations in Neural Machine Translation
Vikas Raunak, Arul Menezes, Marcin Junczys-Dowmunt
In this work, we study hallucinations in Neural Machine Translation (NMT), which lie at an extreme end on the spectrum of NMT pathologies. Firstly, we connect the phenomenon of hallucinations under source perturbation to the Long-Tail theory of Feldman, and present an empirically validated hypothesis that explains hall...
https://aclanthology.org/2021.naacl-main.92
https://aclanthology.org/2021.naacl-main.92.pdf
NAACL 2021
Towards Continual Learning for Multilingual Machine Translation via Vocabulary Substitution
Xavier Garcia, Noah Constant, Ankur Parikh, Orhan Firat
We propose a straightforward vocabulary adaptation scheme to extend the language capacity of multilingual machine translation models, paving the way towards efficient continual learning for multilingual machine translation. Our approach is suitable for large-scale datasets, applies to distant languages with unseen scri...
https://aclanthology.org/2021.naacl-main.93
https://aclanthology.org/2021.naacl-main.93.pdf
NAACL 2021
Towards Modeling the Style of Translators in Neural Machine Translation
Yue Wang, Cuong Hoang, Marcello Federico
One key ingredient of neural machine translation is the use of large datasets from different domains and resources (e.g. Europarl, TED talks). These datasets contain documents translated by professional translators using different but consistent translation styles. Despite that, the model is usually trained in a way th...
https://aclanthology.org/2021.naacl-main.94
https://aclanthology.org/2021.naacl-main.94.pdf
NAACL 2021
Self-Supervised Test-Time Learning for Reading Comprehension
Pratyay Banerjee, Tejas Gokhale, Chitta Baral
{'i': 'context-question-answer', '#text': 'Recent work on unsupervised question answering has shown that models can be trained with procedurally generated question-answer pairs and can achieve performance competitive with supervised methods. In this work, we consider the task of unsupervised reading comprehension and p...
https://aclanthology.org/2021.naacl-main.95
https://aclanthology.org/2021.naacl-main.95.pdf
NAACL 2021
Capturing Row and Column Semantics in Transformer Based Question Answering over Tables
Michael Glass, Mustafa Canim, Alfio Gliozzo, Saneem Chemmengath, Vishwajeet Kumar, Rishav Chakravarti, Avi Sil, Feifei Pan, Samarth Bharadwaj, Nicolas Rodolfo Fauceglia
Transformer based architectures are recently used for the task of answering questions over tables. In order to improve the accuracy on this task, specialized pre-training techniques have been developed and applied on millions of open-domain web tables. In this paper, we propose two novel approaches demonstrating that o...
https://aclanthology.org/2021.naacl-main.96
https://aclanthology.org/2021.naacl-main.96.pdf
NAACL 2021
Explainable Multi-hop Verbal Reasoning Through Internal Monologue
Zhengzhong Liang, Steven Bethard, Mihai Surdeanu
Many state-of-the-art (SOTA) language models have achieved high accuracy on several multi-hop reasoning problems. However, these approaches tend to not be interpretable because they do not make the intermediate reasoning steps explicit. Moreover, models trained on simpler tasks tend to fail when directly tested on more...
https://aclanthology.org/2021.naacl-main.97
https://aclanthology.org/2021.naacl-main.97.pdf
NAACL 2021
Robust Question Answering Through Sub-part Alignment
Jifan Chen, Greg Durrett
Current textual question answering (QA) models achieve strong performance on in-domain test sets, but often do so by fitting surface-level patterns, so they fail to generalize to out-of-distribution settings. To make a more robust and understandable QA system, we model question answering as an alignment problem. We dec...
https://aclanthology.org/2021.naacl-main.98
https://aclanthology.org/2021.naacl-main.98.pdf
NAACL 2021
Text Modular Networks: Learning to Decompose Tasks in the Language of Existing Models
Tushar Khot, Daniel Khashabi, Kyle Richardson, Peter Clark, Ashish Sabharwal
{'i': 'language', '#text': 'We propose a general framework called Text Modular Networks(TMNs) for building interpretable systems that learn to solve complex tasks by decomposing them into simpler ones solvable by existing models. To ensure solvability of simpler tasks, TMNs learn the textual input-output behavior (i.e....
https://aclanthology.org/2021.naacl-main.99
https://aclanthology.org/2021.naacl-main.99.pdf
NAACL 2021
RECONSIDER: Improved Re-Ranking using Span-Focused Cross-Attention for Open Domain Question Answering
Srinivasan Iyer, Sewon Min, Yashar Mehdad, Wen-tau Yih
State-of-the-art Machine Reading Comprehension (MRC) models for Open-domain Question Answering (QA) are typically trained for span selection using distantly supervised positive examples and heuristically retrieved negative examples. This training scheme possibly explains empirical observations that these models achieve...
https://aclanthology.org/2021.naacl-main.100
https://aclanthology.org/2021.naacl-main.100.pdf
NAACL 2021
End of preview. Expand in Data Studio
README.md exists but content is empty.
Downloads last month
13