title stringlengths 23 150 | authors stringlengths 12 456 | abstract stringlengths 262 1.75k | url stringlengths 42 44 | detail_url stringclasses 1
value | abs stringclasses 1
value | OpenReview stringclasses 1
value | Download PDF stringlengths 46 48 | tags stringclasses 1
value |
|---|---|---|---|---|---|---|---|---|
Knowledge Router: Learning Disentangled Representations for Knowledge Graphs | Shuai Zhang, Xi Rao, Yi Tay, Ce Zhang | The design of expressive representations of entities and relations in a knowledge graph is an important endeavor. While many of the existing approaches have primarily focused on learning from relational patterns and structural information, the intrinsic complexity of KG entities has been more or less overlooked. More c... | https://aclanthology.org/2021.naacl-main.1 | https://aclanthology.org/2021.naacl-main.1.pdf | NAACL 2021 | |||
Distantly Supervised Relation Extraction with Sentence Reconstruction and Knowledge Base Priors | Fenia Christopoulou, Makoto Miwa, Sophia Ananiadou | We propose a multi-task, probabilistic approach to facilitate distantly supervised relation extraction by bringing closer the representations of sentences that contain the same Knowledge Base pairs. To achieve this, we bias the latent space of sentences via a Variational Autoencoder (VAE) that is trained jointly with a... | https://aclanthology.org/2021.naacl-main.2 | https://aclanthology.org/2021.naacl-main.2.pdf | NAACL 2021 | |||
Cross-Task Instance Representation Interactions and Label Dependencies for Joint Information Extraction with Graph Convolutional Networks | Minh Van Nguyen, Viet Dac Lai, Thien Huu Nguyen | Existing works on information extraction (IE) have mainly solved the four main tasks separately (entity mention recognition, relation extraction, event trigger detection, and argument extraction), thus failing to benefit from inter-dependencies between tasks. This paper presents a novel deep learning model to simultane... | https://aclanthology.org/2021.naacl-main.3 | https://aclanthology.org/2021.naacl-main.3.pdf | NAACL 2021 | |||
Abstract Meaning Representation Guided Graph Encoding and Decoding for Joint Information Extraction | Zixuan Zhang, Heng Ji | The tasks of Rich Semantic Parsing, such as Abstract Meaning Representation (AMR), share similar goals with Information Extraction (IE) to convert natural language texts into structured semantic representations. To take advantage of such similarity, we propose a novel AMR-guided framework for joint information extracti... | https://aclanthology.org/2021.naacl-main.4 | https://aclanthology.org/2021.naacl-main.4.pdf | NAACL 2021 | |||
A Frustratingly Easy Approach for Entity and Relation Extraction | Zexuan Zhong, Danqi Chen | End-to-end relation extraction aims to identify named entities and extract relations between them. Most recent work models these two subtasks jointly, either by casting them in one structured prediction framework, or performing multi-task learning through shared representations. In this work, we present a simple pipeli... | https://aclanthology.org/2021.naacl-main.5 | https://aclanthology.org/2021.naacl-main.5.pdf | NAACL 2021 | |||
Event Time Extraction and Propagation via Graph Attention Networks | Haoyang Wen, Yanru Qu, Heng Ji, Qiang Ning, Jiawei Han, Avi Sil, Hanghang Tong, Dan Roth | Grounding events into a precise timeline is important for natural language understanding but has received limited attention in recent work. This problem is challenging due to the inherent ambiguity of language and the requirement for information propagation over inter-related events. This paper first formulates this pr... | https://aclanthology.org/2021.naacl-main.6 | https://aclanthology.org/2021.naacl-main.6.pdf | NAACL 2021 | |||
Probing Word Translations in the Transformer and Trading Decoder for Encoder Layers | Hongfei Xu, Josef van Genabith, Qiuhui Liu, Deyi Xiong | Due to its effectiveness and performance, the Transformer translation model has attracted wide attention, most recently in terms of probing-based approaches. Previous work focuses on using or probing source linguistic features in the encoder. To date, the way word translation evolves in Transformer layers has not yet b... | https://aclanthology.org/2021.naacl-main.7 | https://aclanthology.org/2021.naacl-main.7.pdf | NAACL 2021 | |||
Mediators in Determining what Processing BERT Performs First | Aviv Slobodkin, Leshem Choshen, Omri Abend | Probing neural models for the ability to perform downstream tasks using their activation patterns is often used to localize what parts of the network specialize in performing what tasks. However, little work addressed potential mediating factors in such comparisons. As a test-case mediating factor, we consider the pred... | https://aclanthology.org/2021.naacl-main.8 | https://aclanthology.org/2021.naacl-main.8.pdf | NAACL 2021 | |||
Automatic Generation of Contrast Sets from Scene Graphs: Probing the Compositional Consistency of GQA | Yonatan Bitton, Gabriel Stanovsky, Roy Schwartz, Michael Elhadad | Recent works have shown that supervised models often exploit data artifacts to achieve good test scores while their performance severely degrades on samples outside their training distribution. Contrast sets (Gardneret al., 2020) quantify this phenomenon by perturbing test samples in a minimal way such that the output ... | https://aclanthology.org/2021.naacl-main.9 | https://aclanthology.org/2021.naacl-main.9.pdf | NAACL 2021 | |||
Multilingual Language Models Predict Human Reading Behavior | Nora Hollenstein, Federico Pirovano, Ce Zhang, Lena Jäger, Lisa Beinborn | We analyze if large language models are able to predict patterns of human reading behavior. We compare the performance of language-specific and multilingual pretrained transformer models to predict reading time measures reflecting natural human sentence processing on Dutch, English, German, and Russian texts. This resu... | https://aclanthology.org/2021.naacl-main.10 | https://aclanthology.org/2021.naacl-main.10.pdf | NAACL 2021 | |||
Do Syntactic Probes Probe Syntax? Experiments with Jabberwocky Probing | Rowan Hall Maudslay, Ryan Cotterell | Analysing whether neural language models encode linguistic information has become popular in NLP. One method of doing so, which is frequently cited to support the claim that models like BERT encode syntax, is called probing; probes are small supervised models trained to extract linguistic information from another model... | https://aclanthology.org/2021.naacl-main.11 | https://aclanthology.org/2021.naacl-main.11.pdf | NAACL 2021 | |||
A Non-Linear Structural Probe | Jennifer C. White, Tiago Pimentel, Naomi Saphra, Ryan Cotterell | Probes are models devised to investigate the encoding of knowledge—e.g. syntactic structure—in contextual representations. Probes are often designed for simplicity, which has led to restrictions on probe design that may not allow for the full exploitation of the structure of encoded information; one such restriction is... | https://aclanthology.org/2021.naacl-main.12 | https://aclanthology.org/2021.naacl-main.12.pdf | NAACL 2021 | |||
Concealed Data Poisoning Attacks on NLP Models | Eric Wallace, Tony Zhao, Shi Feng, Sameer Singh | Adversarial attacks alter NLP model predictions by perturbing test-time inputs. However, it is much less understood whether, and how, predictions can be manipulated with small, concealed changes to the training data. In this work, we develop a new data poisoning attack that allows an adversary to control model predicti... | https://aclanthology.org/2021.naacl-main.13 | https://aclanthology.org/2021.naacl-main.13.pdf | NAACL 2021 | |||
Backtranslation Feedback Improves User Confidence in MT, Not Quality | Vilém Zouhar, Michal Novák, Matúš Žilinec, Ondřej Bojar, Mateo Obregón, Robin L. Hill, Frédéric Blain, Marina Fomicheva, Lucia Specia, Lisa Yankovskaya | Translating text into a language unknown to the text’s author, dubbed outbound translation, is a modern need for which the user experience has significant room for improvement, beyond the basic machine translation facility. We demonstrate this by showing three ways in which user confidence in the outbound translation, ... | https://aclanthology.org/2021.naacl-main.14 | https://aclanthology.org/2021.naacl-main.14.pdf | NAACL 2021 | |||
Data Filtering using Cross-Lingual Word Embeddings | Christian Herold, Jan Rosendahl, Joris Vanvinckenroye, Hermann Ney | Data filtering for machine translation (MT) describes the task of selecting a subset of a given, possibly noisy corpus with the aim to maximize the performance of an MT system trained on this selected data. Over the years, many different filtering approaches have been proposed. However, varying task definitions and dat... | https://aclanthology.org/2021.naacl-main.15 | https://aclanthology.org/2021.naacl-main.15.pdf | NAACL 2021 | |||
Improving the Lexical Ability of Pretrained Language Models for Unsupervised Neural Machine Translation | Alexandra Chronopoulou, Dario Stojanovski, Alexander Fraser | Successful methods for unsupervised neural machine translation (UNMT) employ cross-lingual pretraining via self-supervision, often in the form of a masked language modeling or a sequence generation task, which requires the model to align the lexical- and high-level representations of the two languages. While cross-ling... | https://aclanthology.org/2021.naacl-main.16 | https://aclanthology.org/2021.naacl-main.16.pdf | NAACL 2021 | |||
Neural Machine Translation without Embeddings | Uri Shaham, Omer Levy | Many NLP models operate over sequences of subword tokens produced by hand-crafted tokenization rules and heuristic subword induction algorithms. A simple universal alternative is to represent every computerized text as a sequence of bytes via UTF-8, obviating the need for an embedding layer since there are fewer token ... | https://aclanthology.org/2021.naacl-main.17 | https://aclanthology.org/2021.naacl-main.17.pdf | NAACL 2021 | |||
Counterfactual Data Augmentation for Neural Machine Translation | Qi Liu, Matt Kusner, Phil Blunsom | We propose a data augmentation method for neural machine translation. It works by interpreting language models and phrasal alignment causally. Specifically, it creates augmented parallel translation corpora by generating (path-specific) counterfactual aligned phrases. We generate these by sampling new source phrases fr... | https://aclanthology.org/2021.naacl-main.18 | https://aclanthology.org/2021.naacl-main.18.pdf | NAACL 2021 | |||
Cultural and Geographical Influences on Image Translatability of Words across Languages | Nikzad Khani, Isidora Tourni, Mohammad Sadegh Rasooli, Chris Callison-Burch, Derry Tanti Wijaya | Neural Machine Translation (NMT) models have been observed to produce poor translations when there are few/no parallel sentences to train the models. In the absence of parallel data, several approaches have turned to the use of images to learn translations. Since images of words, e.g., horse may be unchanged across lan... | https://aclanthology.org/2021.naacl-main.19 | https://aclanthology.org/2021.naacl-main.19.pdf | NAACL 2021 | |||
Multilingual BERT Post-Pretraining Alignment | Lin Pan, Chung-Wei Hang, Haode Qi, Abhishek Shah, Saloni Potdar, Mo Yu | We propose a simple method to align multilingual contextual embeddings as a post-pretraining step for improved cross-lingual transferability of the pretrained language models. Using parallel data, our method aligns embeddings on the word level through the recently proposed Translation Language Modeling objective as wel... | https://aclanthology.org/2021.naacl-main.20 | https://aclanthology.org/2021.naacl-main.20.pdf | NAACL 2021 | |||
A Million Tweets Are Worth a Few Points: Tuning Transformers for Customer Service Tasks | Amir Hadifar, Sofie Labat, Veronique Hoste, Chris Develder, Thomas Demeester | In online domain-specific customer service applications, many companies struggle to deploy advanced NLP models successfully, due to the limited availability of and noise in their datasets. While prior research demonstrated the potential of migrating large open-domain pretrained models for domain-specific tasks, the app... | https://aclanthology.org/2021.naacl-main.21 | https://aclanthology.org/2021.naacl-main.21.pdf | NAACL 2021 | |||
Paragraph-level Rationale Extraction through Regularization: A case study on European Court of Human Rights Cases | Ilias Chalkidis, Manos Fergadiotis, Dimitrios Tsarapatsanis, Nikolaos Aletras, Ion Androutsopoulos, Prodromos Malakasiotis | Interpretability or explainability is an emerging research field in NLP. From a user-centric point of view, the goal is to build models that provide proper justification for their decisions, similar to those of humans, by requiring the models to satisfy additional constraints. To this end, we introduce a new applicatio... | https://aclanthology.org/2021.naacl-main.22 | https://aclanthology.org/2021.naacl-main.22.pdf | NAACL 2021 | |||
Answering Product-Questions by Utilizing Questions from Other Contextually Similar Products | Ohad Rozen, David Carmel, Avihai Mejer, Vitaly Mirkis, Yftah Ziser | Predicting the answer to a product-related question is an emerging field of research that recently attracted a lot of attention. Answering subjective and opinion-based questions is most challenging due to the dependency on customer generated content. Previous works mostly focused on review-aware answer prediction; howe... | https://aclanthology.org/2021.naacl-main.23 | https://aclanthology.org/2021.naacl-main.23.pdf | NAACL 2021 | |||
EnSidNet: Enhanced Hybrid Siamese-Deep Network for grouping clinical trials into drug-development pathways | Lucia Pagani | Siamese Neural Networks have been widely used to perform similarity classification in multi-class settings. Their architecture can be used to group the clinical trials belonging to the same drug-development pathway along the several clinical trial phases. Here we present an approach for the unmet need of drug-developme... | https://aclanthology.org/2021.naacl-main.24 | https://aclanthology.org/2021.naacl-main.24.pdf | NAACL 2021 | |||
DATE: Detecting Anomalies in Text via Self-Supervision of Transformers | Andrei Manolache, Florin Brad, Elena Burceanu | Leveraging deep learning models for Anomaly Detection (AD) has seen widespread use in recent years due to superior performances over traditional methods. Recent deep methods for anomalies in images learn better features of normality in an end-to-end self-supervised setting. These methods train a model to discriminate b... | https://aclanthology.org/2021.naacl-main.25 | https://aclanthology.org/2021.naacl-main.25.pdf | NAACL 2021 | |||
A Simple Approach for Handling Out-of-Vocabulary Identifiers in Deep Learning for Source Code | Nadezhda Chirkova, Sergey Troshin | There is an emerging interest in the application of natural language processing models to source code processing tasks. One of the major problems in applying deep learning to software engineering is that source code often contains a lot of rare identifiers, resulting in huge vocabularies. We propose a simple, yet effec... | https://aclanthology.org/2021.naacl-main.26 | https://aclanthology.org/2021.naacl-main.26.pdf | NAACL 2021 | |||
Fast and Scalable Dialogue State Tracking with Explicit Modular Decomposition | Dingmin Wang, Chenghua Lin, Qi Liu, Kam-Fai Wong | We present a fast and scalable architecture called Explicit Modular Decomposition (EMD), in which we incorporate both classification-based and extraction-based methods and design four modules (for clas- sification and sequence labelling) to jointly extract dialogue states. Experimental results based on the MultiWoz 2.0... | https://aclanthology.org/2021.naacl-main.27 | https://aclanthology.org/2021.naacl-main.27.pdf | NAACL 2021 | |||
Augmented SBERT: Data Augmentation Method for Improving Bi-Encoders for Pairwise Sentence Scoring Tasks | Nandan Thakur, Nils Reimers, Johannes Daxenberger, Iryna Gurevych | There are two approaches for pairwise sentence scoring: Cross-encoders, which perform full-attention over the input pair, and Bi-encoders, which map each input independently to a dense vector space. While cross-encoders often achieve higher performance, they are too slow for many practical use cases. Bi-encoders, on th... | https://aclanthology.org/2021.naacl-main.28 | https://aclanthology.org/2021.naacl-main.28.pdf | NAACL 2021 | |||
SmBoP: Semi-autoregressive Bottom-up Semantic Parsing | Ohad Rubin, Jonathan Berant | The de-facto standard decoding method for semantic parsing in recent years has been to autoregressively decode the abstract syntax tree of the target program using a top-down depth-first traversal. In this work, we propose an alternative approach: a Semi-autoregressive Bottom-up Parser (SmBoP) that constructs at decodi... | https://aclanthology.org/2021.naacl-main.29 | https://aclanthology.org/2021.naacl-main.29.pdf | NAACL 2021 | |||
SGL: Speaking the Graph Languages of Semantic Parsing via Multilingual Translation | Luigi Procopio, Rocco Tripodi, Roberto Navigli | {'url': 'https://github.com/SapienzaNLP/sgl', '#text': 'Graph-based semantic parsing aims to represent textual meaning through directed graphs. As one of the most promising general-purpose meaning representations, these structures and their parsing have gained a significant interest momentum during recent years, with s... | https://aclanthology.org/2021.naacl-main.30 | https://aclanthology.org/2021.naacl-main.30.pdf | NAACL 2021 | |||
Unifying Cross-Lingual Semantic Role Labeling with Heterogeneous Linguistic Resources | Simone Conia, Andrea Bacciu, Roberto Navigli | {'url': 'https://github.com/SapienzaNLP/unify-srl', '#text': 'While cross-lingual techniques are finding increasing success in a wide range of Natural Language Processing tasks, their application to Semantic Role Labeling (SRL) has been strongly limited by the fact that each language adopts its own linguistic formalism... | https://aclanthology.org/2021.naacl-main.31 | https://aclanthology.org/2021.naacl-main.31.pdf | NAACL 2021 | |||
Fool Me Twice: Entailment from Wikipedia Gamification | Julian Eisenschlos, Bhuwan Dhingra, Jannis Bulian, Benjamin Börschinger, Jordan Boyd-Graber | We release FoolMeTwice (FM2 for short), a large dataset of challenging entailment pairs collected through a fun multi-player game. Gamification encourages adversarial examples, drastically lowering the number of examples that can be solved using “shortcuts” compared to other popular entailment datasets. Players are pre... | https://aclanthology.org/2021.naacl-main.32 | https://aclanthology.org/2021.naacl-main.32.pdf | NAACL 2021 | |||
Meta-Learning for Domain Generalization in Semantic Parsing | Bailin Wang, Mirella Lapata, Ivan Titov | The importance of building semantic parsers which can be applied to new domains and generate programs unseen at training has long been acknowledged, and datasets testing out-of-domain performance are becoming increasingly available. However, little or no attention has been devoted to learning algorithms or objectives w... | https://aclanthology.org/2021.naacl-main.33 | https://aclanthology.org/2021.naacl-main.33.pdf | NAACL 2021 | |||
Aspect-Controlled Neural Argument Generation | Benjamin Schiller, Johannes Daxenberger, Iryna Gurevych | We rely on arguments in our daily lives to deliver our opinions and base them on evidence, making them more convincing in turn. However, finding and formulating arguments can be challenging. In this work, we present the Arg-CTRL - a language model for argument generation that can be controlled to generate sentence-leve... | https://aclanthology.org/2021.naacl-main.34 | https://aclanthology.org/2021.naacl-main.34.pdf | NAACL 2021 | |||
Text Generation from Discourse Representation Structures | Jiangming Liu, Shay B. Cohen, Mirella Lapata | We propose neural models to generate text from formal meaning representations based on Discourse Representation Structures (DRSs). DRSs are document-level representations which encode rich semantic detail pertaining to rhetorical relations, presupposition, and co-reference within and across sentences. We formalize the ... | https://aclanthology.org/2021.naacl-main.35 | https://aclanthology.org/2021.naacl-main.35.pdf | NAACL 2021 | |||
APo-VAE: Text Generation in Hyperbolic Space | Shuyang Dai, Zhe Gan, Yu Cheng, Chenyang Tao, Lawrence Carin, Jingjing Liu | Natural language often exhibits inherent hierarchical structure ingrained with complex syntax and semantics. However, most state-of-the-art deep generative models learn embeddings only in Euclidean vector space, without accounting for this structural property of language. In this paper, we investigate text generation i... | https://aclanthology.org/2021.naacl-main.36 | https://aclanthology.org/2021.naacl-main.36.pdf | NAACL 2021 | |||
DART: Open-Domain Structured Data Record to Text Generation | Linyong Nan, Dragomir Radev, Rui Zhang, Amrit Rau, Abhinand Sivaprasad, Chiachun Hsieh, Xiangru Tang, Aadit Vyas, Neha Verma, Pranav Krishna, Yangxiaokang Liu, Nadia Irwanto, Jessica Pan, Faiaz Rahman, Ahmad Zaidi, Mutethia Mutuma, Yasin Tarabar, Ankit Gupta, Tao Yu, Yi Chern Tan, Xi Victoria Lin, Caiming Xiong, Richar... | {'url': 'https://github.com/Yale-LILY/dart', '#text': 'We present DART, an open domain structured DAta Record to Text generation dataset with over 82k instances (DARTs). Data-to-text annotations can be a costly process, especially when dealing with tables which are the major source of structured data and contain nontri... | https://aclanthology.org/2021.naacl-main.37 | https://aclanthology.org/2021.naacl-main.37.pdf | NAACL 2021 | |||
When Being Unseen from mBERT is just the Beginning: Handling New Languages With Multilingual Language Models | Benjamin Muller, Antonios Anastasopoulos, Benoît Sagot, Djamé Seddah | Transfer learning based on pretraining language models on a large amount of raw data has become a new norm to reach state-of-the-art performance in NLP. Still, it remains unclear how this approach should be applied for unseen languages that are not covered by any available large-scale multilingual language model and fo... | https://aclanthology.org/2021.naacl-main.38 | https://aclanthology.org/2021.naacl-main.38.pdf | NAACL 2021 | |||
Multi-Adversarial Learning for Cross-Lingual Word Embeddings | Haozhou Wang, James Henderson, Paola Merlo | Generative adversarial networks (GANs) have succeeded in inducing cross-lingual word embeddings - maps of matching words across languages - without supervision. Despite these successes, GANs’ performance for the difficult case of distant languages is still not satisfactory. These limitations have been explained by GANs... | https://aclanthology.org/2021.naacl-main.39 | https://aclanthology.org/2021.naacl-main.39.pdf | NAACL 2021 | |||
Multi-view Subword Regularization | Xinyi Wang, Sebastian Ruder, Graham Neubig | Multilingual pretrained representations generally rely on subword segmentation algorithms to create a shared multilingual vocabulary. However, standard heuristic algorithms often lead to sub-optimal segmentation, especially for languages with limited amounts of data. In this paper, we take two major steps towards allev... | https://aclanthology.org/2021.naacl-main.40 | https://aclanthology.org/2021.naacl-main.40.pdf | NAACL 2021 | |||
mT5: A Massively Multilingual Pre-trained Text-to-Text Transformer | Linting Xue, Noah Constant, Adam Roberts, Mihir Kale, Rami Al-Rfou, Aditya Siddhant, Aditya Barua, Colin Raffel | The recent “Text-to-Text Transfer Transformer” (T5) leveraged a unified text-to-text format and scale to attain state-of-the-art results on a wide variety of English-language NLP tasks. In this paper, we introduce mT5, a multilingual variant of T5 that was pre-trained on a new Common Crawl-based dataset covering 101 la... | https://aclanthology.org/2021.naacl-main.41 | https://aclanthology.org/2021.naacl-main.41.pdf | NAACL 2021 | |||
MetaXL: Meta Representation Transformation for Low-resource Cross-lingual Learning | Mengzhou Xia, Guoqing Zheng, Subhabrata Mukherjee, Milad Shokouhi, Graham Neubig, Ahmed Hassan Awadallah | The combination of multilingual pre-trained representations and cross-lingual transfer learning is one of the most effective methods for building functional NLP systems for low-resource languages. However, for extremely low-resource languages without large-scale monolingual corpora for pre-training or sufficient annota... | https://aclanthology.org/2021.naacl-main.42 | https://aclanthology.org/2021.naacl-main.42.pdf | NAACL 2021 | |||
Open Domain Question Answering over Tables via Dense Retrieval | Jonathan Herzig, Thomas Müller, Syrine Krichene, Julian Eisenschlos | Recent advances in open-domain QA have led to strong models based on dense retrieval, but only focused on retrieving textual passages. In this work, we tackle open-domain QA over tables for the first time, and show that retrieval can be improved by a retriever designed to handle tabular context. We present an effective... | https://aclanthology.org/2021.naacl-main.43 | https://aclanthology.org/2021.naacl-main.43.pdf | NAACL 2021 | |||
Open-Domain Question Answering Goes Conversational via Question Rewriting | Raviteja Anantha, Svitlana Vakulenko, Zhucheng Tu, Shayne Longpre, Stephen Pulman, Srinivas Chappidi | We introduce a new dataset for Question Rewriting in Conversational Context (QReCC), which contains 14K conversations with 80K question-answer pairs. The task in QReCC is to find answers to conversational questions within a collection of 10M web pages (split into 54M passages). Answers to questions in the same conversa... | https://aclanthology.org/2021.naacl-main.44 | https://aclanthology.org/2021.naacl-main.44.pdf | NAACL 2021 | |||
QA-GNN: Reasoning with Language Models and Knowledge Graphs for Question Answering | Michihiro Yasunaga, Hongyu Ren, Antoine Bosselut, Percy Liang, Jure Leskovec | The problem of answering questions using knowledge from pre-trained language models (LMs) and knowledge graphs (KGs) presents two challenges: given a QA context (question and answer choice), methods need to (i) identify relevant knowledge from large KGs, and (ii) perform joint reasoning over the QA context and KG. Here... | https://aclanthology.org/2021.naacl-main.45 | https://aclanthology.org/2021.naacl-main.45.pdf | NAACL 2021 | |||
XOR QA: Cross-lingual Open-Retrieval Question Answering | Akari Asai, Jungo Kasai, Jonathan Clark, Kenton Lee, Eunsol Choi, Hannaneh Hajishirzi | {'url': 'https://nlp.cs.washington.edu/xorqa/', '#text': 'Multilingual question answering tasks typically assume that answers exist in the same language as the question. Yet in practice, many languages face both information scarcity—where languages have few reference articles—and information asymmetry—where questions r... | https://aclanthology.org/2021.naacl-main.46 | https://aclanthology.org/2021.naacl-main.46.pdf | NAACL 2021 | |||
SPARTA: Efficient Open-Domain Question Answering via Sparse Transformer Matching Retrieval | Tiancheng Zhao, Xiaopeng Lu, Kyusong Lee | We introduce SPARTA, a novel neural retrieval method that shows great promise in performance, generalization, and interpretability for open-domain question answering. Unlike many neural ranking methods that use dense vector nearest neighbor search, SPARTA learns a sparse representation that can be efficiently implement... | https://aclanthology.org/2021.naacl-main.47 | https://aclanthology.org/2021.naacl-main.47.pdf | NAACL 2021 | |||
Implicitly Abusive Language – What does it actually look like and why are we not getting there? | Michael Wiegand, Josef Ruppenhofer, Elisabeth Eder | Abusive language detection is an emerging field in natural language processing which has received a large amount of attention recently. Still the success of automatic detection is limited. Particularly, the detection of implicitly abusive language, i.e. abusive language that is not conveyed by abusive words (e.g. dumba... | https://aclanthology.org/2021.naacl-main.48 | https://aclanthology.org/2021.naacl-main.48.pdf | NAACL 2021 | |||
The Importance of Modeling Social Factors of Language: Theory and Practice | Dirk Hovy, Diyi Yang | Natural language processing (NLP) applications are now more powerful and ubiquitous than ever before. With rapidly developing (neural) models and ever-more available data, current NLP models have access to more information than any human speaker during their life. Still, it would be hard to argue that NLP models have r... | https://aclanthology.org/2021.naacl-main.49 | https://aclanthology.org/2021.naacl-main.49.pdf | NAACL 2021 | |||
On learning and representing social meaning in NLP: a sociolinguistic perspective | Dong Nguyen, Laura Rosseel, Jack Grieve | The field of NLP has made substantial progress in building meaning representations. However, an important aspect of linguistic meaning, social meaning, has been largely overlooked. We introduce the concept of social meaning to NLP and discuss how insights from sociolinguistics can inform work on representation learning... | https://aclanthology.org/2021.naacl-main.50 | https://aclanthology.org/2021.naacl-main.50.pdf | NAACL 2021 | |||
Preregistering NLP research | Emiel van Miltenburg, Chris van der Lee, Emiel Krahmer | Preregistration refers to the practice of specifying what you are going to do, and what you expect to find in your study, before carrying out the study. This practice is increasingly common in medicine and psychology, but is rarely discussed in NLP. This paper discusses preregistration in more detail, explores how NLP ... | https://aclanthology.org/2021.naacl-main.51 | https://aclanthology.org/2021.naacl-main.51.pdf | NAACL 2021 | |||
Get Your Vitamin C! Robust Fact Verification with Contrastive Evidence | Tal Schuster, Adam Fisch, Regina Barzilay | Typical fact verification models use retrieved written evidence to verify claims. Evidence sources, however, often change over time as more information is gathered and revised. In order to adapt, models must be sensitive to subtle differences in supporting evidence. We present VitaminC, a benchmark infused with challen... | https://aclanthology.org/2021.naacl-main.52 | https://aclanthology.org/2021.naacl-main.52.pdf | NAACL 2021 | |||
Representing Numbers in NLP: a Survey and a Vision | Avijit Thawani, Jay Pujara, Filip Ilievski, Pedro Szekely | NLP systems rarely give special consideration to numbers found in text. This starkly contrasts with the consensus in neuroscience that, in the brain, numbers are represented differently from words. We arrange recent NLP work on numeracy into a comprehensive taxonomy of tasks and methods. We break down the subjective no... | https://aclanthology.org/2021.naacl-main.53 | https://aclanthology.org/2021.naacl-main.53.pdf | NAACL 2021 | |||
Extending Multi-Document Summarization Evaluation to the Interactive Setting | Ori Shapira, Ramakanth Pasunuru, Hadar Ronen, Mohit Bansal, Yael Amsterdamer, Ido Dagan | Allowing users to interact with multi-document summarizers is a promising direction towards improving and customizing summary results. Different ideas for interactive summarization have been proposed in previous work but these solutions are highly divergent and incomparable. In this paper, we develop an end-to-end eval... | https://aclanthology.org/2021.naacl-main.54 | https://aclanthology.org/2021.naacl-main.54.pdf | NAACL 2021 | |||
Identifying Helpful Sentences in Product Reviews | Iftah Gamzu, Hila Gonen, Gilad Kutiel, Ran Levy, Eugene Agichtein | In recent years online shopping has gained momentum and became an important venue for customers wishing to save time and simplify their shopping process. A key advantage of shopping online is the ability to read what other customers are saying about products of interest. In this work, we aim to maintain this advantage ... | https://aclanthology.org/2021.naacl-main.55 | https://aclanthology.org/2021.naacl-main.55.pdf | NAACL 2021 | |||
Noisy Self-Knowledge Distillation for Text Summarization | Yang Liu, Sheng Shen, Mirella Lapata | In this paper we apply self-knowledge distillation to text summarization which we argue can alleviate problems with maximum-likelihood training on single reference and noisy datasets. Instead of relying on one-hot annotation labels, our student summarization model is trained with guidance from a teacher which generates... | https://aclanthology.org/2021.naacl-main.56 | https://aclanthology.org/2021.naacl-main.56.pdf | NAACL 2021 | |||
Improving Zero and Few-Shot Abstractive Summarization with Intermediate Fine-tuning and Data Augmentation | Alexander Fabbri, Simeng Han, Haoyuan Li, Haoran Li, Marjan Ghazvininejad, Shafiq Joty, Dragomir Radev, Yashar Mehdad | Models pretrained with self-supervised objectives on large text corpora achieve state-of-the-art performance on English text summarization tasks. However, these models are typically fine-tuned on hundreds of thousands of data points, an infeasible requirement when applying summarization to new, niche domains. In this w... | https://aclanthology.org/2021.naacl-main.57 | https://aclanthology.org/2021.naacl-main.57.pdf | NAACL 2021 | |||
Enhancing Factual Consistency of Abstractive Summarization | Chenguang Zhu, William Hinthorn, Ruochen Xu, Qingkai Zeng, Michael Zeng, Xuedong Huang, Meng Jiang | Automatic abstractive summaries are found to often distort or fabricate facts in the article. This inconsistency between summary and original text has seriously impacted its applicability. We propose a fact-aware summarization model FASum to extract and integrate factual relations into the summary generation process vi... | https://aclanthology.org/2021.naacl-main.58 | https://aclanthology.org/2021.naacl-main.58.pdf | NAACL 2021 | |||
Few-shot Intent Classification and Slot Filling with Retrieved Examples | Dian Yu, Luheng He, Yuan Zhang, Xinya Du, Panupong Pasupat, Qi Li | Few-shot learning arises in important practical scenarios, such as when a natural language understanding system needs to learn new semantic labels for an emerging, resource-scarce domain. In this paper, we explore retrieval-based methods for intent classification and slot filling tasks in few-shot settings. Retrieval-b... | https://aclanthology.org/2021.naacl-main.59 | https://aclanthology.org/2021.naacl-main.59.pdf | NAACL 2021 | |||
“Nice Try, Kiddo”: Investigating Ad Hominems in Dialogue Responses | Emily Sheng, Kai-Wei Chang, Prem Natarajan, Nanyun Peng | Ad hominem attacks are those that target some feature of a person’s character instead of the position the person is maintaining. These attacks are harmful because they propagate implicit biases and diminish a person’s credibility. Since dialogue systems respond directly to user input, it is important to study ad homine... | https://aclanthology.org/2021.naacl-main.60 | https://aclanthology.org/2021.naacl-main.60.pdf | NAACL 2021 | |||
Human-like informative conversations: Better acknowledgements using conditional mutual information | Ashwin Paranjape, Christopher Manning | This work aims to build a dialogue agent that can weave new factual content into conversations as naturally as humans. We draw insights from linguistic principles of conversational analysis and annotate human-human conversations from the Switchboard Dialog Act Corpus to examine humans strategies for acknowledgement, tr... | https://aclanthology.org/2021.naacl-main.61 | https://aclanthology.org/2021.naacl-main.61.pdf | NAACL 2021 | |||
A Comparative Study on Schema-Guided Dialogue State Tracking | Jie Cao, Yi Zhang | Frame-based state representation is widely used in modern task-oriented dialog systems to model user intentions and slot values. However, a fixed design of domain ontology makes it difficult to extend to new services and APIs. Recent work proposed to use natural language descriptions to define the domain ontology inste... | https://aclanthology.org/2021.naacl-main.62 | https://aclanthology.org/2021.naacl-main.62.pdf | NAACL 2021 | |||
Spoken Language Understanding for Task-oriented Dialogue Systems with Augmented Memory Networks | Jie Wu, Ian Harris, Hongzhi Zhao | Spoken language understanding, usually including intent detection and slot filling, is a core component to build a spoken dialog system. Recent research shows promising results by jointly learning of those two tasks based on the fact that slot filling and intent detection are sharing semantic knowledge. Furthermore, at... | https://aclanthology.org/2021.naacl-main.63 | https://aclanthology.org/2021.naacl-main.63.pdf | NAACL 2021 | |||
How to Motivate Your Dragon: Teaching Goal-Driven Agents to Speak and Act in Fantasy Worlds | Prithviraj Ammanabrolu, Jack Urbanek, Margaret Li, Arthur Szlam, Tim Rocktäschel, Jason Weston | We seek to create agents that both act and communicate with other agents in pursuit of a goal. Towards this end, we extend LIGHT (Urbanek et al. 2019)—a large-scale crowd-sourced fantasy text-game—with a dataset of quests. These contain natural language motivations paired with in-game goals and human demonstrations; co... | https://aclanthology.org/2021.naacl-main.64 | https://aclanthology.org/2021.naacl-main.64.pdf | NAACL 2021 | |||
Linking Entities to Unseen Knowledge Bases with Arbitrary Schemas | Yogarshi Vyas, Miguel Ballesteros | In entity linking, mentions of named entities in raw text are disambiguated against a knowledge base (KB). This work focuses on linking to unseen KBs that do not have training data and whose schema is unknown during training. Our approach relies on methods to flexibly convert entities with several attribute-value pairs... | https://aclanthology.org/2021.naacl-main.65 | https://aclanthology.org/2021.naacl-main.65.pdf | NAACL 2021 | |||
Self-Training with Weak Supervision | Giannis Karamanolakis, Subhabrata Mukherjee, Guoqing Zheng, Ahmed Hassan Awadallah | State-of-the-art deep neural networks require large-scale labeled training data that is often expensive to obtain or not available for many tasks. Weak supervision in the form of domain-specific rules has been shown to be useful in such settings to automatically generate weakly labeled training data. However, learning ... | https://aclanthology.org/2021.naacl-main.66 | https://aclanthology.org/2021.naacl-main.66.pdf | NAACL 2021 | |||
Neural Language Modeling for Contextualized Temporal Graph Generation | Aman Madaan, Yiming Yang | This paper presents the first study on using large-scale pre-trained language models for automated generation of an event-level temporal graph for a document. Despite the huge success of neural pre-training methods in NLP tasks, its potential for temporal reasoning over event graphs has not been sufficiently explored. ... | https://aclanthology.org/2021.naacl-main.67 | https://aclanthology.org/2021.naacl-main.67.pdf | NAACL 2021 | |||
Probabilistic Box Embeddings for Uncertain Knowledge Graph Reasoning | Xuelu Chen, Michael Boratko, Muhao Chen, Shib Sankar Dasgupta, Xiang Lorraine Li, Andrew McCallum | Knowledge bases often consist of facts which are harvested from a variety of sources, many of which are noisy and some of which conflict, resulting in a level of uncertainty for each triple. Knowledge bases are also often incomplete, prompting the use of embedding methods to generalize from known facts, however, existi... | https://aclanthology.org/2021.naacl-main.68 | https://aclanthology.org/2021.naacl-main.68.pdf | NAACL 2021 | |||
Document-Level Event Argument Extraction by Conditional Generation | Sha Li, Heng Ji, Jiawei Han | Event extraction has long been treated as a sentence-level task in the IE community. We argue that this setting does not match human informative seeking behavior and leads to incomplete and uninformative extraction results. We propose a document-level neural event argument extraction model by formulating the task as co... | https://aclanthology.org/2021.naacl-main.69 | https://aclanthology.org/2021.naacl-main.69.pdf | NAACL 2021 | |||
Template Filling with Generative Transformers | Xinya Du, Alexander Rush, Claire Cardie | Template filling is generally tackled by a pipeline of two separate supervised systems – one for role-filler extraction and another for template/event recognition. Since pipelines consider events in isolation, they can suffer from error propagation. We introduce a framework based on end-to-end generative transformers f... | https://aclanthology.org/2021.naacl-main.70 | https://aclanthology.org/2021.naacl-main.70.pdf | NAACL 2021 | |||
Towards Interpreting and Mitigating Shortcut Learning Behavior of NLU models | Mengnan Du, Varun Manjunatha, Rajiv Jain, Ruchi Deshpande, Franck Dernoncourt, Jiuxiang Gu, Tong Sun, Xia Hu | Recent studies indicate that NLU models are prone to rely on shortcut features for prediction, without achieving true language understanding. As a result, these models fail to generalize to real-world out-of-distribution data. In this work, we show that the words in the NLU training set can be modeled as a long-tailed ... | https://aclanthology.org/2021.naacl-main.71 | https://aclanthology.org/2021.naacl-main.71.pdf | NAACL 2021 | |||
On Attention Redundancy: A Comprehensive Study | Yuchen Bian, Jiaji Huang, Xingyu Cai, Jiahong Yuan, Kenneth Church | Multi-layer multi-head self-attention mechanism is widely applied in modern neural language models. Attention redundancy has been observed among attention heads but has not been deeply studied in the literature. Using BERT-base model as an example, this paper provides a comprehensive study on attention redundancy which... | https://aclanthology.org/2021.naacl-main.72 | https://aclanthology.org/2021.naacl-main.72.pdf | NAACL 2021 | |||
Does BERT Pretrained on Clinical Notes Reveal Sensitive Data? | Eric Lehman, Sarthak Jain, Karl Pichotta, Yoav Goldberg, Byron Wallace | {'url': 'https://github.com/elehman16/exposing_patient_data_release', '#text': 'Large Transformers pretrained over clinical notes from Electronic Health Records (EHR) have afforded substantial gains in performance on predictive clinical tasks. The cost of training such models (and the necessity of data access to do so)... | https://aclanthology.org/2021.naacl-main.73 | https://aclanthology.org/2021.naacl-main.73.pdf | NAACL 2021 | |||
Low-Complexity Probing via Finding Subnetworks | Steven Cao, Victor Sanh, Alexander Rush | The dominant approach in probing neural networks for linguistic properties is to train a new shallow multi-layer perceptron (MLP) on top of the model’s internal representations. This approach can detect properties encoded in the model, but at the cost of adding new parameters that may learn the task directly. We instea... | https://aclanthology.org/2021.naacl-main.74 | https://aclanthology.org/2021.naacl-main.74.pdf | NAACL 2021 | |||
An Empirical Comparison of Instance Attribution Methods for NLP | Pouya Pezeshkpour, Sarthak Jain, Byron Wallace, Sameer Singh | {'url': 'https://github.com/successar/instance_attributions_NLP', '#text': 'Widespread adoption of deep models has motivated a pressing need for approaches to interpret network outputs and to facilitate model debugging. Instance attribution methods constitute one means of accomplishing these goals by retrieving trainin... | https://aclanthology.org/2021.naacl-main.75 | https://aclanthology.org/2021.naacl-main.75.pdf | NAACL 2021 | |||
Generalization in Instruction Following Systems | Soham Dan, Michael Zhou, Dan Roth | Understanding and executing natural language instructions in a grounded domain is one of the hallmarks of artificial intelligence. In this paper, we focus on instruction understanding in the blocks world domain and investigate the language understanding abilities of two top-performing systems for the task. We aim to un... | https://aclanthology.org/2021.naacl-main.76 | https://aclanthology.org/2021.naacl-main.76.pdf | NAACL 2021 | |||
LightningDOT: Pre-training Visual-Semantic Embeddings for Real-Time Image-Text Retrieval | Siqi Sun, Yen-Chun Chen, Linjie Li, Shuohang Wang, Yuwei Fang, Jingjing Liu | Multimodal pre-training has propelled great advancement in vision-and-language research. These large-scale pre-trained models, although successful, fatefully suffer from slow inference speed due to enormous computational cost mainly from cross-modal attention in Transformer architecture. When applied to real-life appli... | https://aclanthology.org/2021.naacl-main.77 | https://aclanthology.org/2021.naacl-main.77.pdf | NAACL 2021 | |||
Measuring Social Biases in Grounded Vision and Language Embeddings | Candace Ross, Boris Katz, Andrei Barbu | We generalize the notion of measuring social biases in word embeddings to visually grounded word embeddings. Biases are present in grounded embeddings, and indeed seem to be equally or more significant than for ungrounded embeddings. This is despite the fact that vision and language can suffer from different biases, wh... | https://aclanthology.org/2021.naacl-main.78 | https://aclanthology.org/2021.naacl-main.78.pdf | NAACL 2021 | |||
MTAG: Modal-Temporal Attention Graph for Unaligned Human Multimodal Language Sequences | Jianing Yang, Yongxin Wang, Ruitao Yi, Yuying Zhu, Azaan Rehman, Amir Zadeh, Soujanya Poria, Louis-Philippe Morency | Human communication is multimodal in nature; it is through multiple modalities such as language, voice, and facial expressions, that opinions and emotions are expressed. Data in this domain exhibits complex multi-relational and temporal interactions. Learning from this data is a fundamentally challenging research probl... | https://aclanthology.org/2021.naacl-main.79 | https://aclanthology.org/2021.naacl-main.79.pdf | NAACL 2021 | |||
Grounding Open-Domain Instructions to Automate Web Support Tasks | Nancy Xu, Sam Masling, Michael Du, Giovanni Campagna, Larry Heck, James Landay, Monica Lam | Grounding natural language instructions on the web to perform previously unseen tasks enables accessibility and automation. We introduce a task and dataset to train AI agents from open-domain, step-by-step instructions originally written for people. We build RUSS (Rapid Universal Support Service) to tackle this problem... | https://aclanthology.org/2021.naacl-main.80 | https://aclanthology.org/2021.naacl-main.80.pdf | NAACL 2021 | |||
Modular Networks for Compositional Instruction Following | Rodolfo Corona, Daniel Fried, Coline Devin, Dan Klein, Trevor Darrell | Standard architectures used in instruction following often struggle on novel compositions of subgoals (e.g. navigating to landmarks or picking up objects) observed during training. We propose a modular architecture for following natural language instructions that describe sequences of diverse subgoals. In our approach,... | https://aclanthology.org/2021.naacl-main.81 | https://aclanthology.org/2021.naacl-main.81.pdf | NAACL 2021 | |||
Improving Cross-Modal Alignment in Vision Language Navigation via Syntactic Information | Jialu Li, Hao Tan, Mohit Bansal | Vision language navigation is the task that requires an agent to navigate through a 3D environment based on natural language instructions. One key challenge in this task is to ground instructions with the current visual information that the agent perceives. Most of the existing work employs soft attention over individu... | https://aclanthology.org/2021.naacl-main.82 | https://aclanthology.org/2021.naacl-main.82.pdf | NAACL 2021 | |||
Improving Pretrained Models for Zero-shot Multi-label Text Classification through Reinforced Label Hierarchy Reasoning | Hui Liu, Danqing Zhang, Bing Yin, Xiaodan Zhu | Exploiting label hierarchies has become a promising approach to tackling the zero-shot multi-label text classification (ZS-MTC) problem. Conventional methods aim to learn a matching model between text and labels, using a graph encoder to incorporate label hierarchies to obtain effective label representations (Rios and ... | https://aclanthology.org/2021.naacl-main.83 | https://aclanthology.org/2021.naacl-main.83.pdf | NAACL 2021 | |||
Fine-Tuning Pre-trained Language Model with Weak Supervision: A Contrastive-Regularized Self-Training Approach | Yue Yu, Simiao Zuo, Haoming Jiang, Wendi Ren, Tuo Zhao, Chao Zhang | {'url': 'https://github.com/yueyu1030/COSINE', '#text': 'Fine-tuned pre-trained language models (LMs) have achieved enormous success in many natural language processing (NLP) tasks, but they still require excessive labeled data in the fine-tuning stage. We study the problem of fine-tuning pre-trained LMs using only wea... | https://aclanthology.org/2021.naacl-main.84 | https://aclanthology.org/2021.naacl-main.84.pdf | NAACL 2021 | |||
Posterior Differential Regularization with f-divergence for Improving Model Robustness | Hao Cheng, Xiaodong Liu, Lis Pereira, Yaoliang Yu, Jianfeng Gao | We address the problem of enhancing model robustness through regularization. Specifically, we focus on methods that regularize the model posterior difference between clean and noisy inputs. Theoretically, we provide a connection of two recent methods, Jacobian Regularization and Virtual Adversarial Training, under this... | https://aclanthology.org/2021.naacl-main.85 | https://aclanthology.org/2021.naacl-main.85.pdf | NAACL 2021 | |||
Understanding Hard Negatives in Noise Contrastive Estimation | Wenzheng Zhang, Karl Stratos | The choice of negative examples is important in noise contrastive estimation. Recent works find that hard negatives—highest-scoring incorrect examples under the model—are effective in practice, but they are used without a formal justification. We develop analytical tools to understand the role of hard negatives. Specif... | https://aclanthology.org/2021.naacl-main.86 | https://aclanthology.org/2021.naacl-main.86.pdf | NAACL 2021 | |||
Certified Robustness to Word Substitution Attack with Differential Privacy | Wenjie Wang, Pengfei Tang, Jian Lou, Li Xiong | The robustness and security of natural language processing (NLP) models are significantly important in real-world applications. In the context of text classification tasks, adversarial examples can be designed by substituting words with synonyms under certain semantic and syntactic constraints, such that a well-trained... | https://aclanthology.org/2021.naacl-main.87 | https://aclanthology.org/2021.naacl-main.87.pdf | NAACL 2021 | |||
DReCa: A General Task Augmentation Strategy for Few-Shot Natural Language Inference | Shikhar Murty, Tatsunori B. Hashimoto, Christopher Manning | Meta-learning promises few-shot learners that can adapt to new distributions by repurposing knowledge acquired from previous training. However, we believe meta-learning has not yet succeeded in NLP due to the lack of a well-defined task distribution, leading to attempts that treat datasets as tasks. Such an ad hoc task... | https://aclanthology.org/2021.naacl-main.88 | https://aclanthology.org/2021.naacl-main.88.pdf | NAACL 2021 | |||
Harnessing Multilinguality in Unsupervised Machine Translation for Rare Languages | Xavier Garcia, Aditya Siddhant, Orhan Firat, Ankur Parikh | Unsupervised translation has reached impressive performance on resource-rich language pairs such as English-French and English-German. However, early studies have shown that in more realistic settings involving low-resource, rare languages, unsupervised translation performs poorly, achieving less than 3.0 BLEU. In this... | https://aclanthology.org/2021.naacl-main.89 | https://aclanthology.org/2021.naacl-main.89.pdf | NAACL 2021 | |||
Macro-Average: Rare Types Are Important Too | Thamme Gowda, Weiqiu You, Constantine Lignos, Jonathan May | While traditional corpus-level evaluation metrics for machine translation (MT) correlate well with fluency, they struggle to reflect adequacy. Model-based MT metrics trained on segment-level human judgments have emerged as an attractive replacement due to strong correlation results. These models, however, require poten... | https://aclanthology.org/2021.naacl-main.90 | https://aclanthology.org/2021.naacl-main.90.pdf | NAACL 2021 | |||
Assessing Reference-Free Peer Evaluation for Machine Translation | Sweta Agrawal, George Foster, Markus Freitag, Colin Cherry | Reference-free evaluation has the potential to make machine translation evaluation substantially more scalable, allowing us to pivot easily to new languages or domains. It has been recently shown that the probabilities given by a large, multilingual model can achieve state of the art results when used as a reference-fr... | https://aclanthology.org/2021.naacl-main.91 | https://aclanthology.org/2021.naacl-main.91.pdf | NAACL 2021 | |||
The Curious Case of Hallucinations in Neural Machine Translation | Vikas Raunak, Arul Menezes, Marcin Junczys-Dowmunt | In this work, we study hallucinations in Neural Machine Translation (NMT), which lie at an extreme end on the spectrum of NMT pathologies. Firstly, we connect the phenomenon of hallucinations under source perturbation to the Long-Tail theory of Feldman, and present an empirically validated hypothesis that explains hall... | https://aclanthology.org/2021.naacl-main.92 | https://aclanthology.org/2021.naacl-main.92.pdf | NAACL 2021 | |||
Towards Continual Learning for Multilingual Machine Translation via Vocabulary Substitution | Xavier Garcia, Noah Constant, Ankur Parikh, Orhan Firat | We propose a straightforward vocabulary adaptation scheme to extend the language capacity of multilingual machine translation models, paving the way towards efficient continual learning for multilingual machine translation. Our approach is suitable for large-scale datasets, applies to distant languages with unseen scri... | https://aclanthology.org/2021.naacl-main.93 | https://aclanthology.org/2021.naacl-main.93.pdf | NAACL 2021 | |||
Towards Modeling the Style of Translators in Neural Machine Translation | Yue Wang, Cuong Hoang, Marcello Federico | One key ingredient of neural machine translation is the use of large datasets from different domains and resources (e.g. Europarl, TED talks). These datasets contain documents translated by professional translators using different but consistent translation styles. Despite that, the model is usually trained in a way th... | https://aclanthology.org/2021.naacl-main.94 | https://aclanthology.org/2021.naacl-main.94.pdf | NAACL 2021 | |||
Self-Supervised Test-Time Learning for Reading Comprehension | Pratyay Banerjee, Tejas Gokhale, Chitta Baral | {'i': 'context-question-answer', '#text': 'Recent work on unsupervised question answering has shown that models can be trained with procedurally generated question-answer pairs and can achieve performance competitive with supervised methods. In this work, we consider the task of unsupervised reading comprehension and p... | https://aclanthology.org/2021.naacl-main.95 | https://aclanthology.org/2021.naacl-main.95.pdf | NAACL 2021 | |||
Capturing Row and Column Semantics in Transformer Based Question Answering over Tables | Michael Glass, Mustafa Canim, Alfio Gliozzo, Saneem Chemmengath, Vishwajeet Kumar, Rishav Chakravarti, Avi Sil, Feifei Pan, Samarth Bharadwaj, Nicolas Rodolfo Fauceglia | Transformer based architectures are recently used for the task of answering questions over tables. In order to improve the accuracy on this task, specialized pre-training techniques have been developed and applied on millions of open-domain web tables. In this paper, we propose two novel approaches demonstrating that o... | https://aclanthology.org/2021.naacl-main.96 | https://aclanthology.org/2021.naacl-main.96.pdf | NAACL 2021 | |||
Explainable Multi-hop Verbal Reasoning Through Internal Monologue | Zhengzhong Liang, Steven Bethard, Mihai Surdeanu | Many state-of-the-art (SOTA) language models have achieved high accuracy on several multi-hop reasoning problems. However, these approaches tend to not be interpretable because they do not make the intermediate reasoning steps explicit. Moreover, models trained on simpler tasks tend to fail when directly tested on more... | https://aclanthology.org/2021.naacl-main.97 | https://aclanthology.org/2021.naacl-main.97.pdf | NAACL 2021 | |||
Robust Question Answering Through Sub-part Alignment | Jifan Chen, Greg Durrett | Current textual question answering (QA) models achieve strong performance on in-domain test sets, but often do so by fitting surface-level patterns, so they fail to generalize to out-of-distribution settings. To make a more robust and understandable QA system, we model question answering as an alignment problem. We dec... | https://aclanthology.org/2021.naacl-main.98 | https://aclanthology.org/2021.naacl-main.98.pdf | NAACL 2021 | |||
Text Modular Networks: Learning to Decompose Tasks in the Language of Existing Models | Tushar Khot, Daniel Khashabi, Kyle Richardson, Peter Clark, Ashish Sabharwal | {'i': 'language', '#text': 'We propose a general framework called Text Modular Networks(TMNs) for building interpretable systems that learn to solve complex tasks by decomposing them into simpler ones solvable by existing models. To ensure solvability of simpler tasks, TMNs learn the textual input-output behavior (i.e.... | https://aclanthology.org/2021.naacl-main.99 | https://aclanthology.org/2021.naacl-main.99.pdf | NAACL 2021 | |||
RECONSIDER: Improved Re-Ranking using Span-Focused Cross-Attention for Open Domain Question Answering | Srinivasan Iyer, Sewon Min, Yashar Mehdad, Wen-tau Yih | State-of-the-art Machine Reading Comprehension (MRC) models for Open-domain Question Answering (QA) are typically trained for span selection using distantly supervised positive examples and heuristically retrieved negative examples. This training scheme possibly explains empirical observations that these models achieve... | https://aclanthology.org/2021.naacl-main.100 | https://aclanthology.org/2021.naacl-main.100.pdf | NAACL 2021 |
End of preview. Expand in Data Studio
README.md exists but content is empty.
- Downloads last month
- 13