Paper Digest: EMNLP 2021 Highlights
The Conference on Empirical Methods in Natural Language Processing (EMNLP) is one of the top natural language processing conferences in the world. In 2021, it will be held both online and in Punta Cana, Dominican Republic.
To help the community quickly catch up on the work presented in this conference, Paper Digest Team processed all accepted papers, and generated one highlight sentence (typically the main topic) for each paper. Readers are encouraged to read these machine generated highlights / summaries to quickly get the main idea of each paper. Based in New York, Paper Digest is dedicated to producing high-quality text analysis results that people can acturally use on a daily basis. In the past 4 years, we have been serving users across the world with a number of exclusive services on ranking, search, tracking and review. This month we feature Literature Review Generator, which automatically generates literature review around any topic.
If you do not want to miss any interesting academic paper, you are welcome to sign up our free daily paper digest service to get updates on new papers published in your area every day. You are also welcome to follow us on Twitter and Linkedin to get updated with new conference digests.
Paper Digest Team
team@paperdigest.org
TABLE 1: Paper Digest: EMNLP 2021 Highlights
Paper | Author(s) | Code | |
---|---|---|---|
1 | AligNART: Non-autoregressive Neural Machine Translation By Jointly Learning to Estimate Alignment and Translate Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we introduce AligNART, which leverages full alignment information to explicitly reduce the modality of the target distribution. |
Jongyoon Song; Sungwon Kim; Sungroh Yoon; | |
2 | Zero-Shot Cross-Lingual Transfer of Neural Machine Translation with Multilingual Pretrained Encoders Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we focus on a zero-shot cross-lingual transfer task in NMT. |
Guanhua Chen; Shuming Ma; Yun Chen; Li Dong; Dongdong Zhang; Jia Pan; Wenping Wang; Furu Wei; | |
3 | ERNIE-M: Enhanced Multilingual Representation By Aligning Cross-lingual Semantics with Monolingual Corpora Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we propose Ernie-M, a new training method that encourages the model to align the representation of multiple languages with monolingual corpora, to overcome the constraint that the parallel corpus size places on the model performance. |
Xuan Ouyang; Shuohuan Wang; Chao Pang; Yu Sun; Hao Tian; Hua Wu; Haifeng Wang; | |
4 | Cross Attention Augmented Transducer Networks for Simultaneous Translation Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: This paper proposes a novel architecture, Cross Attention Augmented Transducer (CAAT), for simultaneous translation. |
Dan Liu; Mengge Du; Xiaoxi Li; Ya Li; Enhong Chen; | |
5 | Translating Headers of Tabular Data: A Pilot Study of Schema Translation Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: To facilitate the research study, we construct the first parallel dataset for schema translation, which consists of 3,158 tables with 11,979 headers written in 6 different languages, including English, Chinese, French, German, Spanish, and Japanese. Also, we propose the first schema translation model called CAST, which is a header-to-header neural machine translation model augmented with schema context. |
Kunrui Zhu; Yan Gao; Jiaqi Guo; Jian-Guang Lou; | code |
6 | Towards Making The Most of Dialogue Characteristics for Neural Chat Translation Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we propose to promote the chat translation by introducing the modeling of dialogue characteristics into the NCT model. |
Yunlong Liang; Chulun Zhou; Fandong Meng; Jinan Xu; Yufeng Chen; Jinsong Su; Jie Zhou; | |
7 | Low-Resource Dialogue Summarization with Domain-Agnostic Multi-Source Pretraining Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: To bridge the gap between out-of-domain pretraining and in-domain fine-tuning, in this work, we propose a multi-source pretraining paradigm to better leverage the external summary data. |
Yicheng Zou; Bolin Zhu; Xingwu Hu; Tao Gui; Qi Zhang; | |
8 | Controllable Neural Dialogue Summarization with Personal Named Entity Planning Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we propose a controllable neural generation framework that can flexibly guide dialogue summarization with personal named entity planning. |
Zhengyuan Liu; Nancy Chen; | |
9 | Fine-grained Factual Consistency Assessment for Abstractive Summarization Models Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: This paper proposes a fine-grained two-stage Fact Consistency assessment framework for Summarization models (SumFC). |
Sen Zhang; Jianwei Niu; Chuyuan Wei; | |
10 | Decision-Focused Summarization Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We propose a novel problem, decision-focused summarization, where the goal is to summarize relevant information for a decision. |
Chao-Chun Hsu; Chenhao Tan; | |
11 | Multiplex Graph Neural Network for Extractive Text Summarization Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: To address these problems, we propose a novel Multiplex Graph Convolutional Network (Multi-GCN) to jointly model different types of relationships among sentences and words. |
Baoyu Jing; Zeyu You; Tao Yang; Wei Fan; Hanghang Tong; | |
12 | A Thorough Evaluation of Task-Specific Pretraining for Summarization Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We compare three summarization specific pretraining objectives with the task agnostic corrupted span prediction pretraining in controlled study. |
Sascha Rothe; Joshua Maynez; Shashi Narayan; | |
13 | HETFORMER: Heterogeneous Transformer with Sparse Attention for Long-Text Extractive Summarization Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: To mitigate these issues, this paper proposes HetFormer, a Transformer-based pre-trained model with multi-granularity sparse attentions for long-text extractive summarization. |
Ye Liu; Jianguo Zhang; Yao Wan; Congying Xia; Lifang He; Philip Yu; | |
14 | Unsupervised Keyphrase Extraction By Jointly Modeling Local and Global Context Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we propose a novel method for UKE, where local and global contexts are jointly modeled. |
Xinnian Liang; Shuangzhi Wu; Mu Li; Zhoujun Li; | |
15 | Distantly Supervised Relation Extraction Using Multi-Layer Revision Network and Confidence-based Multi-Instance Learning Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we propose a novel Multi-Layer Revision Network (MLRN) which alleviates the effects of word-level noise by emphasizing inner-sentence correlations before extracting relevant information within sentences. |
Xiangyu Lin; Tianyi Liu; Weijia Jia; Zhiguo Gong; | |
16 | Logic-level Evidence Retrieval and Graph-based Verification Network for Table-based Fact Verification Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: To address the aforementioned problems, in this work, we formulate the table-based fact verification task as an evidence retrieval and reasoning framework, proposing the Logic-level Evidence Retrieval and Graph-based Verification network (LERGV). |
Qi Shi; Yu Zhang; Qingyu Yin; Ting Liu; | |
17 | A Partition Filter Network for Joint Entity and Relation Extraction Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We propose a partition filter network to model two-way interaction between tasks properly, where feature encoding is decomposed into two steps: partition and filter. |
Zhiheng Yan; Chong Zhang; Jinlan Fu; Qi Zhang; Zhongyu Wei; | code |
18 | TEBNER: Domain Specific Named Entity Recognition with Type Expanded Boundary-aware Network Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we propose a novel dictionary extension method which extracts new entities through the type expanded model. |
Zheng Fang; Yanan Cao; Tai Li; Ruipeng Jia; Fang Fang; Yanmin Shang; Yuhai Lu; | |
19 | Beta Distribution Guided Aspect-aware Graph for Aspect Category Sentiment Analysis with Affective Knowledge Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we investigate the Aspect Category Sentiment Analysis (ACSA) task from a novel perspective by exploring a Beta Distribution guided aspect-aware graph construction based on external knowledge. |
Bin Liang; Hang Su; Rongdi Yin; Lin Gui; Min Yang; Qin Zhao; Xiaoqi Yu; Ruifeng Xu; | |
20 | DILBERT: Customized Pre-Training for Domain Adaptation with Category Shift, with An Application to Aspect Extraction Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: This paper presents a new fine-tuning scheme for BERT, which aims to address the above challenges. |
Entony Lekhtman; Yftah Ziser; Roi Reichart; | |
21 | Improving Multimodal Fusion Via Mutual Dependency Maximisation Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this work, we investigate unexplored penalties and propose a set of new objectives that measure the dependency between modalities. |
Pierre Colombo; Emile Chapuis; Matthieu Labeau; Chlo? Clavel; | |
22 | Learning Implicit Sentiment in Aspect-based Sentiment Analysis with Supervised Contrastive Pre-Training Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: To overcome this issue, we adopt Supervised Contrastive Pre-training on large-scale sentiment-annotated corpora retrieved from in-domain language resources. |
Zhengyan Li; Yicheng Zou; Chong Zhang; Qi Zhang; Zhongyu Wei; | |
23 | Progressive Self-Training with Discriminator for Aspect Term Extraction Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we use two means to alleviate the noise in the pseudo-labels. |
Qianlong Wang; Zhiyuan Wen; Qin Zhao; Min Yang; Ruifeng Xu; | |
24 | Reinforced Counterfactual Data Augmentation for Dual Sentiment Classification Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this work, we propose an end-to-end reinforcement learning framework, which jointly performs counterfactual data generation and dual sentiment classification. |
Hao Chen; Rui Xia; Jianfei Yu; | |
25 | Idiosyncratic But Not Arbitrary: Learning Idiolects in Online Registers Reveals Distinctive Yet Consistent Individual Styles Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We introduce a new approach to studying idiolects through a massive cross-author comparison to identify and encode stylistic features. |
Jian Zhu; David Jurgens; | |
26 | Narrative Theory for Computational Narrative Understanding Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this position paper, we introduce the dominant theoretical frameworks to the NLP community, situate current research in NLP within distinct narratological traditions, and argue that linking computational work in NLP to theory opens up a range of new empirical questions that would both help advance our understanding of narrative and open up new practical applications. |
Andrew Piper; Richard Jean So; David Bamman; | |
27 | (Mis)alignment Between Stance Expressed in Social Media Data and Public Opinion Surveys Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: By presenting a framework for assessing the limitations of stance detection models, this work provides important insight into what stance detection truly measures. |
Kenneth Joseph; Sarah Shugars; Ryan Gallagher; Jon Green; Alexi Quintana Math?; Zijian An; David Lazer; | |
28 | How Does Counterfactually Augmented Data Impact Models for Social Computing Constructs? Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We investigate the benefits of CAD for social NLP models by focusing on three social computing constructs – sentiment, sexism, and hate speech. |
Indira Sen; Mattia Samory; Fabian Fl?ck; Claudia Wagner; Isabelle Augenstein; | |
29 | Latent Hatred: A Benchmark for Understanding Implicit Hate Speech Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: To fill this gap, this work introduces a theoretically-justified taxonomy of implicit hate speech and a benchmark corpus with fine-grained labels for each message and its implication. |
Mai ElSherief; Caleb Ziems; David Muchlinski; Vaishnavi Anupindi; Jordyn Seybolt; Munmun De Choudhury; Diyi Yang; | |
30 | Distilling Linguistic Context for Language Model Compression Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, inspired by the recent observations that language representations are relatively positioned and have more semantic knowledge as a whole, we present a new knowledge distillation objective for language representation learning that transfers the contextual knowledge via two types of relationships across representations: Word Relation and Layer Transforming Relation. |
Geondo Park; Gyeongman Kim; Eunho Yang; | |
31 | Dynamic Knowledge Distillation for Pre-trained Language Models Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we explore whether a dynamic knowledge distillation that empowers the student to adjust the learning procedure according to its competency, regarding the student performance and learning efficiency. |
Lei Li; Yankai Lin; Shuhuai Ren; Peng Li; Jie Zhou; Xu Sun; | |
32 | Few-Shot Text Generation with Natural Language Instructions Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we show how these challenges can be tackled: We introduce GenPET, a method for text generation that is based on pattern-exploiting training, a recent approach for combining textual instructions with supervised learning that only works for classification tasks. |
Timo Schick; Hinrich Sch?tze; | |
33 | SOM-NCSCM : An Efficient Neural Chinese Sentence Compression Model Enhanced with Self-Organizing Map Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this work, we construct an SC dataset of Chinese colloquial sentences from a real-life question answering system in the telecommunication domain, and then, we propose a neural Chinese SC model enhanced with a Self-Organizing Map (SOM-NCSCM), to gain a valuable insight from the data and improve the performance of the whole neural Chinese SC model in a valid manner. |
Kangli Zi; Shi Wang; Yu Liu; Jicun Li; Yanan Cao; Cungen Cao; | |
34 | Efficient Multi-Task Auxiliary Learning: Selecting Auxiliary Data By Feature Similarity Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: Therefore, this paper focuses on addressing these problems and proposes a time-efficient sampling method to select the data that is most relevant to the primary task. |
Po-Nien Kung; Sheng-Siang Yin; Yi-Cheng Chen; Tse-Hsuan Yang; Yun-Nung Chen; | |
35 | GOLD: Improving Out-of-Scope Detection in Dialogues Using Data Augmentation Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We introduce GOLD as an orthogonal technique that augments existing data to train better OOS detectors operating in low-data regimes. |
Derek Chen; Zhou Yu; | |
36 | Graph Based Network with Contextualized Representations of Turns in Dialogue Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we propose the TUrn COntext awaRE Graph Convolutional Network (TUCORE-GCN) modeled by paying attention to the way people understand dialogues. |
Bongseok Lee; Yong Suk Choi; | code |
37 | Automatically Exposing Problems with Neural Dialog Models Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we propose two methods including reinforcement learning to automatically trigger a dialog model into generating problematic responses. |
Dian Yu; Kenji Sagae; | |
38 | Event Coreference Data (Almost) for Free: Mining Hyperlinks from Online News Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We demonstrate that collecting hyperlinks which point to the same article(s) produces extensive and high-quality CDCR data and create a corpus of 2M documents and 2.7M silver-standard event mentions called HyperCoref. |
Michael Bugert; Iryna Gurevych; | |
39 | Inducing Stereotypical Character Roles from Plot Structure Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We present a fully unsupervised k-means clustering approach for learning stereotypical roles given only structural plot information. |
Labiba Jahan; Rahul Mittal; Mark Finlayson; | |
40 | Multitask Semi-Supervised Learning for Class-Imbalanced Discourse Classification Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this work, we show that a multitask learning approach can combine discourse datasets from similar and diverse domains to improve discourse classification. |
Alexander Spangher; Jonathan May; Sz-Rung Shiang; Lingjia Deng; | |
41 | Low Frequency Names Exhibit Bias and Overfitting in Contextualizing Language Models Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We use a dataset of U.S. first names with labels based on predominant gender and racial group to examine the effect of training corpus frequency on tokenization, contextualization, similarity to initial representation, and bias in BERT, GPT-2, T5, and XLNet. |
Robert Wolfe; Aylin Caliskan; | |
42 | Mitigating Language-Dependent Ethnic Bias in BERT Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we study ethnic bias and how it varies across languages by analyzing and mitigating ethnic bias in monolingual BERT for English, German, Spanish, Korean, Turkish, and Chinese. |
Jaimeen Ahn; Alice Oh; | |
43 | Adversarial Scrubbing of Demographic Information for Text Classification Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we present an adversarial learning framework Adversarial Scrubber (AdS), to debias contextual representations. |
Somnath Basu Roy Chowdhury; Sayan Ghosh; Yiyuan Li; Junier Oliva; Shashank Srivastava; Snigdha Chaturvedi; | |
44 | Open-domain Clarification Question Generation Without Question Examples Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We propose a framework for building a visually grounded question-asking model capable of producing polar (yes-no) clarification questions to resolve misunderstandings in dialogue. |
Julia White; Gabriel Poesia; Robert Hawkins; Dorsa Sadigh; Noah Goodman; | |
45 | Improving Sequence-to-Sequence Pre-training Via Sequence Span Rewriting Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we propose Sequence Span Rewriting (SSR), a self-supervised task for sequence-to-sequence (Seq2Seq) pre-training. |
Wangchunshu Zhou; Tao Ge; Canwen Xu; Ke Xu; Furu Wei; | |
46 | Coarse2Fine: Fine-grained Text Classification on Coarsely-grained Annotated Data Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: To accommodate such requirements, we introduce a new problem called coarse-to-fine grained classification, which aims to perform fine-grained classification on coarsely annotated data. |
Dheeraj Mekala; Varun Gangal; Jingbo Shang; | |
47 | Text2Mol: Cross-Modal Molecule Retrieval with Natural Language Queries Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We propose a new task, Text2Mol, to retrieve molecules using natural language descriptions as queries. |
Carl Edwards; ChengXiang Zhai; Heng Ji; | |
48 | Classification of Hierarchical Text Using Geometric Deep Learning: The Case of Clinical Trials Corpus Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We consider the hierarchical representation of documents as graphs and use geometric deep learning to classify them into different categories. |
Sohrab Ferdowsi; Nikolay Borissov; Julien Knafou; Poorya Amini; Douglas Teodoro; | |
49 | The Devil Is in The Detail: Simple Tricks Improve Systematic Generalization of Transformers Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: Here we demonstrate that by revisiting model configurations as basic as scaling of embeddings, early stopping, relative positional embedding, and Universal Transformer variants, we can drastically improve the performance of Transformers on systematic generalization. |
R?bert Csord?s; Kazuki Irie; Juergen Schmidhuber; | |
50 | Artificial Text Detection Via Examining The Topology of Attention Maps Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: To this end, we propose three novel types of interpretable topological features for this task based on Topological Data Analysis (TDA) which is currently understudied in the field of NLP. |
Laida Kushnareva; Daniil Cherniavskii; Vladislav Mikhailov; Ekaterina Artemova; Serguei Barannikov; Alexander Bernstein; Irina Piontkovskaya; Dmitri Piontkovski; Evgeny Burnaev; | |
51 | Active Learning By Acquiring Contrastive Examples Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this work, leveraging the best of both worlds, we propose an acquisition function that opts for selecting contrastive examples, i.e. data points that are similar in the model feature space and yet the model outputs maximally different predictive likelihoods. |
Katerina Margatina; Giorgos Vernikos; Lo?c Barrault; Nikolaos Aletras; | |
52 | Conditional Poisson Stochastic Beams Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this work, we propose a new method for turning beam search into a stochastic process: Conditional Poisson stochastic beam search. |
Clara Meister; Afra Amini; Tim Vieira; Ryan Cotterell; | |
53 | Building Adaptive Acceptability Classifiers for Neural NLG Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We propose a novel framework to train models to classify acceptability of responses generated by natural language generation (NLG) models, improving upon existing sentence transformation and model-based approaches. |
Soumya Batra; Shashank Jain; Peyman Heidari; Ankit Arun; Catharine Youngs; Xintong Li; Pinar Donmez; Shawn Mei; Shiunzu Kuo; Vikas Bhardwaj; Anuj Kumar; Michael White; | |
54 | Moral Stories: Situated Reasoning About Norms, Intents, Actions, and Their Consequences Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: For this purpose, we introduce Moral Stories, a crowd-sourced dataset of structured, branching narratives for the study of grounded, goal-oriented social reasoning. |
Denis Emelin; Ronan Le Bras; Jena D. Hwang; Maxwell Forbes; Yejin Choi; | |
55 | Truth-Conditional Captions for Time Series Data Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we explore the task of automatically generating natural language descriptions of salient patterns in a time series, such as stock prices of a company over a week. |
Harsh Jhamtani; Taylor Berg-Kirkpatrick; | |
56 | Injecting Entity Types Into Entity-Guided Text Generation Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: To enhance the role of entity in NLG, in this paper, we aim to model the entity type in the decoding phase to generate contextual words accurately. |
Xiangyu Dong; Wenhao Yu; Chenguang Zhu; Meng Jiang; | |
57 | Smelting Gold and Silver for Improved Multilingual AMR-to-Text Generation Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we investigate different techniques for automatically generating AMR annotations, where we aim to study which source of information yields better multilingual results. |
Leonardo F. R. Ribeiro; Jonas Pfeiffer; Yue Zhang; Iryna Gurevych; | |
58 | Learning Compact Metrics for MT Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We present a series of experiments which show that model size is indeed a bottleneck for cross-lingual transfer, then demonstrate how distillation can help addressing this bottleneck, by leveraging synthetic data generation and transferring knowledge from one teacher to multiple students trained on related languages. |
Amy Pu; Hyung Won Chung; Ankur Parikh; Sebastian Gehrmann; Thibault Sellam; | |
59 | The Impact of Positional Encodings on Multilingual Compression Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We first show that surprisingly, while these modifications tend to improve monolingual language models, none of them result in better multilingual language models. We then answer why that is: sinusoidal encodings were explicitly designed to facilitate compositionality by allowing linear projections over arbitrary time steps. |
Vinit Ravishankar; Anders S?gaard; | |
60 | Disentangling Representations of Text By Masking Transformers Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper we explore whether it is possible to learn disentangled representations by identifying existing subnetworks within pretrained models that encode distinct, complementary aspects. |
Xiongyi Zhang; Jan-Willem van de Meent; Byron Wallace; | |
61 | Exploring The Role of BERT Token Representations to Explain Sentence Probing Results Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we provide a more in-depth analysis on the representation space of BERT in search for distinct and meaningful subspaces that can explain the reasons behind these probing results. |
Hosein Mohebbi; Ali Modarressi; Mohammad Taher Pilehvar; | |
62 | Do Long-Range Language Models Actually Use Long-Range Context? Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we perform a fine-grained analysis of two long-range Transformer language models (including the Routing Transformer, which achieves state-of-the-art perplexity on the PG-19 long-sequence LM benchmark dataset) that accept input sequences of up to 8K tokens. |
Simeng Sun; Kalpesh Krishna; Andrew Mattarella-Micke; Mohit Iyyer; | |
63 | The World of An Octopus: How Reporting Bias Influences A Language Model’s Perception of Color Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we first demonstrate that reporting bias, the tendency of people to not state the obvious, is one of the causes of this limitation, and then investigate to what extent multimodal training can mitigate this issue. |
Cory Paik; St?phane Aroca-Ouellette; Alessandro Roncone; Katharina Kann; | |
64 | SELFEXPLAIN: A Self-Explaining Architecture for Neural Text Classifiers Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We introduce SelfExplain, a novel self-explaining model that explains a text classifier’s predictions using phrase-based concepts. |
Dheeraj Rajagopal; Vidhisha Balachandran; Eduard H Hovy; Yulia Tsvetkov; | |
65 | Memory and Knowledge Augmented Language Models for Inferring Salience in Long-Form Stories Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: This paper takes a recent unsupervised method for salience detection derived from Barthes Cardinal Functions and theories of surprise and applies it to longer narrative forms. |
David Wilmot; Frank Keller; | |
66 | Semantic Novelty Detection in Natural Language Descriptions Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: This paper proposes to study a fine-grained semantic novelty detection task, which can be illustrated with the following example. |
Nianzu Ma; Alexander Politowicz; Sahisnu Mazumder; Jiahua Chen; Bing Liu; Eric Robertson; Scott Grigsby; | |
67 | Jump-Starting Item Parameters for Adaptive Language Tests Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: While prior work has addressed ‘cold start’ estimation of item difficulties without piloting, we devise a multi-task generalized linear model with BERT features to jump-start these estimates, rapidly improving their quality with as few as 500 test-takers and a small sample of item exposures (~6 each) from a large item bank (~4,000 items). |
Arya D. McCarthy; Kevin P. Yancey; Geoff T. LaFlair; Jesse Egbert; Manqian Liao; Burr Settles; | |
68 | Voice Query Auto Completion Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we extend QAC to the streaming voice search setting, where automatic speech recognition systems produce intermediate transcriptions as users speak. |
Raphael Tang; Karun Kumar; Kendra Chalkley; Ji Xin; Liming Zhang; Wenyan Li; Gefei Yang; Yajie Mao; Junho Shin; Geoffrey Craig Murray; Jimmy Lin; | |
69 | CoPHE: A Count-Preserving Hierarchical Evaluation Metric in Large-Scale Multi-Label Text Classification Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this work we argue for hierarchical evaluation of the predictions of neural LMTC models. |
Mat?? Falis; Hang Dong; Alexandra Birch; Beatrice Alex; | |
70 | Learning Universal Authorship Representations Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: To study these questions, we conduct the first large-scale study of cross-domain transfer for authorship verification considering zero-shot transfers involving three disparate domains: Amazon reviews, fanfiction short stories, and Reddit comments. |
Rafael A. Rivera-Soto; Olivia Elizabeth Miano; Juanita Ordonez; Barry Y. Chen; Aleem Khan; Marcus Bishop; Nicholas Andrews; | |
71 | Predicting Emergent Linguistic Compositions Through Time: Syntactic Frame Extension Via Multimodal Chaining Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We present the syntactic frame extension model (SFEM) that draws on the theory of chaining and knowledge from percept, concept, and language to infer how verbs extend their frames to form new compositions with existing and novel nouns. |
Lei Yu; Yang Xu; | |
72 | Frequency Effects on Syntactic Rule Learning in Transformers Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: Pre-trained language models perform well on a variety of linguistic tasks that require symbolic reasoning, raising the question of whether such models implicitly represent abstract symbols and rules. We investigate this question using the case study of BERT’s performance on English subject-verb agreement. |
Jason Wei; Dan Garrette; Tal Linzen; Ellie Pavlick; | |
73 | A Surprisal-duration Trade-off Across and Within The World’s Languages Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: Specifically, we find that, on average, phones are produced faster in languages where they are less surprising, and vice versa. |
Tiago Pimentel; Clara Meister; Elizabeth Salesky; Simone Teufel; Dami?n Blasi; Ryan Cotterell; | |
74 | Revisiting The Uniform Information Density Hypothesis Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: Here we investigate these facets of the UID hypothesis using reading time and acceptability data. |
Clara Meister; Tiago Pimentel; Patrick Haller; Lena J?ger; Ryan Cotterell; Roger Levy; | |
75 | Condenser: A Pre-training Architecture for Dense Retrieval Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We propose to pre-train towards dense encoder with a novel Transformer architecture, Condenser, where LM prediction CONditions on DENSE Representation. |
Luyu Gao; Jamie Callan; | |
76 | Monitoring Geometrical Properties of Word Embeddings for Detecting The Emergence of New Topics Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this work, we tackle the problem of early detection of slowly emerging new topics. |
Cl?ment Christophe; Julien Velcin; Jairo Cugliari; Manel Boumghar; Philippe Suignard; | |
77 | Contextualized Query Embeddings for Conversational Search Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: This paper describes a compact and effective model for low-latency passage retrieval in conversational search based on learned dense representations. |
Sheng-Chieh Lin; Jheng-Hong Yang; Jimmy Lin; | |
78 | Ultra-High Dimensional Sparse Representations with Binarization for Efficient Text Retrieval Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: Taking the merits of the sparse and dense representations, we propose an ultra-high dimensional (UHD) representation scheme equipped with directly controllable sparsity. |
Kyoung-Rok Jang; Junmo Kang; Giwon Hong; Sung-Hyon Myaeng; Joohee Park; Taewon Yoon; Heecheol Seo; | |
79 | IR Like A SIR: Sense-enhanced Information Retrieval for Multiple Languages Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper we present SIR (Sense-enhanced Information Retrieval) to mitigate both problems by leveraging word sense information. |
Rexhina Blloshmi; Tommaso Pasini; Niccol? Campolungo; Somnath Banerjee; Roberto Navigli; Gabriella Pasi; | code |
80 | Neural Attention-Aware Hierarchical Topic Model Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: To address these issues, we propose a variational autoencoder (VAE) NTM model that jointly reconstructs the sentence and document word counts using combinations of bag-of-words (BoW) topical embeddings and pre-trained semantic embeddings. |
Yuan Jin; He Zhao; Ming Liu; Lan Du; Wray Buntine; | |
81 | Relational World Knowledge Representation in Contextual Language Models: A Review Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this review, we take a natural language processing perspective to these limitations, examining how they may be addressed in part by training deep contextual language models (LMs) to internalize and express relational knowledge in more flexible forms. |
Tara Safavi; Danai Koutra; | |
82 | Certified Robustness to Programmable Transformations in LSTMs Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We present an approach to certifying the robustness of LSTMs (and extensions of LSTMs) and training models that can be efficiently certified. |
Yuhao Zhang; Aws Albarghouthi; Loris D?Antoni; | |
83 | ReGen: Reinforcement Learning for Text and Knowledge Base Generation Using Pretrained Language Models Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we present ReGen, a bidirectional generation of text and graph leveraging Reinforcement Learning to improve performance. |
Pierre Dognin; Inkit Padhi; Igor Melnyk; Payel Das; | code |
84 | Contrastive Out-of-Distribution Detection for Pretrained Transformers Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we develop an unsupervised OOD detection method, in which only the in-distribution (ID) data are used in training. |
Wenxuan Zhou; Fangyu Liu; Muhao Chen; | |
85 | MindCraft: Theory of Mind Modeling for Situated Dialogue in Collaborative Tasks Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: To enable theory of mind modeling in situated interactions, we introduce a fine-grained dataset of collaborative tasks performed by pairs of human subjects in the 3D virtual blocks world of Minecraft. |
Cristian-Paul Bara; Sky CH-Wang; Joyce Chai; | |
86 | Detecting Speaker Personas from Conversational Texts Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: To tackle this issue, we study a new task, named Speaker Persona Detection (SPD), which aims to detect speaker personas based on the plain conversational text. |
Jia-Chen Gu; Zhenhua Ling; Yu Wu; Quan Liu; Zhigang Chen; Xiaodan Zhu; | |
87 | Cross-lingual Intermediate Fine-tuning Improves Dialogue State Tracking Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this work, we enhance the transfer learning process by intermediate fine-tuning of pretrained multilingual models, where the multilingual models are fine-tuned with different but related data and/or tasks. |
Nikita Moghe; Mark Steedman; Alexandra Birch; | |
88 | ConvFiT: Conversational Fine-Tuning of Pretrained Language Models Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this work, we propose ConvFiT, a simple and efficient two-stage procedure which turns any pretrained LM into a universal conversational encoder (after Stage 1 ConvFiT-ing) and task-specialised sentence encoder (after Stage 2). |
Ivan Vulic; Pei-Hao Su; Samuel Coope; Daniela Gerz; Pawel Budzianowski; I?igo Casanueva; Nikola Mrk?ic; Tsung-Hsien Wen; | |
89 | We’ve Had This Conversation Before: A Novel Approach to Measuring Dialog Similarity Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We propose a novel adaptation of the edit distance metric to the scenario of dialog similarity. |
Ofer Lavi; Ella Rabinovich; Segev Shlomov; David Boaz; Inbal Ronen; Ateret Anaby Tavor; | |
90 | Towards Incremental Transformers: An Empirical Analysis of Transformer Models for Incremental NLU Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this work, we examine the feasibility of LT for incremental NLU in English. |
Patrick Kahardipraja; Brielen Madureira; David Schlangen; | |
91 | Feedback Attribution for Counterfactual Bandit Learning in Multi-Domain Spoken Language Understanding Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we study the feedback attribution problem that arises when using counterfactual bandit learning for multi-domain spoken language understanding. |
Tobias Falke; Patrick Lehnen; | |
92 | Label Verbalization and Entailment for Effective Zero and Few-Shot Relation Extraction Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this work we reformulate relation extraction as an entailment task, with simple, hand-made, verbalizations of relations produced in less than 15 min per relation. |
Oscar Sainz; Oier Lopez de Lacalle; Gorka Labaka; Ander Barrena; Eneko Agirre; | |
93 | Extend, Don’t Rebuild: Phrasing Conditional Graph Modification As Autoregressive Sequence Labelling Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this work, we show that we can considerably increase performance on this problem by phrasing it as graph extension instead of graph generation. |
Leon Weber; Jannes M?nchmeyer; Samuele Garda; Ulf Leser; | |
94 | Zero-Shot Information Extraction As A Unified Text-to-Triple Translation Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We cast a suite of information extraction tasks into a text-to-triple translation framework. |
Chenguang Wang; Xiao Liu; Zui Chen; Haoyun Hong; Jie Tang; Dawn Song; | |
95 | Learning Logic Rules for Document-Level Relation Extraction Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: To tackle this challenge, in this paper, we propose LogiRE, a novel probabilistic model for document-level relation extraction by learning logic rules. |
Dongyu Ru; Changzhi Sun; Jiangtao Feng; Lin Qiu; Hao Zhou; Weinan Zhang; Yong Yu; Lei Li; | code |
96 | A Large-Scale Dataset for Empathetic Response Generation Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we describe a large-scale silver dataset consisting of 1M dialogues annotated with 32 fine-grained emotions, eight empathetic response intents, and the Neutral category. |
Anuradha Welivita; Yubo Xie; Pearl Pu; | |
97 | The Perils of Using Mechanical Turk to Evaluate Open-Ended Text Generation Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we first conduct a survey of 45 open-ended text generation papers and find that the vast majority of them fail to report crucial details about their AMT tasks, hindering reproducibility. We then run a series of story evaluation experiments with both AMT workers and English teachers and discover that even with strict qualification filters, AMT workers (unlike teachers) fail to distinguish between model-generated text and human-generated references. |
Marzena Karpinska; Nader Akoury; Mohit Iyyer; | |
98 | Documenting Large Webtext Corpora: A Case Study on The Colossal Clean Crawled Corpus Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this work we provide some of the first documentation for the Colossal Clean Crawled Corpus (C4; Raffel et al., 2020), a dataset created by applying a set of filters to a single snapshot of Common Crawl. |
Jesse Dodge; Maarten Sap; Ana Marasovic; William Agnew; Gabriel Ilharco; Dirk Groeneveld; Margaret Mitchell; Matt Gardner; | |
99 | AfroMT: Pretraining Strategies and Reproducible Benchmarks for Translation of 8 African Languages Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: To tackle these challenges, we propose AfroMT, a standardized, clean, and reproducible machine translation benchmark for eight widely spoken African languages. |
Machel Reid; Junjie Hu; Graham Neubig; Yutaka Matsuo; | |
100 | Evaluating The Evaluation Metrics for Style Transfer: A Case Study in Multilingual Formality Transfer Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we evaluate leading automatic metrics on the oft-researched task of formality style transfer. |
Eleftheria Briakou; Sweta Agrawal; Joel Tetreault; Marine Carpuat; | |
101 | MS-Mentions: Consistently Annotating Entity Mentions in Materials Science Procedural Text Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We present a new corpus of entity mention annotations over 595 Material Science synthesis procedural texts (157,488 tokens), which greatly expands the training data available for the Named Entity Recognition task. |
Tim O?Gorman; Zach Jensen; Sheshera Mysore; Kevin Huang; Rubayyat Mahbub; Elsa Olivetti; Andrew McCallum; | |
102 | Understanding Politics Via Contextualized Discourse Processing Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we propose a Compositional Reader model consisting of encoder and composer modules, that captures and leverages such information to generate more effective representations for entities, issues, and events. |
Rajkumar Pujari; Dan Goldwasser; | |
103 | Conundrums in Event Coreference Resolution: Making Sense of The State of The Art Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We present an empirical analysis of a state-of-the-art span-based event reference systems with the goal of providing the general NLP audience with a better understanding of the state of the art and reference researchers with directions for future research. |
Jing Lu; Vincent Ng; | |
104 | Weakly Supervised Discourse Segmentation for Multiparty Oral Conversations Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We develop a weak supervision approach to adapt, using minimal annotation, a state of the art discourse segmenter trained on written text to French conversation transcripts. |
Lila Gravellier; Julie Hunter; Philippe Muller; Thomas Pellegrini; Isabelle Ferran?; | |
105 | Narrative Embedding: Re-Contextualization Through Attention Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We present a novel approach for narrative event representation using attention to re-contextualize events across the whole story. |
Sean Wilner; Daniel Woolridge; Madeleine Glick; | |
106 | Focus on What Matters: Applying Discourse Coherence Theory to Cross Document Coreference Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We model the entities/events in a reader’s focus as a neighborhood within a learned latent embedding space which minimizes the distance between mentions and the centroids of their gold coreference clusters. We then use these neighborhoods to sample only hard negatives to train a fine-grained classifier on mention pairs and their local discourse features. |
William Held; Dan Iter; Dan Jurafsky; | |
107 | Salience-Aware Event Chain Modeling for Narrative Understanding Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We introduce methods for extracting this principal chain from natural language text, by filtering away non-salient events and supportive sentences. |
Xiyang Zhang; Muhao Chen; Jonathan May; | |
108 | Asking It All: Generating Contextualized Questions for Any Semantic Role Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: To this end, we introduce the task of role question generation, which, given a predicate mention and a passage, requires producing a set of questions asking about all possible semantic roles of the predicate. |
Valentina Pyatkin; Paul Roit; Julian Michael; Yoav Goldberg; Reut Tsarfaty; Ido Dagan; | |
109 | Fast, Effective, and Self-Supervised: Transforming Masked Language Models Into Universal Lexical and Sentence Encoders Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this work, we demonstrate that it is possible to turn MLMs into effective lexical and sentence encoders even without any additional data, relying simply on self-supervision. |
Fangyu Liu; Ivan Vulic; Anna Korhonen; Nigel Collier; | |
110 | RuleBERT: Teaching Soft Rules to Pre-Trained Language Models Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: Here, we aim to bridge this gap by teaching PLMs how to reason with soft Horn rules. |
Mohammed Saeed; Naser Ahmadi; Preslav Nakov; Paolo Papotti; | |
111 | Stepmothers Are Mean and Academics Are Pretentious: What Do Pretrained Language Models Learn About You? Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we investigate what types of stereotypical information are captured by pretrained language models. |
Rochelle Choenni; Ekaterina Shutova; Robert van Rooij; | |
112 | ConSeC: Word Sense Disambiguation As Continuous Sense Comprehension Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: To address this limitation and drop this assumption, we propose CONtinuous SEnse Comprehension (ConSeC), a novel approach to WSD: leveraging a recent re-framing of this task as a text extraction problem, we adapt it to our formulation and introduce a feedback loop strategy that allows the disambiguation of a target word to be conditioned not only on its context but also on the explicit senses assigned to nearby words. |
Edoardo Barba; Luigi Procopio; Roberto Navigli; | code |
113 | Shortcutted Commonsense: Data Spuriousness in Deep Learning of Commonsense Reasoning Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper we seek to further pursue this analysis into the realm of commonsense related language processing tasks. |
Ruben Branco; Ant?nio Branco; Jo?o Ant?nio Rodrigues; Jo?o Ricardo Silva; | |
114 | When Differential Privacy Meets NLP: The Devil Is in The Detail Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: Our contribution is a formal analysis of ADePT, a differentially private auto-encoder for text rewriting (Krishna et al, 2021). |
Ivan Habernal; | |
115 | Achieving Model Robustness Through Discrete Adversarial Training Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this work, we address this gap and leverage discrete attacks for online augmentation, where adversarial examples are generated at every training step, adapting to the changing nature of the model. |
Maor Ivgi; Jonathan Berant; | |
116 | Debiasing Methods in Natural Language Understanding Make Bias More Accessible Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We propose a general probing-based framework that allows for post-hoc interpretation of biases in language models, and use an information-theoretic approach to measure the extractability of certain biases from the model’s representations. |
Michael Mendelson; Yonatan Belinkov; | |
117 | Evaluating The Robustness of Neural Language Models to Input Perturbations Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this study, we design and implement various types of character-level and word-level perturbation methods to simulate realistic scenarios in which input texts may be slightly noisy or different from the data distribution on which NLP systems were trained. |
Milad Moradi; Matthias Samwald; | |
118 | How Much Pretraining Data Do Language Models Need to Learn Syntax? Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: This calls for a study of the impact of pretraining data size on the knowledge of the models. We explore this impact on the syntactic capabilities of RoBERTa, using models trained on incremental sizes of raw text data. |
Laura P?rez-Mayos; Miguel Ballesteros; Leo Wanner; | |
119 | Sorting Through The Noise: Testing Robustness of Information Processing in Pre-trained Language Models Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper we tackle a component of this question by examining robustness of models’ ability to deploy relevant context information in the face of distracting content. |
Lalchand Pandia; Allyson Ettinger; | |
120 | Contrastive Explanations for Model Interpretability Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We propose a method to produce contrastive explanations in the latent space, via a projection of the input representation, such that only the features that differentiate two potential decisions are captured. |
Alon Jacovi; Swabha Swayamdipta; Shauli Ravfogel; Yanai Elazar; Yejin Choi; Yoav Goldberg; | |
121 | On The Transferability of Adversarial Attacks Against Neural Text Classifier Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: Based on these studies, we propose a genetic algorithm to find an ensemble of models that can be used to induce adversarial examples to fool almost all existing models. |
Liping Yuan; Xiaoqing Zheng; Yi Zhou; Cho-Jui Hsieh; Kai-Wei Chang; | |
122 | Conditional Probing: Measuring Usable Information Beyond A Baseline Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this work, we extend a theory of usable information called V-information and propose conditional probing, which explicitly conditions on the information in the baseline. |
John Hewitt; Kawin Ethayarajh; Percy Liang; Christopher Manning; | |
123 | GFST: Gender-Filtered Self-Training for More Accurate Gender in Translation Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We propose gender-filtered self-training (GFST) to improve gender translation accuracy on unambiguously gendered inputs. |
Prafulla Kumar Choubey; Anna Currey; Prashant Mathur; Georgiana Dinu; | |
124 | Wikily Supervised Neural Translation Tailored to Cross-Lingual Tasks Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We present a simple but effective approach for leveraging Wikipedia for neural machine translation as well as cross-lingual tasks of image captioning and dependency parsing without using any direct supervision from external parallel data or supervised models in the target language. |
Mohammad Sadegh Rasooli; Chris Callison-Burch; Derry Tanti Wijaya; | |
125 | MT6: Multilingual Pretrained Text-to-Text Transformer with Translation Pairs Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we improve multilingual text-to-text transfer Transformer with translation pairs (mT6). |
Zewen Chi; Li Dong; Shuming Ma; Shaohan Huang; Saksham Singhal; Xian-Ling Mao; Heyan Huang; Xia Song; Furu Wei; | |
126 | Improving Zero-Shot Cross-Lingual Transfer Learning Via Robust Training Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this work, we propose a learning strategy for training robust models by drawing connections between adversarial examples and the failure cases of zero-shot cross-lingual transfer. |
Kuan-Hao Huang; Wasi Ahmad; Nanyun Peng; Kai-Wei Chang; | |
127 | Speechformer: Reducing Information Loss in Direct Speech Translation Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: To solve this issue, we propose Speechformer, an architecture that, thanks to reduced memory usage in the attention layers, avoids the initial lossy compression and aggregates information only at a higher level according to more informed linguistic criteria. |
Sara Papi; Marco Gaido; Matteo Negri; Marco Turchi; | |
128 | Is moby Dick A Whale or A Bird? Named Entities and Terminology in Speech Translation Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: To fill this gap, we i) present the first systematic analysis of the behavior of state-of-the-art ST systems in translating NEs and terminology, and ii) release NEuRoparl-ST, a novel benchmark built from European Parliament speeches annotated with NEs and terminology. |
Marco Gaido; Susana Rodr?guez; Matteo Negri; Luisa Bentivogli; Marco Turchi; | |
129 | HintedBT: Augmenting Back-Translation with Quality and Transliteration Hints Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: To improve effectiveness of the available BT data, we introduce HintedBT-a family of techniques which provides hints (through tags) to the encoder and decoder. |
Sahana Ramnath; Melvin Johnson; Abhirut Gupta; Aravindan Raghuveer; | |
130 | Translation-based Supervision for Policy Generation in Simultaneous Neural Machine Translation Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We propose a novel supervised learning approach for training an agent that can detect the minimum number of reads required for generating each target token by comparing simultaneous translations against full-sentence translations during training to generate oracle action sequences. |
Ashkan Alinejad; Hassan S. Shavarani; Anoop Sarkar; | |
131 | Nearest Neighbour Few-Shot Learning for Cross-lingual Classification Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this work, we investigate cross-lingual adaptation using a simple nearest-neighbor few-shot (<15 samples) inference technique for classification tasks. |
M Saiful Bari; Batool Haider; Saab Mansour; | |
132 | Cross-Attention Is All You Need: Adapting Pretrained Transformers for Machine Translation Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We study the power of cross-attention in the Transformer architecture within the context of transfer learning for machine translation, and extend the findings of studies into cross-attention when training from scratch. |
Mozhdeh Gheini; Xiang Ren; Jonathan May; | |
133 | Effects of Parameter Norm Growth During Transformer Training: Inductive Bias from Gradient Descent Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: To better understand this bias, we study the tendency for transformer parameters to grow in magnitude (l2 norm) during training, and its implications for the emergent representations within self attention layers. |
William Merrill; Vivek Ramanujan; Yoav Goldberg; Roy Schwartz; Noah A. Smith; | |
134 | Foreseeing The Benefits of Incidental Supervision Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We propose a unified PAC-Bayesian motivated informativeness measure, PABI, that characterizes the uncertainty reduction provided by incidental supervision signals. |
Hangfeng He; Mingyuan Zhang; Qiang Ning; Dan Roth; | |
135 | Competency Problems: On Finding and Removing Artifacts in Language Data Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this work we argue that for complex language understanding tasks, all simple feature correlations are spurious, and we formalize this notion into a class of problems which we call competency problems. |
Matt Gardner; William Merrill; Jesse Dodge; Matthew Peters; Alexis Ross; Sameer Singh; Noah A. Smith; | |
136 | Knowledge-Aware Meta-learning for Low-Resource Text Classification Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: This paper studies a low-resource text classification problem and bridges the gap between meta-training and meta-testing tasks by leveraging the external knowledge bases. |
Huaxiu Yao; Ying-xin Wu; Maruan Al-Shedivat; Eric Xing; | |
137 | Sentence Bottleneck Autoencoders from Transformer Language Models Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We therefore explore the construction of a sentence-level autoencoder from a pretrained, frozen transformer language model. |
Ivan Montero; Nikolaos Pappas; Noah A. Smith; | |
138 | Efficient Contrastive Learning Via Novel Data Augmentation and Curriculum Learning Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We introduce EfficientCL, a memory-efficient continual pretraining method that applies contrastive learning with novel data augmentation and curriculum learning. |
Seonghyeon Ye; Jiseon Kim; Alice Oh; | |
139 | CR-Walker: Tree-Structured Graph Reasoning and Dialog Acts for Conversational Recommendation Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: To address these issues, we propose CR-Walker in this paper, a model that performs tree-structured reasoning on a knowledge graph, and generates informative dialog acts to guide language generation. |
Wenchang Ma; Ryuichi Takanobu; Minlie Huang; | |
140 | DIALKI: Knowledge Identification in Conversational Systems Through Dialogue-Document Contextualization Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We introduce a knowledge identification model that leverages the document structure to provide dialogue-contextualized passage encodings and better locate knowledge relevant to the conversation. |
Zeqiu Wu; Bo-Ru Lu; Hannaneh Hajishirzi; Mari Ostendorf; | |
141 | Iconary: A Pictionary-Based Game for Testing Multimodal Communication with Drawings and Text Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We propose models to play Iconary and train them on over 55,000 games between human players. |
Christopher Clark; Jordi Salvador; Dustin Schwenk; Derrick Bonafilia; Mark Yatskar; Eric Kolve; Alvaro Herrasti; Jonghyun Choi; Sachin Mehta; Sam Skjonsberg; Carissa Schoenick; Aaron Sarnat; Hannaneh Hajishirzi; Aniruddha Kembhavi; Oren Etzioni; Ali Farhadi; | |
142 | Self-training Improves Pre-training for Few-shot Learning in Task-oriented Dialog Systems Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: Specifically, we propose a self-training approach that iteratively labels the most confident unlabeled data to train a stronger Student model. |
Fei Mi; Wanhao Zhou; Lingjing Kong; Fengyu Cai; Minlie Huang; Boi Faltings; | |
143 | Contextual Rephrase Detection for Reducing Friction in Dialogue Systems Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: To this end, we propose a contextual rephrase detection model ContReph to automatically identify rephrases from multi-turn dialogues. |
Zhuoyi Wang; Saurabh Gupta; Jie Hao; Xing Fan; Dingcheng Li; Alexander Hanbo Li; Chenlei Guo; | |
144 | Few-Shot Intent Detection Via Contrastive Pre-Training and Fine-Tuning Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this work, we focus on a more challenging few-shot intent detection scenario where many intents are fine-grained and semantically similar. |
Jianguo Zhang; Trung Bui; Seunghyun Yoon; Xiang Chen; Zhiwei Liu; Congying Xia; Quan Hung Tran; Walter Chang; Philip Yu; | |
145 | It Doesn’t Look Good for A Date: Transforming Critiques Into Preferences for Conversational Recommendation Systems Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this work, we present a method for transforming a user critique into a positive preference (e.g., I prefer more romantic) in order to retrieve reviews pertaining to potentially better recommendations (e.g., Perfect for a romantic dinner). |
Victor Bursztyn; Jennifer Healey; Nedim Lipka; Eunyee Koh; Doug Downey; Larry Birnbaum; | |
146 | AttentionRank: Unsupervised Keyphrase Extraction Using Self and Cross Attentions Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: This paper proposes the AttentionRank, a hybrid attention model, to identify keyphrases from a document in an unsupervised manner. |
Haoran Ding; Xiao Luo; | |
147 | Unsupervised Relation Extraction: A Variational Autoencoder Approach Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We propose a VAE-based unsupervised relation extraction technique that overcomes this limitation by using the classifications as an intermediate variable instead of a latent variable. |
Chenhan Yuan; Hoda Eldardiry; | |
148 | Robust Retrieval Augmented Generation for Zero-shot Slot Filling Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we present a novel approach to zero-shot slot filling that extends dense passage retrieval with hard negatives and robust training procedures for retrieval augmented generation models. |
Michael Glass; Gaetano Rossiello; Md Faisal Mahbub Chowdhury; Alfio Gliozzo; | |
149 | Everything Is All It Takes: A Multipronged Strategy for Zero-Shot Cross-Lingual Information Extraction Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: While the advance of pretrained multilingual encoders suggests an easy optimism of train on English, run on any language, we find through a thorough exploration and extension of techniques that a combination of approaches, both new and old, leads to better performance than any one cross-lingual strategy in particular. |
Mahsa Yarmohammadi; Shijie Wu; Marc Marone; Haoran Xu; Seth Ebner; Guanghui Qin; Yunmo Chen; Jialiang Guo; Craig Harman; Kenton Murray; Aaron Steven White; Mark Dredze; Benjamin Van Durme; | |
150 | Harms of Gender Exclusivity and Challenges in Non-Binary Representation in Language Technologies Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we explain the complexity of gender and language around it, and survey non-binary persons to understand harms associated with the treatment of gender as binary in English language technologies. |
Sunipa Dev; Masoud Monajatipoor; Anaelia Ovalle; Arjun Subramonian; Jeff Phillips; Kai-Wei Chang; | |
151 | Are Gender-Neutral Queries Really Gender-Neutral? Mitigating Gender Bias in Image Search Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: Therefore, we introduce two novel debiasing approaches: an in-processing fair sampling method to address the gender imbalance issue for training models, and a post-processing feature clipping method base on mutual information to debias multimodal representations of pre-trained models. |
Jialu Wang; Yang Liu; Xin Wang; | |
152 | Style Pooling: Automatic Text Style Obfuscation for Improved Classification Fairness Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We propose a VAE-based framework that obfuscates stylistic features of human-generated text through style transfer, by automatically re-writing the text itself. |
Fatemehsadat Mireshghallah; Taylor Berg-Kirkpatrick; | |
153 | Modeling Disclosive Transparency in NLP Application Descriptions Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: To improve this state of affairs, We introduce neural language model-based probabilistic metrics to directly model disclosive transparency, and demonstrate that they correlate with user and expert opinions of system transparency, making them a valid objective proxy. |
Michael Saxon; Sharon Levy; Xinyi Wang; Alon Albalak; William Yang Wang; | |
154 | Reconstruction Attack on Instance Encoding for Language Understanding Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We propose a novel reconstruction attack to break TextHide by recovering the private training data, and thus unveil the privacy risks of instance encoding. |
Shangyu Xie; Yuan Hong; | |
155 | Fairness-aware Class Imbalanced Learning Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this work we evaluate long-tail learning methods for tweet sentiment and occupation classification, and extend a margin-loss based approach with methods to enforce fairness. |
Shivashankar Subramanian; Afshin Rahimi; Timothy Baldwin; Trevor Cohn; Lea Frermann; | |
156 | CRYPTOGRU: Low Latency Privacy-Preserving Text Analysis With GRU Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we present a novel hybrid structure of HE and GC gated recurrent unit (GRU) network, , for low-latency secure inferences. |
Bo Feng; Qian Lou; Lei Jiang; Geoffrey Fox; | |
157 | Local Word Discovery for Interactive Transcription Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: Accordingly, we propose a new computational task which is tuned to the available knowledge and interests in an Indigenous community, and which supports the construction of high quality texts and lexicons. |
William Lane; Steven Bird; | |
158 | Segment, Mask, and Predict: Augmenting Chinese Word Segmentation with Self-Supervision Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this work, we propose a self-supervised CWS approach with a straightforward and effective architecture. |
Mieradilijiang Maimaiti; Yang Liu; Yuanhang Zheng; Gang Chen; Kaiyu Huang; Ji Zhang; Huanbo Luan; Maosong Sun; | |
159 | Minimal Supervision for Morphological Inflection Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this work we aim to overcome this annotation bottleneck by bootstrapping labeled data from a seed as small as five labeled inflection tables, accompanied by a large bulk of unlabeled text. |
Omer Goldman; Reut Tsarfaty; | |
160 | Fast WordPiece Tokenization Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we propose efficient algorithms for the WordPiece tokenization used in BERT, from single-word tokenization to general text (e.g., sentence) tokenization. |
Xinying Song; Alex Salcianu; Yang Song; Dave Dopson; Denny Zhou; | |
161 | You Should Evaluate Your Language Model on Marginal Likelihood Over Tokenisations Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we argue that instead, language models should be evaluated on their marginal likelihood over tokenisations. |
Kris Cao; Laura Rimell; | |
162 | Broaden The Vision: Geo-Diverse Visual Commonsense Reasoning Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we construct a Geo-Diverse Visual Commonsense Reasoning dataset (GD-VCR) to test vision-and-language models’ ability to understand cultural and geo-location-specific commonsense. |
Da Yin; Liunian Harold Li; Ziniu Hu; Nanyun Peng; Kai-Wei Chang; | code |
163 | Reference-Centric Models for Grounded Collaborative Dialogue Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We present a grounded neural dialogue model that successfully collaborates with people in a partially-observable reference game. |
Daniel Fried; Justin Chiu; Dan Klein; | |
164 | CrossVQA: Scalably Generating Benchmarks for Systematically Testing VQA Generalization Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we propose a semi-automatic framework for generating disentangled shifts by introducing a controllable visual question-answer generation (VQAG) module that is capable of generating highly-relevant and diverse question-answer pairs with the desired dataset style. |
Arjun Akula; Soravit Changpinyo; Boqing Gong; Piyush Sharma; Song-Chun Zhu; Radu Soricut; | |
165 | Visual Goal-Step Inference Using WikiHow Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We propose the Visual Goal-Step Inference (VGSI) task, where a model is given a textual goal and must choose which of four images represents a plausible step towards that goal. |
Yue Yang; Artemis Panagopoulou; Qing Lyu; Li Zhang; Mark Yatskar; Chris Callison-Burch; | |
166 | Systematic Generalization on GSCAN: What Is Nearly Solved and What Is Next? Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We find that a general-purpose Transformer-based model with cross-modal attention achieves strong performance on a majority of the gSCAN splits, surprisingly outperforming more specialized approaches from prior work. Furthermore, our analysis suggests that many of the remaining errors reveal the same fundamental challenge in systematic generalization of linguistic constructs regardless of visual context. Second, inspired by this finding, we propose challenging new tasks for gSCAN by generating data to incorporate relations between objects in the visual environment. |
Linlu Qiu; Hexiang Hu; Bowen Zhang; Peter Shaw; Fei Sha; | |
167 | Effect of Visual Extensions on Natural Language Understanding in Vision-and-Language Models Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: To see how well this is achieved, we propose to evaluate V&L models using an NLU benchmark (GLUE). |
Taichi Iki; Akiko Aizawa; | |
168 | Neural Path Hunter: Reducing Hallucination in Dialogue Systems Via Path Grounding Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we focus on the task of improving faithfulness and reducing hallucination of neural dialogue systems to known facts supplied by a Knowledge Graph (KG). |
Nouha Dziri; Andrea Madotto; Osmar Za?ane; Avishek Joey Bose; | code |
169 | Thinking Clearly, Talking Fast: Concept-Guided Non-Autoregressive Generation for Open-Domain Dialogue Systems Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: To facilitate a controllable and coherent dialogue, in this work, we devise a concept-guided non-autoregressive model (CG-nAR) for open-domain dialogue generation. |
Yicheng Zou; Zhihua Liu; Xingwu Hu; Qi Zhang; | |
170 | Perspective-taking and Pragmatics for Generating Empathetic Responses Focused on Emotion Causes Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: Taking inspiration from social cognition, we leverage a generative estimator to infer emotion cause words from utterances with no word-level label. Also, we introduce a novel method based on pragmatics to make dialogue models focus on targeted words in the input during generation. |
Hyunwoo Kim; Byeongchang Kim; Gunhee Kim; | |
171 | Generation and Extraction Combined Dialogue State Tracking with Hierarchical Ontology Integration Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: To address the problem, we explore the hierarchical semantic of ontology and enhance the interrelation between slots with masked hierarchical attention. |
Xinmeng Li; Qian Li; Wansen Wu; Quanjun Yin; | |
172 | CoLV: A Collaborative Latent Variable Model for Knowledge-Grounded Dialogue Generation Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, in order to improve the diversity of both knowledge selection and knowledge-aware response generation, we propose a collaborative latent variable (CoLV) model to integrate these two aspects simultaneously in separate yet collaborative latent spaces, so as to capture the inherent correlation between knowledge selection and response generation. |
Haolan Zhan; Lei Shen; Hongshen Chen; Hainan Zhang; | |
173 | A Three-Stage Learning Framework for Low-Resource Knowledge-Grounded Dialogue Generation Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we propose a novel three-stage learning framework based on weakly supervised learning which benefits from large scale ungrounded dialogues and unstructured knowledge base. |
Shilei Liu; Xiaofeng Zhao; Bochao Li; Feiliang Ren; Longhui Zhang; Shujuan Yin; | |
174 | Intention Reasoning Network for Multi-Domain End-to-end Task-Oriented Dialogue Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: To address these issues, we propose a novel intention mechanism to better model deterministic entity knowledge. |
Zhiyuan Ma; Jianjun Li; Zezheng Zhang; Guohui Li; Yongjing Cheng; | |
175 | More Is Better: Enhancing Open-Domain Dialogue Generation Via Multi-Source Heterogeneous Knowledge Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: This paper proposes a novel dialogue generation model, MSKE-Dialog, to solve this issue with three unique advantages: (1) Rather than only one, MSKE-Dialog can simultaneously leverage multiple heterogeneous knowledge sources (it includes but is not limited to commonsense knowledge facts, text knowledge, infobox knowledge) to improve the knowledge coverage; (2) To avoid the topic conflict among the context and different knowledge sources, we propose a Multi-Reference Selection to better select context/knowledge; (3) We propose a Multi-Reference Generation to generate informative responses by referring to multiple generation references at the same time. |
Sixing Wu; Ying Li; Minghui Wang; Dawei Zhang; Yang Zhou; Zhonghai Wu; | |
176 | Domain-Lifelong Learning for Dialogue State Tracking Via Knowledge Preservation Networks Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: To this end, we propose a novel domain-lifelong learning method, called Knowledge Preservation Networks (KPN), which consists of multi-prototype enhanced retrospection and multi-strategy knowledge distillation, to solve the problems of expression diversity and combinatorial explosion in the DLL-DST task. |
Qingbin Liu; Pengfei Cao; Cao Liu; Jiansong Chen; Xunliang Cai; Fan Yang; Shizhu He; Kang Liu; Jun Zhao; | |
177 | CSAGN: Conversational Structure Aware Graph Network for Conversational Semantic Role Labeling Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we present a simple and effective architecture for CSRL which aims to address this problem. |
Han Wu; Kun Xu; Linqi Song; | |
178 | Different Strokes for Different Folks: Investigating Appropriate Further Pre-training Approaches for Diverse Dialogue Tasks Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: To investigate this, we carry out a study for improving multiple task-oriented dialogue downstream tasks through designing various tasks at the further pre-training phase. |
Yao Qiu; Jinchao Zhang; Jie Zhou; | |
179 | Knowledge Enhanced Fine-Tuning for Better Handling Unseen Entities in Dialogue Generation Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: Specifically, with the help of a knowledge base, we introduce two auxiliary training objectives: 1) Interpret Masked Word, which conjectures the meaning of the masked entity given the context; 2) Hypernym Generation, which predicts the hypernym of the entity based on the context. |
Leyang Cui; Yu Wu; Shujie Liu; Yue Zhang; | |
180 | An Evaluation Dataset and Strategy for Building Robust Multi-turn Response Selection Model Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this study, we analyze the weaknesses of the open-domain Korean Multi-turn response selection models and publish an adversarial dataset to evaluate these weaknesses. |
Kijong Han; Seojin Lee; Dong-hun Lee; | |
181 | Unsupervised Conversation Disentanglement Through Co-Training Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this work, we explore training a conversation disentanglement model without referencing any human annotations. |
Hui Liu; Zhan Shi; Xiaodan Zhu; | |
182 | Don’t Be Contradicted with Anything! CI-ToD: Towards Benchmarking Consistency for Task-oriented Dialogue System Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we argue that consistency problem is more urgent in task-oriented domain. |
Libo Qin; Tianbao Xie; Shijue Huang; Qiguang Chen; Xiao Xu; Wanxiang Che; | |
183 | Transferable Persona-Grounded Dialogues Via Grounded Minimal Edits Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: To address the challenges, we propose the grounded minimal editing framework, which minimally edits existing responses to be grounded on the given concept. |
Chen Henry Wu; Yinhe Zheng; Xiaoxi Mao; Minlie Huang; | |
184 | EARL: Informative Knowledge-Grounded Conversation Generation with Entity-Agnostic Representation Learning Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we propose an Entity-Agnostic Representation Learning (EARL) method to introduce knowledge graphs to informative conversation generation. |
Hao Zhou; Minlie Huang; Yong Liu; Wei Chen; Xiaoyan Zhu; | |
185 | DialogueCSE: Dialogue-based Contrastive Learning of Sentence Embeddings Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we propose DialogueCSE, a dialogue-based contrastive learning approach to tackle this issue. |
Che Liu; Rui Wang; Jinghua Liu; Jian Sun; Fei Huang; Luo Si; | |
186 | Improving Graph-based Sentence Ordering with Iteratively Predicted Pairwise Orderings Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we propose a novel sentence ordering framework which introduces two classifiers to make better use of pairwise orderings for graph-based sentence ordering (Yin et al. 2019, 2021). |
Shaopeng Lai; Ante Wang; Fandong Meng; Jie Zhou; Yubin Ge; Jiali Zeng; Junfeng Yao; Degen Huang; Jinsong Su; | code |
187 | Not Just Classification: Recognizing Implicit Discourse Relation on Joint Modeling of Classification and Generation Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: Specifically, we propose a joint model, CG-T5, to recognize the relation label and generate the target sentence containing the meaning of relations simultaneously. |
Feng Jiang; Yaxin Fan; Xiaomin Chu; Peifeng Li; Qiaoming Zhu; | |
188 | A Language Model-based Generative Classifier for Sentence-level Discourse Parsing Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: To solve the problem, we propose a language model-based generative classifier (LMGC) for using more information from labels by treating the labels as an input while enhancing label representations by embedding descriptions for each label. |
Ying Zhang; Hidetaka Kamigaito; Manabu Okumura; | |
189 | Multimodal Phased Transformer for Sentiment Analysis Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We propose multimodal Sparse Phased Transformer (SPT) to alleviate the problem of self-attention complexity and memory footprint. |
Junyan Cheng; Iordanis Fostiropoulos; Barry Boehm; Mohammad Soleymani; | |
190 | Hierarchical Multi-label Text Classification with Horizontal and Vertical Category Correlations Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we propose a novel HMTC framework that considers both vertical and horizontal category correlations. |
Linli Xu; Sijie Teng; Ruoyu Zhao; Junliang Guo; Chi Xiao; Deqiang Jiang; Bo Ren; | |
191 | RankNAS: Efficient Neural Architecture Search By Pairwise Ranking Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: This paper addresses the efficiency challenge of Neural Architecture Search (NAS) by formulating the task as a ranking problem. |
Chi Hu; Chenglong Wang; Xiangnan Ma; Xia Meng; Yinqiao Li; Tong Xiao; Jingbo Zhu; Changliang Li; | |
192 | FLiText: A Faster and Lighter Semi-Supervised Text Classification with Convolution Networks Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this work, we develop a new SSL framework called FLiText, which stands for Faster and Lighter semi-supervised Text classification. |
Chen Liu; Zhang Mengchao; Fu Zhibing; Panpan Hou; Yu Li; | |
193 | Evaluating Debiasing Techniques for Intersectional Biases Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper we argue that a truly fair model must consider ‘gerrymandering’ groups which comprise not only single attributes, but also intersectional groups. |
Shivashankar Subramanian; Xudong Han; Timothy Baldwin; Trevor Cohn; Lea Frermann; | |
194 | Definition Modelling for Appropriate Specificity Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: Herein, we propose a method for definition generation with appropriate specificity. |
Han Huang; Tomoyuki Kajiwara; Yuki Arase; | |
195 | Transductive Learning for Unsupervised Text Style Transfer Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: To tackle this problem, we propose a novel transductive learning approach in this paper, based on a retrieval-based context-aware style representation. |
Fei Xiao; Liang Pang; Yanyan Lan; Yan Wang; Huawei Shen; Xueqi Cheng; | |
196 | Integrating Semantic Scenario and Word Relations for Abstractive Sentence Summarization Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: To simultaneously capture the word relations and structure information from sentences, we propose a novel Dual Graph network for Abstractive Sentence Summarization. |
Yong Guan; Shaoru Guo; Ru Li; Xiaoli Li; Hu Zhang; | |
197 | Coupling Context Modeling with Zero Pronoun Recovering for Document-Level Natural Language Generation Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this case, we propose a highly adaptive two-stage approach to couple context modeling with ZP recovering to mitigate the ZP problem in NLG tasks. |
Xin Tan; Longyin Zhang; Guodong Zhou; | |
198 | Adaptive Bridge Between Training and Inference for Dialogue Generation Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we propose a novel adaptive switching mechanism, which learns to automatically transit between ground-truth learning and generated learning regarding the word-level matching score, such as the cosine similarity. |
Haoran Xu; Hainan Zhang; Yanyan Zou; Hongshen Chen; Zhuoye Ding; Yanyan Lan; | |
199 | ConRPG: Paraphrase Generation Using Contexts As Regularizer Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we propose a new unsupervised paradigm for paraphrase generation based on the assumption that the probabilities of generating two sentences with the same meaning given the same context should be the same. |
Yuxian Meng; Xiang Ao; Qing He; Xiaofei Sun; Qinghong Han; Fei Wu; Chun Fan; Jiwei Li; | |
200 | Building The Directed Semantic Graph for Coherent Long Text Generation Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: To address aforementioned research issue, this paper proposes a novel two-stage approach to generate coherent long text. |
Ziao Wang; Xiaofeng Zhang; Hongwei Du; | |
201 | Iterative GNN-based Decoder for Question Generation Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: To address these problems, we design the Iterative Graph Network-based Decoder (IGND) to model the previous generation using a Graph Neural Network at each decoding step. |
Zichu Fei; Qi Zhang; Yaqian Zhou; | |
202 | Asking Questions Like Educational Experts: Automatically Generating Question-Answer Pairs on Real-World Examination Data Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: This paper for the first time addresses the question-answer pair generation task on the real-world examination data, and proposes a new unified framework on RACE. |
Fanyi Qu; Xin Jia; Yunfang Wu; | |
203 | Syntactically-Informed Unsupervised Paraphrasing with Non-Parallel Data Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we take this research direction to the extreme and investigate whether it is possible to learn syntactically controlled paraphrase generation with nonparallel data. |
Erguang Yang; Mingtong Liu; Deyi Xiong; Yujie Zhang; Yao Meng; Changjian Hu; Jinan Xu; Yufeng Chen; | |
204 | Exploring Task Difficulty for Few-Shot Relation Extraction Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we introduce a novel approach based on contrastive learning that learns better representations by exploiting relation label information. |
Jiale Han; Bo Cheng; Wei Lu; | |
205 | MuVER: Improving First-Stage Entity Retrieval with Multi-View Entity Representations Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this work, we propose Multi-View Entity Representations (MuVER), a novel approach for entity retrieval that constructs multi-view representations for entity descriptions and approximates the optimal view for mentions via a heuristic searching method. |
Xinyin Ma; Yong Jiang; Nguyen Bach; Tao Wang; Zhongqiang Huang; Fei Huang; Weiming Lu; | |
206 | Treasures Outside Contexts: Improving Event Detection Via Global Statistics Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: Specifically, we propose a Semantic and Statistic-Joint Discriminative Network (SS-JDN) consisting of a semantic feature extractor, a statistical feature extractor, and a joint event discriminator. |
Rui Li; Wenlin Zhao; Cheng Yang; Sen Su; | |
207 | Uncertain Local-to-Global Networks for Document-Level Event Factuality Identification Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we propose an Uncertain Local-to-Global Network (ULGN) to make use of these two characteristics. |
Pengfei Cao; Yubo Chen; Yuqing Yang; Kang Liu; Jun Zhao; | |
208 | A Novel Global Feature-Oriented Relational Triple Extraction Model Based on Table Filling Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: To overcome this deficiency, we propose a global feature-oriented triple extraction model that makes full use of the mentioned two kinds of global associations. |
Feiliang Ren; Longhui Zhang; Shujuan Yin; Xiaofeng Zhao; Shilei Liu; Bochao Li; Yaduo Liu; | code |
209 | Structure-Augmented Keyphrase Generation Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: Our contribution is generating/augmenting structure then injecting these information in the encoding, using existing keyphrases of other documents, complementing missing/incomplete titles. |
Jihyuk Kim; Myeongho Jeong; Seungtaek Choi; Seung-won Hwang; | |
210 | An Empirical Study on Multiple Information Sources for Zero-Shot Fine-Grained Entity Typing Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we empirically study three kinds of auxiliary information: context consistency, type hierarchy and background knowledge (e.g., prototypes and descriptions) of types, and propose a multi-source fusion model (MSF) targeting these sources. |
Yi Chen; Haiyun Jiang; Lemao Liu; Shuming Shi; Chuang Fan; Min Yang; Ruifeng Xu; | |
211 | DyLex: Incorporating Dynamic Lexicons Into BERT for Sequence Labeling Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we propose DyLex, a plug-in lexicon incorporation approach for BERT based sequence labeling tasks. |
Baojun Wang; Zhao Zhang; Kun Xu; Guang-Yuan Hao; Yuyang Zhang; Lifeng Shang; Linlin Li; Xiao Chen; Xin Jiang; Qun Liu; | |
212 | MapRE: An Effective Semantic Mapping Approach for Low-resource Relation Extraction Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this work, we propose a framework considering both label-agnostic and label-aware semantic mapping information for low resource relation extraction. |
Manqing Dong; Chunguang Pan; Zhipeng Luo; | |
213 | Heterogeneous Graph Neural Networks for Keyphrase Generation Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: To address these problems, we propose a novel graph-based method that can capture explicit knowledge from related references. |
Jiacheng Ye; Ruijian Cai; Tao Gui; Qi Zhang; | |
214 | Machine Reading Comprehension As Data Augmentation: A Case Study on Implicit Event Argument Extraction Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we take a new perspective to address the data sparsity issue faced by implicit EAE, by bridging the task with machine reading comprehension (MRC). |
Jian Liu; Yufeng Chen; Jinan Xu; | |
215 | Importance Estimation from Multiple Perspectives for Keyphrase Extraction Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we propose a new approach to estimate the importance of keyphrase from multiple perspectives (called as KIEMP) and further improve the performance of keyphrase extraction. |
Mingyang Song; Liping Jing; Lin Xiao; | |
216 | Gradient Imitation Reinforcement Learning for Low Resource Relation Extraction Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: To alleviate selection bias due to the lack of feedback loops in existing LRE learning paradigms, we developed a Gradient Imitation Reinforcement Learning method to encourage pseudo label data to imitate the gradient descent direction on labeled data and bootstrap its optimization capability through trial and error. |
Xuming Hu; Chenwei Zhang; Yawen Yang; Xiaohe Li; Li Lin; Lijie Wen; Philip S. Yu; | |
217 | Low-resource Taxonomy Enrichment with Pretrained Language Models Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: To tackle the problem of low-resource taxonomy enrichment, we propose Musubu, an efficient framework for taxonomy enrichment in low-resource settings with pretrained language models (LMs) as knowledge bases to compensate for the shortage of information. |
Kunihiro Takeoka; Kosuke Akimoto; Masafumi Oyamada; | |
218 | Entity Relation Extraction As Dependency Parsing in Visually Rich Documents Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we adapt the popular dependency parsing model, the biaffine parser, to this entity relation extraction task. |
Yue Zhang; Zhang Bo; Rui Wang; Junjie Cao; Chen Li; Zuyi Bao; | |
219 | Synchronous Dual Network with Cross-Type Attention for Joint Entity and Relation Extraction Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we design a novel synchronous dual network (SDN) with cross-type attention via separately and interactively considering the entity types and relation types. |
Hui Wu; Xiaodong Shi; | |
220 | Less Is More: Pretrain A Strong Siamese Encoder for Dense Text Retrieval Using A Weak Decoder Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: To address this, we propose a new self-learning method that pre-trains the autoencoder using a weak decoder, with restricted capacity and attention flexibility to push the encoder to provide better text representations. |
Shuqi Lu; Di He; Chenyan Xiong; Guolin Ke; Waleed Malik; Zhicheng Dou; Paul Bennett; Tie-Yan Liu; Arnold Overwijk; | code |
221 | TransPrompt: Towards An Automatic Transferable Prompting Framework for Few-shot Text Classification Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: Based on continuous prompt embeddings, we propose TransPrompt, a transferable prompting framework for few-shot learning across similar tasks. |
Chengyu Wang; Jianing Wang; Minghui Qiu; Jun Huang; Ming Gao; | |
222 | Weakly-supervised Text Classification Based on Keyword Graph Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we propose a novel framework called ClassKG to explore keyword-keyword correlation on keyword graph by GNN. |
Lu Zhang; Jiandong Ding; Yi Xu; Yingyao Liu; Shuigeng Zhou; | |
223 | Efficient-FedRec: Efficient Federated Learning Framework for Privacy-Preserving News Recommendation Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we propose an efficient federated learning framework for privacy-preserving news recommendation. |
Jingwei Yi; Fangzhao Wu; Chuhan Wu; Ruixuan Liu; Guangzhong Sun; Xing Xie; | |
224 | RocketQAv2: A Joint Training Method for Dense Passage Retrieval and Passage Re-ranking Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we propose a novel joint training approach for dense passage retrieval and passage reranking. |
Ruiyang Ren; Yingqi Qu; Jing Liu; Wayne Xin Zhao; QiaoQiao She; Hua Wu; Haifeng Wang; Ji-Rong Wen; | code |
225 | Dealing with Typos for BERT-based Passage Retrieval and Ranking Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper we consider the Dense Retriever (DR), a passage retrieval method, and the BERT re-ranker, a popular passage re-ranking method. |
Shengyao Zhuang; Guido Zuccon; | |
226 | From Alignment to Assignment: Frustratingly Simple Unsupervised Entity Alignment Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: Based on this re-definition, we propose a frustratingly Simple but Effective Unsupervised entity alignment method (SEU) without neural networks. |
Xin Mao; Wenting Wang; Yuanbin Wu; Man Lan; | |
227 | Simple and Effective Unsupervised Redundancy Elimination to Compress Dense Vectors for Passage Retrieval Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we analyze the redundancy present in encoded dense vectors and show that the default dimension of 768 is unnecessarily large. |
Xueguang Ma; Minghan Li; Kai Sun; Ji Xin; Jimmy Lin; | code |
228 | Relation Extraction with Word Graphs from N-grams Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: To address this limitation, in this paper, we propose attentive graph convolutional networks (A-GCN) to improve neural RE methods with an unsupervised manner to build the context graph, without relying on the existence of a dependency parser. |
Han Qin; Yuanhe Tian; Yan Song; | |
229 | A Bayesian Framework for Information-Theoretic Probing Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: This paper proposes a new framework to measure what we term Bayesian mutual information, which analyses information from the perspective of Bayesian agents-allowing for more intuitive findings in scenarios with finite data. |
Tiago Pimentel; Ryan Cotterell; | |
230 | Masked Language Modeling and The Distributional Hypothesis: Order Word Matters Pre-training for Little Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we propose a different explanation: MLMs succeed on downstream tasks almost entirely due to their ability to model higher-order word co-occurrence statistics. |
Koustuv Sinha; Robin Jia; Dieuwke Hupkes; Joelle Pineau; Adina Williams; Douwe Kiela; | |
231 | What’s Hidden in A One-layer Randomly Weighted Transformer? Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We demonstrate that, hidden within one-layer randomly weighted neural networks, there exist subnetworks that can achieve impressive performance, without ever modifying the weight initializations, on machine translation tasks. |
Sheng Shen; Zhewei Yao; Douwe Kiela; Kurt Keutzer; Michael Mahoney; | |
232 | Rethinking Denoised Auto-Encoding in Language Pre-Training Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: However, such pre-training approaches are prone to learning representations that are covariant with the noise, leading to the discrepancy between the pre-training and fine-tuning stage. To remedy this, we present ContrAstive Pre-Training (CAPT) to learn noise invariant sequence representations. |
Fuli Luo; Pengcheng Yang; Shicheng Li; Xuancheng Ren; Xu Sun; Songfang Huang; Fei Huang; | |
233 | Lifelong Explainer for Lifelong Learners Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we propose a novel Lifelong Explanation (LLE) approach that continuously trains a student explainer under the supervision of a teacher – an arbitrary explanation algorithm – on different tasks undertaken in LL. |
Xuelin Situ; Sameen Maruf; Ingrid Zukerman; Cecile Paris; Gholamreza Haffari; | code |
234 | Linguistic Dependencies and Statistical Dependence Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this work we contribute an extensive analysis of the relationship between linguistic dependencies and statistical dependence between words. |
Jacob Louis Hoover; Wenyu Du; Alessandro Sordoni; Timothy J. O?Donnell; | |
235 | Modeling Human Sentence Processing with Left-Corner Recurrent Neural Network Grammars Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we investigated whether hierarchical structures make LMs more human-like, and if so, which parsing strategy is most cognitively plausible. |
Ryo Yoshida; Hiroshi Noji; Yohei Oseki; | |
236 | A Simple and Effective Positional Encoding for Transformers Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: Motivated by this, we introduce Decoupled Positional Attention for Transformers (DIET), a simple yet effective mechanism to encode position and segment information into the Transformer models. |
Pu-Chin Chen; Henry Tsai; Srinadh Bhojanapalli; Hyung Won Chung; Yin-Wen Chang; Chun-Sung Ferng; | |
237 | Explore Better Relative Position Embeddings from Encoding Perspective for Transformer Models Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we investigate the potential problems in Shaw-RPE and XL-RPE, which are the most representative and prevalent RPEs, and propose two novel RPEs called Low-level Fine-grained High-level Coarse-grained (LFHC) RPE and Gaussian Cumulative Distribution Function (GCDF) RPE. |
Anlin Qu; Jianwei Niu; Shasha Mo; | |
238 | Adversarial Mixing Policy for Relaxing Locally Linear Constraints in Mixup Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: To address these issues, we propose the Adversarial Mixing Policy (AMP), organized in a min-max-rand formulation, to relax the Locally Linear Constraints in Mixup. |
Guang Liu; Yuzhao Mao; Huang Hailong; Gao Weiguo; Li Xuan; | |
239 | Is This The End of The Gold Standard? A Straightforward Reference-less Grammatical Error Correction Metric Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We propose a reference-less GEC evaluation system that is strongly correlated with human judgement, solves the issues related to the use of a reference, and does not need another annotated dataset for fine-tuning. |
Md Asadul Islam; Enrico Magnani; | |
240 | Augmenting BERT-style Models with Predictive Coding to Improve Discourse-level Representations Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this work, we propose to use ideas from predictive coding theory to augment BERT-style language models with a mechanism that allows them to learn suitable discourse-level representations. |
Vladimir Araujo; Andr?s Villa; Marcelo Mendoza; Marie-Francine Moens; Alvaro Soto; | |
241 | Backdoor Attacks on Pre-trained Models By Layerwise Weight Poisoning Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we propose a stronger weight-poisoning attack method that introduces a layerwise weight poisoning strategy to plant deeper backdoors; we also introduce a combinatorial trigger that cannot be easily detected. |
Linyang Li; Demin Song; Xiaonan Li; Jiehang Zeng; Ruotian Ma; Xipeng Qiu; | |
242 | GAML-BERT: Improving BERT Early Exiting By Gradient Aligned Mutual Learning Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this work, we propose a novel framework, Gradient Aligned Mutual Learning BERT (GAML-BERT), for improving the early exiting of BERT. |
Wei Zhu; Xiaoling Wang; Yuan Ni; Guotong Xie; | |
243 | The Power of Scale for Parameter-Efficient Prompt Tuning Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this work, we explore prompt tuning, a simple yet effective mechanism for learning soft prompts to condition frozen language models to perform specific downstream tasks. |
Brian Lester; Rami Al-Rfou; Noah Constant; | |
244 | Scalable Font Reconstruction with Dual Latent Manifolds Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We propose a deep generative model that performs typography analysis and font reconstruction by learning disentangled manifolds of both font style and character shape. |
Nikita Srivatsan; Si Wu; Jonathan Barron; Taylor Berg-Kirkpatrick; | |
245 | Neuro-Symbolic Approaches for Text-Based Policy Learning Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We present SymboLic Action policy for Textual Environments (SLATE), that learns interpretable action policy rules from symbolic abstractions of textual observations for improved generalization. |
Subhajit Chaudhury; Prithviraj Sen; Masaki Ono; Daiki Kimura; Michiaki Tatsubori; Asim Munawar; | |
246 | Layer-wise Model Pruning Based on Mutual Information Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: Inspired by mutual information (MI) based feature selection in SVMs and logistic regression, in this paper, we propose MI-based layer-wise pruning: for each layer of a multi-layer neural network, neurons with higher values of MI with respect to preserved neurons in the upper layer are preserved. |
Chun Fan; Jiwei Li; Tianwei Zhang; Xiang Ao; Fei Wu; Yuxian Meng; Xiaofei Sun; | |
247 | Hierarchical Heterogeneous Graph Representation Learning for Short Text Classification Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we propose a new method called SHINE, which is based on graph neural network (GNN), for short text classification. |
Yaqing Wang; Song Wang; Quanming Yao; Dejing Dou; | |
248 | KFolden: K-Fold Ensemble for Out-Of-Distribution Detection Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this work, we propose a simple yet effective framework kFolden, which mimics the behaviors of OOD detection during training without the use of any external data. |
Xiaoya Li; Jiwei Li; Xiaofei Sun; Chun Fan; Tianwei Zhang; Fei Wu; Yuxian Meng; Jun Zhang; | |
249 | Frustratingly Simple Pretraining Alternatives to Masked Language Modeling Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we explore five simple pretraining objectives based on token-level classification tasks as replacements of MLM. |
Atsuki Yamaguchi; George Chrysostomou; Katerina Margatina; Nikolaos Aletras; | |
250 | HRKD: Hierarchical Relational Knowledge Distillation for Cross-domain Language Model Compression Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we target to compress PLMs with knowledge distillation, and propose a hierarchical relational knowledge distillation (HRKD) method to capture both hierarchical and domain relational information. |
Chenhe Dong; Yaliang Li; Ying Shen; Minghui Qiu; | code |
251 | Searching for An Effective Defender: Benchmarking Defense Against Adversarial Word Substitution Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we seek to fill the gap of systematic studies through comprehensive researches on understanding the behavior of neural text classifiers trained by various defense methods under representative adversarial attacks. |
Zongyi Li; Jianhan Xu; Jiehang Zeng; Linyang Li; Xiaoqing Zheng; Qi Zhang; Kai-Wei Chang; Cho-Jui Hsieh; | code |
252 | Re-embedding Difficult Samples Via Mutual Information Constrained Semantically Oversampling for Imbalanced Text Classification Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we propose a Mutual Information constrained Semantically Oversampling framework (MISO) that can generate anchor instances to help the backbone network determine the re-embedding position of a non-overlapping representation for each difficult sample. |
Jiachen Tian; Shizhan Chen; Xiaowang Zhang; Zhiyong Feng; Deyi Xiong; Shaojuan Wu; Chunliu Dou; | |
253 | Beyond Text: Incorporating Metadata and Label Structure for Multi-Label Document Classification Using Heterogeneous Graphs Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: Therefore, in this paper, we propose a novel neural network based approach for multi-label document classification, in which two heterogeneous graphs are constructed and learned using heterogeneous graph transformers. |
Chenchen Ye; Linhai Zhang; Yulan He; Deyu Zhou; Jie Wu; | |
254 | Natural Language Processing Meets Quantum Physics: A Survey and Categorization Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this survey, we review representative methods at the intersection of NLP and quantum physics in the past ten years, categorizing them according to the use of quantum theory, the linguistic targets that are modeled, and the downstream application. |
Sixuan Wu; Jian Li; Peng Zhang; Yue Zhang; | |
255 | MetaTS: Meta Teacher-Student Network for Multilingual Sequence Labeling with Minimal Supervision Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this work, we explore multilingual sequence labeling with minimal supervision using a single unified model for multiple languages. |
Zheng Li; Danqing Zhang; Tianyu Cao; Ying Wei; Yiwei Song; Bing Yin; | |
256 | Neural Machine Translation with Heterogeneous Topic Knowledge Embeddings Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we propose heterogeneous ways of embedding topic information at the sentence level into an NMT model to improve translation performance. |
Weixuan Wang; Wei Peng; Meng Zhang; Qun Liu; | |
257 | Allocating Large Vocabulary Capacity for Cross-Lingual Language Model Pre-Training Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: To this end, we propose an algorithm VoCap to determine the desired vocabulary capacity of each language. |
Bo Zheng; Li Dong; Shaohan Huang; Saksham Singhal; Wanxiang Che; Ting Liu; Xia Song; Furu Wei; | code |
258 | Recurrent Attention for Neural Machine Translation Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we push further in this research line and propose a novel substitute mechanism for self-attention: Recurrent AtteNtion (RAN) . |
Jiali Zeng; Shuangzhi Wu; Yongjing Yin; Yufan Jiang; Mu Li; | |
259 | Learning from Multiple Noisy Augmented Data Sets for Better Cross-Lingual Spoken Language Understanding Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper we focus on mitigating noise in augmented data. |
Yingmei Guo; Linjun Shou; Jian Pei; Ming Gong; Mingxing Xu; Zhiyong Wu; Daxin Jiang; | |
260 | Enlivening Redundant Heads in Multi-head Self-attention for Machine Translation Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We propose a redundant head enlivening (RHE) method to precisely identify redundant heads, and then vitalize their potential by learning syntactic relations and prior knowledge in the text without sacrificing the roles of important heads. |
Tianfu Zhang; Heyan Huang; Chong Feng; Longbing Cao; | |
261 | Unsupervised Neural Machine Translation with Universal Grammar Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: Therefore, in this paper, we seek to leverage such shared grammar clues to provide more explicit language parallel signals to enhance the training of unsupervised machine translation models. |
Zuchao Li; Masao Utiyama; Eiichiro Sumita; Hai Zhao; | |
262 | Encouraging Lexical Translation Consistency for Document-Level Neural Machine Translation Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper we apply one translation per discourse in NMT, and aim to encourage lexical translation consistency for document-level NMT. |
Xinglin Lyu; Junhui Li; Zhengxian Gong; Min Zhang; | |
263 | Improving Neural Machine Translation By Bidirectional Training Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We present a simple and effective pretraining strategy – bidirectional training (BiT) for neural machine translation. |
Liang Ding; Di Wu; Dacheng Tao; | |
264 | Scheduled Sampling Based on Decoding Steps for Neural Machine Translation Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: To alleviate the above discrepancy, we propose scheduled sampling methods based on decoding steps, increasing the selection chance of predicted tokens with the growth of decoding steps. |
Yijin Liu; Fandong Meng; Yufeng Chen; Jinan Xu; Jie Zhou; | |
265 | Learning to Rewrite for Non-Autoregressive Neural Machine Translation Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we propose an architecture named RewriteNAT to explicitly learn to rewrite the erroneous translation pieces. |
Xinwei Geng; Xiaocheng Feng; Bing Qin; | |
266 | SHAPE : Shifted Absolute Position Embedding for Transformers Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: Existing position representations suffer from a lack of generalization to test data with unseen lengths or high computational cost. We investigate shifted absolute position embedding (SHAPE) to address both issues. |
Shun Kiyono; Sosuke Kobayashi; Jun Suzuki; Kentaro Inui; | |
267 | Self-Supervised Quality Estimation for Machine Translation Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: To reduce the negative impact of noises, we propose a self-supervised method for both sentence- and word-level QE, which performs quality estimation by recovering the masked target words. |
Yuanhang Zheng; Zhixing Tan; Meng Zhang; Mieradilijiang Maimaiti; Huanbo Luan; Maosong Sun; Qun Liu; Yang Liu; | |
268 | Generalised Unsupervised Domain Adaptation of Neural Machine Translation with Cross-Lingual Data Selection Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We propose a cross-lingual data selection method to extract in-domain sentences in the missing language side from a large generic monolingual corpus. |
Thuy-Trang Vu; Xuanli He; Dinh Phung; Gholamreza Haffari; | |
269 | STANKER: Stacking Network Based on Level-grained Attention-masked BERT for Rumor Detection on Social Media Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: To alleviate these problems, we build a new Chinese microblog dataset named Weibo20 by collecting posts and associated comments from Sina Weibo and propose a new ensemble named STANKER (Stacking neTwork bAsed-on atteNtion-masKed BERT). |
Dongning Rao; Xin Miao; Zhihua Jiang; Ran Li; | |
270 | ActiveEA: Active Learning for Neural Entity Alignment Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we devise a novel Active Learning (AL) framework for neural EA, aiming to create highly informative seed alignment to obtain more effective EA models with less annotation cost. |
Bing Liu; Harrisen Scells; Guido Zuccon; Wen Hua; Genghong Zhao; | |
271 | Cost-effective End-to-end Information Extraction for Semi-structured Document Images Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: Here we present our recent effort on transitioning from our existing pipeline-based IE system to an end-to-end system focusing on practical challenges that are associated with replacing and deploying the system in real, large-scale production. |
Wonseok Hwang; Hyunji Lee; Jinyeong Yim; Geewook Kim; Minjoon Seo; | |
272 | Improving Math Word Problems with Pre-trained Knowledge and Hierarchical Reasoning Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we propose a Reasoning with Pre-trained Knowledge and Hierarchical Structure (RPKHS) network, which contains a pre-trained knowledge encoder and a hierarchical reasoning encoder. |
Weijiang Yu; Yingpeng Wen; Fudan Zheng; Nong Xiao; | |
273 | GraphMR: Graph Neural Network for Mathematical Reasoning Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: Having transformed to the new representations, we proposed a graph-to-sequence neural network GraphMR, which can effectively learn the hierarchical information of graphs inputs to solve mathematics and speculate answers. |
Weijie Feng; Binbin Liu; Dongpeng Xu; Qilong Zheng; Yun Xu; | |
274 | What Changes Can Large-scale Language Models Bring? Intensive Study on HyperCLOVA: Billions-scale Korean Generative Pretrained Transformers Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: To achieve this, we introduce HyperCLOVA, a Korean variant of 82B GPT-3 trained on a Korean-centric corpus of 560B tokens. |
Boseop Kim; HyoungSeok Kim; Sang-Woo Lee; Gichang Lee; Donghyun Kwak; Jeon Dong Hyeon; Sunghyun Park; Sungju Kim; Seonhoon Kim; Dongpil Seo; Heungsub Lee; Minyoung Jeong; Sungjae Lee; Minsub Kim; Suk Hyun Ko; Seokhun Kim; Taeyong Park; Jinuk Kim; Soyoung Kang; Na-Hyeon Ryu; Kang Min Yoo; Minsuk Chang; Soobin Suh; Sookyo In; Jinseong Park; Kyungduk Kim; Hiun Kim; Jisu Jeong; Yong Goo Yeo; Donghoon Ham; Dongju Park; Min Young Lee; Jaewook Kang; Inho Kang; Jung-Woo Ha; Woomyoung Park; Nako Sung; | |
275 | APIRecX: Cross-Library API Recommendation Via Pre-Trained Language Model Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we propose APIRecX, the first cross-library API recommendation approach, which uses BPE to split each API call in each API sequence and pre-trains a GPT based language model. |
Yuning Kang; Zan Wang; Hongyu Zhang; Junjie Chen; Hanmo You; | |
276 | GMH: A General Multi-hop Reasoning Model for KG Completion Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: Therefore, we propose a general model which resolves the issues with three modules: 1) the local-global knowledge module to estimate the possible paths, 2) the differentiated action dropout module to explore a diverse set of paths, and 3) the adaptive stopping search module to avoid over searching. |
Yao Zhang; Hongru Liang; Adam Jatowt; Wenqiang Lei; Xin Wei; Ning Jiang; Zhenglu Yang; | |
277 | BPM_MT: Enhanced Backchannel Prediction Model Using Multi-Task Learning Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: To address this limitation, we present a BC prediction model called BPM_MT (Backchannel prediction model with multitask learning), which utilizes KoBERT, a pre-trained language model. |
Jin Yea Jang; San Kim; Minyoung Jung; Saim Shin; Gahgene Gweon; | |
278 | Graphine: A Dataset for Graph-aware Terminology Definition Generation Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we present a large-scale terminology definition dataset Graphine covering 2,010,648 terminology definition pairs, spanning 227 biomedical subdisciplines. |
Zequn Liu; Shukai Wang; Yiyang Gu; Ruiyi Zhang; Ming Zhang; Sheng Wang; | |
279 | Leveraging Order-Free Tag Relations for Context-Aware Recommendation Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We propose a sequence-oblivious generation method for tag recommendation, in which the next tag to be generated is independent of the order of the generated tags and the order of the ground truth tags occurring in training data. |
Junmo Kang; Jeonghwan Kim; Suwon Shin; Sung-Hyon Myaeng; | |
280 | End-to-End Conversational Search for Online Shopping with Utterance Transfer Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this work we first propose ConvSearch, an end-to-end conversational search system that deeply combines the dialog system with search. |
Liqiang Xiao; Jun Ma; Xin Luna Dong; Pascual Mart?nez-G?mez; Nasser Zalmout; Wei Chen; Tong Zhao; Hao He; Yaohui Jin; | |
281 | Self-Supervised Curriculum Learning for Spelling Error Correction Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we study how to further improve the performance of the state-of-the-art SEC method with CL, and propose a Self-Supervised Curriculum Learning (SSCL) approach. |
Zifa Gan; Hongfei Xu; Hongying Zan; | |
282 | Fix-Filter-Fix: Intuitively Connect Any Models for Effective Bug Fixing Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: Based on these, we propose an intuitive yet effective general framework (called Fix-Filter-Fix or F^3) for bug fixing. |
Haiwen Hong; Jingfeng Zhang; Yin Zhang; Yao Wan; Yulei Sui; | |
283 | Neuro-Symbolic Reinforcement Learning with First-Order Logic Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In order to achieve fast convergence and interpretability for the policy in RL, we propose a novel RL method for text-based games with a recent neuro-symbolic framework called Logical Neural Network, which can learn symbolic and interpretable rules in their differentiable network. |
Daiki Kimura; Masaki Ono; Subhajit Chaudhury; Ryosuke Kohita; Akifumi Wachi; Don Joven Agravante; Michiaki Tatsubori; Asim Munawar; Alexander Gray; | |
284 | Biomedical Concept Normalization By Leveraging Hypernyms Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we exploit biomedical concept hypernyms to facilitate BCN. |
Cheng Yan; Yuanzhe Zhang; Kang Liu; Jun Zhao; Yafei Shi; Shengping Liu; | |
285 | Leveraging Capsule Routing to Associate Knowledge with Medical Literature Hierarchically Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, to alleviate this problem, we propose leveraging capsule routing to associate knowledge with medical literature hierarchically (called HiCapsRKL). |
Xin Liu; Qingcai Chen; Junying Chen; Wenxiu Zhou; Tingyu Liu; Xinlan Yang; Weihua Peng; | |
286 | Label-Enhanced Hierarchical Contextualized Representation for Sequential Metaphor Identification Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we propose a model augmented with hierarchical contextualized representation to extract more information from both sentence-level and discourse-level. |
Shuqun Li; Liang Yang; Weidong He; Shiqi Zhang; Jingjie Zeng; Hongfei Lin; | |
287 | SpellBERT: A Lightweight Pretrained Model for Chinese Spelling Check Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: To that end, we propose SpellBERT, a pretrained model with graph-based extra features and independent on confusion set. |
Tuo Ji; Hang Yan; Xipeng Qiu; | |
288 | Automated Generation of Accurate & Fluent Medical X-ray Reports Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: Our paper aims to automate the generation of medical reports from chest X-ray image inputs, a critical yet time-consuming task for radiologists. |
Hoang Nguyen; Dong Nie; Taivanbat Badamdorj; Yujie Liu; Yingying Zhu; Jason Truong; Li Cheng; | |
289 | Enhancing Document Ranking with Task-adaptive Training and Segmented Token Recovery Mechanism Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we propose a new ranking model DR-BERT, which improves the Document Retrieval (DR) task by a task-adaptive training process and a Segmented Token Recovery Mechanism (STRM). |
Xingwu Sun; Yanling Cui; Hongyin Tang; Fuzheng Zhang; Beihong Jin; Shi Wang; | |
290 | Abstract, Rationale, Stance: A Joint Model for Scientific Claim Verification Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We thus propose an approach, named as ARSJoint, that jointly learns the modules for the three tasks with a machine reading comprehension framework by including claim information. |
Zhiwei Zhang; Jiyi Li; Fumiyo Fukumoto; Yanming Ye; | |
291 | A Fine-Grained Domain Adaption Model for Joint Word Segmentation and POS Tagging Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this work, we start from joint word segmentation and POS tagging, presenting a fine-grained domain adaption method to model the gaps accurately. |
Peijie Jiang; Dingkun Long; Yueheng Sun; Meishan Zhang; Guangwei Xu; Pengjun Xie; | |
292 | Answering Open-Domain Questions of Varying Reasoning Steps from Text Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We employ a single multi-task transformer model to perform all the necessary subtasks-retrieving supporting facts, reranking them, and predicting the answer from all retrieved documents-in an iterative fashion. |
Peng Qi; Haejun Lee; Tg Sido; Christopher Manning; | code |
293 | Adaptive Information Seeking for Open-Domain Question Answering Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we propose a novel adaptive information-seeking strategy for open-domain question answering, namely AISO. |
Yunchang Zhu; Liang Pang; Yanyan Lan; Huawei Shen; Xueqi Cheng; | |
294 | Mapping Probability Word Problems to Executable Representations Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we employ and analyse various neural models for answering such word problems. |
Simon Suster; Pieter Fivez; Pietro Totis; Angelika Kimmig; Jesse Davis; Luc de Raedt; Walter Daelemans; | |
295 | Enhancing Multiple-choice Machine Reading Comprehension By Punishing Illogical Interpretations Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: Based on post-hoc interpretation methods, we assess attributions of paragraphs in multiple-choice MRC and improve the model by punishing the illogical attributions. |
Yiming Ju; Yuanzhe Zhang; Zhixing Tian; Kang Liu; Xiaohuan Cao; Wenting Zhao; Jinlong Li; Jun Zhao; | |
296 | Large-Scale Relation Learning for Question Answering Over Knowledge Bases with Pre-trained Language Models Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: To bridge the gap between the natural language and the structured KB, we propose three relation learning tasks for BERT-based KBQA, including relation extraction, relation matching, and relation reasoning. |
Yuanmeng Yan; Rumei Li; Sirui Wang; Hongzhi Zhang; Zan Daoguang; Fuzheng Zhang; Wei Wu; Weiran Xu; | |
297 | Phrase Retrieval Learns Passage Retrieval, Too Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this work, we follow the intuition that retrieving phrases naturally entails retrieving larger text blocks and study whether phrase retrieval can serve as the basis for coarse-level retrieval including passages and documents. |
Jinhyuk Lee; Alexander Wettig; Danqi Chen; | |
298 | Neural Natural Logic Inference for Interpretable Question Answering Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we investigate a neural-symbolic QA approach that integrates natural logic reasoning within deep learning architectures, towards developing effective and yet explainable question answering models. |
Jihao Shi; Xiao Ding; Li Du; Ting Liu; Bing Qin; | |
299 | Smoothing Dialogue States for Open Conversational Machine Reading Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this work, we propose an effective gating strategy by smoothing the two dialogue states in only one decoder and bridge decision making and question generation to provide a richer dialogue state reference. |
Zhuosheng Zhang; Siru Ouyang; Hai Zhao; Masao Utiyama; Eiichiro Sumita; | |
300 | FinQA: A Dataset of Numerical Reasoning Over Financial Data Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this work, we focus on answering deep questions over financial data, aiming to automate the analysis of a large corpus of financial documents. |
Zhiyu Chen; Wenhu Chen; Charese Smiley; Sameena Shah; Iana Borova; Dylan Langdon; Reema Moussa; Matt Beane; Ting-Hao Huang; Bryan Routledge; William Yang Wang; | code |
301 | FiD-Ex: Improving Sequence-to-Sequence Models for Extractive Rationale Generation Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we develop FiD-Ex, which addresses these shortcomings for seq2seq models by: 1) introducing sentence markers to eliminate explanation fabrication by encouraging extractive generation, 2) using the fusion-in-decoder architecture to handle long input contexts, and 3) intermediate fine-tuning on re-structured open domain QA datasets to improve few-shot performance. |
Kushal Lakhotia; Bhargavi Paranjape; Asish Ghoshal; Scott Yih; Yashar Mehdad; Srini Iyer; | |
302 | RockNER: A Simple Method to Create Adversarial Examples for Evaluating The Robustness of Named Entity Recognition Models Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: To audit the robustness of named entity recognition (NER) models, we propose RockNER, a simple yet effective method to create natural adversarial examples. |
Bill Yuchen Lin; Wenyang Gao; Jun Yan; Ryan Moreno; Xiang Ren; | |
303 | Diagnosing The First-Order Logical Reasoning Ability Through LogicNLI Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this work, we propose a diagnostic method for first-order logic (FOL) reasoning with a new proposed benchmark, LogicNLI. |
Jidong Tian; Yitian Li; Wenqing Chen; Liqiang Xiao; Hao He; Yaohui Jin; | |
304 | Constructing A Psychometric Testbed for Fair Natural Language Processing Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper we describe our efforts to construct a corpus for psychometric natural language processing (NLP) related to important dimensions such as trust, anxiety, numeracy, and literacy, in the health domain. |
Ahmed Abbasi; David Dobolyi; John P. Lalor; Richard G. Netemeyer; Kendall Smith; Yi Yang; | |
305 | COUGH: A Challenge Dataset and Models for COVID-19 FAQ Retrieval Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We present a large, challenging dataset, COUGH, for COVID-19 FAQ retrieval. |
Xinliang Frederick Zhang; Heming Sun; Xiang Yue; Simon Lin; Huan Sun; | code |
306 | Chinese WPLC: A Chinese Dataset for Evaluating Pretrained Language Models on Word Prediction Given Long-Range Context Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: This paper presents a Chinese dataset for evaluating pretrained language models on Word Prediction given Long-term Context (Chinese WPLC). |
Huibin Ge; Chenxi Sun; Deyi Xiong; Qun Liu; | code |
307 | WinoLogic: A Zero-Shot Logic-based Diagnostic Dataset for Winograd Schema Challenge Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: To better evaluate NLMs, we propose a logic-based framework that focuses on high-quality commonsense knowledge. |
Weinan He; Canming Huang; Yongmei Liu; Xiaodan Zhu; | |
308 | Pseudo Zero Pronoun Resolution Improves Zero Anaphora Resolution Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: The first is a new pretraining task that trains MLMs on anaphoric relations with explicit supervision, and the second proposal is a new finetuning method that remedies a notorious issue, the pretrain-finetune discrepancy. |
Ryuto Konno; Shun Kiyono; Yuichiroh Matsubayashi; Hiroki Ouchi; Kentaro Inui; | |
309 | Aligning Cross-lingual Sentence Representations with Dual Momentum Contrast Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we propose to align sentence representations from different languages into a unified embedding space, where semantic similarities (both cross-lingual and monolingual) can be computed with a simple dot product. |
Liang Wang; Wei Zhao; Jingming Liu; | |
310 | Total Recall: A Customized Continual Learning Method for Neural Semantic Parsers Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: Therefore, we propose TotalRecall, a continual learning method designed for neural semantic parsers from two aspects: i) a sampling method for memory replay that diversifies logical form templates and balances distributions of parse actions in a memory; ii) a two-stage training method that significantly improves generalization capability of the parsers across tasks. |
Zhuang Li; Lizhen Qu; Gholamreza Haffari; | |
311 | Exophoric Pronoun Resolution in Dialogues with Topic Regularization Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: Motivated by this, we propose to jointly leverage the local context and global topics of dialogues to solve the out-of-text PCR problem. |
Xintong Yu; Hongming Zhang; Yangqiu Song; Changshui Zhang; Kun Xu; Dong Yu; | |
312 | Context-Aware Interaction Network for Question Matching Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We propose a context-aware interaction network (COIN) to properly align two sequences and infer their semantic relationship. |
Zhe Hu; Zuohui Fu; Yu Yin; Gerard de Melo; | |
313 | TEMP: Taxonomy Expansion with Dynamic Margin Loss Through Taxonomy-Paths Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we propose TEMP, a self-supervised taxonomy expansion method, which predicts the position of new concepts by ranking the generated taxonomy-paths. |
Zichen Liu; Hongyuan Xu; Yanlong Wen; Ning Jiang; HaiYing Wu; Xiaojie Yuan; | |
314 | A Graph-Based Neural Model for End-to-End Frame Semantic Parsing Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this work, we propose an end-to-end neural model to tackle the task jointly. |
ZhiChao Lin; Yueheng Sun; Meishan Zhang; | |
315 | Virtual Data Augmentation: A Robust and General Framework for Fine-tuning Pre-trained Models Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this work, we present Virtual Data Augmentation (VDA), a general framework for robustly fine-tuning PLMs. |
Kun Zhou; Wayne Xin Zhao; Sirui Wang; Fuzheng Zhang; Wei Wu; Ji-Rong Wen; | code |
316 | CATE: A Contrastive Pre-trained Model for Metaphor Detection with Semi-supervised Learning Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we proposed a ContrAstive pre-Trained modEl (CATE) for metaphor detection with semi-supervised learning. |
Zhenxi Lin; Qianli Ma; Jiangyue Yan; Jieyu Chen; | |
317 | To Be Closer: Learning to Link Up Aspects with Opinions Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this work, we aim to shorten the distance between aspects and corresponding opinion words by learning an aspect-centric tree structure. |
Yuxiang Zhou; Lejian Liao; Yang Gao; Zhanming Jie; Wei Lu; | |
318 | Seeking Common But Distinguishing Difference, A Joint Aspect-based Sentiment Analysis Model Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: Therefore, we propose a joint ABSA model, which not only enjoys the benefits of encoder sharing but also focuses on the difference to improve the effectiveness of the model. |
Hongjiang Jing; Zuchao Li; Hai Zhao; Shu Jiang; | |
319 | Argument Pair Extraction with Mutual Guidance and Inter-sentence Relation Graph Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we tackle the APE task by a mutual guidance framework, which could utilize the information of an argument in one passage to guide the identification of arguments that can form pairs with it in another passage. |
Jianzhu Bao; Bin Liang; Jingyi Sun; Yice Zhang; Min Yang; Ruifeng Xu; | |
320 | Emotion Inference in Multi-Turn Conversations with Addressee-Aware Module and Ensemble Strategy Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this study, we focus on investigating the task of emotion inference in multi-turn conversations by modeling the propagation of emotional states among participants in the conversation history, and propose an addressee-aware module to automatically learn whether the participant keeps the historical emotional state or is affected by others in the next upcoming turn. |
Dayu Li; Xiaodan Zhu; Yang Li; Suge Wang; Deyu Li; Jian Liao; Jianxing Zheng; | |
321 | Improving Federated Learning for Aspect-based Sentiment Analysis Via Topic Memories Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: To address this issue and make the best use of all labeled data, we propose a novel ABSA model with federated learning (FL) adopted to overcome the data isolation limitations and incorporate topic memory (TM) proposed to take the cases of data from diverse sources (domains) into consideration. |
Han Qin; Guimin Chen; Yuanhe Tian; Yan Song; | |
322 | Comparative Opinion Quintuple Extraction from Product Reviews Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: To address these limitations, in this work we first introduce a new Comparative Opinion Quintuple Extraction (COQE) task, to identify comparative sentences from product reviews and extract all comparative opinion quintuples (Subject, Object, Comparative Aspect, Comparative Opinion, Comparative Preference). |
Ziheng Liu; Rui Xia; Jianfei Yu; | |
323 | CTAL: Pre-training Cross-modal Transformer for Audio-and-Language Representations Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we present a Cross-modal Transformer for Audio-and-Language, i.e., CTAL, which aims to learn the intra-modality and inter-modality connections between audio and language through two proxy tasks on a large amount of audio-and-language pairs: masked language modeling and masked cross-modal acoustic modeling. |
Hang Li; Wenbiao Ding; Yu Kang; Tianqiao Liu; Zhongqin Wu; Zitao Liu; | code |
324 | Relation-aware Video Reading Comprehension for Temporal Language Grounding Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: This paper will formulate temporal language grounding into video reading comprehension and propose a Relation-aware Network (RaNet) to address it. |
Jialin Gao; Xin Sun; Mengmeng Xu; Xi Zhou; Bernard Ghanem; | code |
325 | Mutual-Learning Improves End-to-End Speech Translation Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we propose an alternative-a trainable mutual-learning scenario, where the MT and the ST models are collaboratively trained and are considered as peers, rather than teacher/student. |
Jiawei Zhao; Wei Luo; Boxing Chen; Andrew Gilman; | |
326 | Vision Guided Generative Pre-trained Language Models for Multimodal Abstractive Summarization Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we present a simple yet effective method to construct vision guided (VG) GPLMs for the MAS task using attention-based add-on layers to incorporate visual information while maintaining their original text generation ability. |
Tiezheng Yu; Wenliang Dai; Zihan Liu; Pascale Fung; | |
327 | Natural Language Video Localization with Learnable Moment Proposals Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we argue that the performance of propose-and-rank models are underestimated due to the predefined manners: 1) Hand-designed rules are hard to guarantee the complete coverage of targeted segments. |
Shaoning Xiao; Long Chen; Jian Shao; Yueting Zhuang; Jun Xiao; | |
328 | Language-Aligned Waypoint (LAW) Supervision for Vision-and-Language Navigation in Continuous Environments Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this work, we propose a simple and effective language-aligned supervision scheme, and a new metric that measures the number of sub-instructions the agent has completed during navigation. |
Sonia Raychaudhuri; Saim Wani; Shivansh Patel; Unnat Jain; Angel Chang; | |
329 | How to Leverage The Multimodal EHR Data for Better Medical Prediction? Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: Therefore, in this paper, we first extract the accompanying clinical notes from EHR and propose a method to integrate these data, we also comprehensively study the different models and the data leverage methods for better medical task prediction performance. |
Bo Yang; Lijun Wu; | |
330 | Considering Nested Tree Structure in Sentence Extractive Summarization with Pre-trained Transformer Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We propose a nested tree-based extractive summarization model on RoBERTa (NeRoBERTa), where nested tree structures consist of syntactic and discourse trees in a given document. |
Jingun Kwon; Naoki Kobayashi; Hidetaka Kamigaito; Manabu Okumura; | |
331 | Frame Semantic-Enhanced Sentence Modeling for Sentence-level Extractive Text Summarization Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we propose a novel Frame Semantic-Enhanced Sentence Modeling for Extractive Summarization, which leverages Frame semantics to model sentences from both intra-sentence level and inter-sentence level, facilitating the text summarization task. |
Yong Guan; Shaoru Guo; Ru Li; Xiaoli Li; Hongye Tan; | |
332 | CAST: Enhancing Code Summarization with Hierarchical Splitting and Reconstruction of Abstract Syntax Trees Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we propose a novel model CAST that hierarchically splits and reconstructs ASTs. |
Ensheng Shi; Yanlin Wang; Lun Du; Hongyu Zhang; Shi Han; Dongmei Zhang; Hongbin Sun; | code |
333 | SgSum:Transforming Multi-document Summarization Into Sub-graph Selection Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we propose a novel MDS framework (SgSum) to formulate the MDS task as a sub-graph selection problem, in which source documents are regarded as a relation graph of sentences (e.g., similarity graph or discourse graph) and the candidate summaries are its sub-graphs. |
Moye Chen; Wei Li; Jiachen Liu; Xinyan Xiao; Hua Wu; Haifeng Wang; | |
334 | Event Graph Based Sentence Fusion Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we explore the effective sentence fusion method in the context of text summarization. |
Ruifeng Yuan; Zili Wang; Wenjie Li; | |
335 | Transformer-based Lexically Constrained Headline Generation Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: Inspired by previous RNN-based methods generating token sequences in backward and forward directions from the given phrase, we propose a simple Transformer-based method that guarantees to include the given phrase in the high-quality generated headline. |
Kosuke Yamada; Yuta Hitomi; Hideaki Tamori; Ryohei Sasano; Naoaki Okazaki; Kentaro Inui; Koichi Takeda; | |
336 | Learn to Copy from The Copying History: Correlational Copy Network for Abstractive Summarization Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we propose a novel copying scheme named Correlational Copying Network (CoCoNet) that enhances the standard copying mechanism by keeping track of the copying history. |
Haoran Li; Song Xu; Peng Yuan; Yujia Wang; Youzheng Wu; Xiaodong He; Bowen Zhou; | |
337 | Gradient-Based Adversarial Factual Consistency Evaluation for Abstractive Summarization Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we proposed an efficient weak-supervised adversarial data augmentation approach to form the factual consistency dataset. |
Zhiyuan Zeng; Jiaze Chen; Weiran Xu; Lei Li; | |
338 | Word Reordering for Zero-shot Cross-lingual Structured Prediction Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we build structured prediction models with bag-of-words inputs, and introduce a new reordering module to organizing words following the source language order, which learns task-specific reordering strategies from a general-purpose order predictor model. |
Tao Ji; Yong Jiang; Tao Wang; Zhongqiang Huang; Fei Huang; Yuanbin Wu; Xiaoling Wang; | |
339 | A Unified Encoding of Structures in Transition Systems Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we propose a novel attention-based encoder unifying representation of all structures in a transition system. |
Tao Ji; Yong Jiang; Tao Wang; Zhongqiang Huang; Fei Huang; Yuanbin Wu; Xiaoling Wang; | |
340 | Improving Unsupervised Question Answering Via Summarization-Informed Question Generation Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In order to overcome these shortcomings, we propose a distantly-supervised QG method which uses questions generated heuristically from summaries as a source of training data for a QG system. |
Chenyang Lyu; Lifeng Shang; Yvette Graham; Jennifer Foster; Xin Jiang; Qun Liu; | |
341 | TransferNet: An Effective and Transparent Framework for Multi-hop Question Answering Over Relation Graph Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we propose TransferNet, an effective and transparent model for multi-hop QA, which supports both label and text relations in a unified framework. |
Jiaxin Shi; Shulin Cao; Lei Hou; Juanzi Li; Hanwang Zhang; | |
342 | Topic Transferable Table Question Answering Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this work we simulate the practical topic shift scenario by designing novel challenge benchmarks WikiSQL-TS and WikiTable-TS, consisting of train-dev-test splits in five distinct topic groups, based on the popular WikiSQL and WikiTable-Questions datasets. |
Saneem Chemmengath; Vishwajeet Kumar; Samarth Bharadwaj; Jaydeep Sen; Mustafa Canim; Soumen Chakrabarti; Alfio Gliozzo; Karthik Sankaranarayanan; | |
343 | WebSRC: A Dataset for Web-Based Structural Reading Comprehension Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we introduce the task of web-based structural reading comprehension. |
Xingyu Chen; Zihan Zhao; Lu Chen; JiaBao Ji; Danyang Zhang; Ao Luo; Yuxuan Xiong; Kai Yu; | |
344 | Cryptonite: A Cryptic Crossword Benchmark for Extreme Ambiguity in Language Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We present Cryptonite, a large-scale dataset based on cryptic crosswords, which is both linguistically complex and naturally sourced. |
Avia Efrat; Uri Shaham; Dan Kilman; Omer Levy; | |
345 | End-to-End Entity Resolution and Question Answering Using Differentiable Knowledge Graphs Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this work, we extend the boundaries of E2E learning for KGQA to include the training of an ER component. |
Amir Saffari; Armin Oliya; Priyanka Sen; Tom Ayoola; | |
346 | Improving Query Graph Generation for Complex Question Answering Over Knowledge Base Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we propose a new solution to query graph generation that works in the opposite manner: we start with the entire knowledge base and gradually shrink it to the desired query graph. |
Kechen Qin; Cheng Li; Virgil Pavlu; Javed Aslam; | |
347 | DiscoDVT: Generating Long Text with Discourse-Aware Discrete Variational Transformer Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we propose DiscoDVT, a discourse-aware discrete variational Transformer to tackle the incoherence issue. |
Haozhe Ji; Minlie Huang; | |
348 | Mathematical Word Problem Generation from Commonsense Knowledge Graph and Equations Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: To address above problem, we develop an end-to-end neural model to generate diverse MWPs in real-world scenarios from commonsense knowledge graph and equations. |
Tianqiao Liu; Qiang Fang; Wenbiao Ding; Hang Li; Zhongqin Wu; Zitao Liu; | code |
349 | Generic Resources Are What You Need: Style Transfer Tasks Without Task-specific Parallel Training Data Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We propose a novel approach to this task that leverages generic resources, and without using any task-specific parallel (source-target) data outperforms existing unsupervised approaches on the two most popular style transfer tasks: formality transfer and polarity swap. |
Huiyuan Lai; Antonio Toral; Malvina Nissim; | |
350 | Revisiting Pivot-Based Paraphrase Generation: Language Is Not The Only Optional Pivot Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we explore the feasibility of using semantic and syntactic representations as the pivot for paraphrase generation. |
Yitao Cai; Yue Cao; Xiaojun Wan; | |
351 | Structural Adapters in Pretrained Language Models for AMR-to-Text Generation Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we propose StructAdapt, an adapter method to encode graph structure into PLMs. |
Leonardo F. R. Ribeiro; Yue Zhang; Iryna Gurevych; | |
352 | Data-to-text Generation By Splicing Together Nearest Neighbors Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We propose to tackle data-to-text generation tasks by directly splicing together retrieved segments of text from neighbor source-target pairs. |
Sam Wiseman; Arturs Backurs; Karl Stratos; | |
353 | Contextualize Knowledge Bases with Transformer for End-to-end Task-Oriented Dialogue Systems Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: To achieve this, we propose a COntext-aware Memory Enhanced Transformer framework (COMET), which treats the KB as a sequence and leverages a novel Memory Mask to enforce the entity to only focus on its relevant entities and dialogue history, while avoiding the distraction from the irrelevant entities. |
Yanjie Gou; Yinjie Lei; Lingqiao Liu; Yong Dai; Chunxu Shen; | |
354 | Efficient Dialogue Complementary Policy Learning Via Deep Q-network Policy and Episodic Memory Policy Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: Inspired by the human brain, this paper proposes a novel complementary policy learning (CPL) framework, which exploits the complementary advantages of the episodic memory (EM) policy and the deep Q-network (DQN) policy to achieve fast and effective dialogue policy learning. |
Yangyang Zhao; Zhenyu Wang; Changxi Zhu; Shihan Wang; | |
355 | CRFR: Improving Conversational Recommender Systems Via Flexible Fragments Reasoning on Knowledge Graphs Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We propose CRFR, which effectively does explicit multi-hop reasoning on KGs with a conversational context-based reinforcement learning model. |
Jinfeng Zhou; Bo Wang; Ruifang He; Yuexian Hou; | |
356 | DuRecDial 2.0: A Bilingual Parallel Corpus for Conversational Recommendation Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we provide a bilingual parallel human-to-human recommendation dialog dataset (DuRecDial 2.0) to enable researchers to explore a challenging task of multilingual and cross-lingual conversational recommendation. |
Zeming Liu; Haifeng Wang; Zheng-Yu Niu; Hua Wu; Wanxiang Che; | |
357 | End-to-End Learning of Flowchart Grounded Task-Oriented Dialogs Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We propose a novel problem within end-to-end learning of task oriented dialogs (TOD), in which the dialog system mimics a troubleshooting agent who helps a user by diagnosing their problem (e.g., car not starting). |
Dinesh Raghu; Shantanu Agarwal; Sachindra Joshi; Mausam; | |
358 | Dimensional Emotion Detection from Categorical Emotion Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We present a model to predict fine-grained emotions along the continuous dimensions of valence, arousal, and dominance (VAD) with a corpus with categorical emotion annotations. |
Sungjoon Park; Jiseon Kim; Seonghyeon Ye; Jaeyeol Jeon; Hee Young Park; Alice Oh; | |
359 | Not All Negatives Are Equal: Label-Aware Contrastive Loss for Fine-grained Text Classification Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this work, we analyse the contrastive fine-tuning of pre-trained language models on two fine-grained text classification tasks, emotion classification and sentiment analysis. |
Varsha Suresh; Desmond Ong; | |
360 | Joint Multi-modal Aspect-Sentiment Analysis with Auxiliary Cross-modal Relation Detection Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: Therefore, in this paper, we are the first to jointly perform multi-modal ATE (MATE) and multi-modal ASC (MASC), and we propose a multi-modal joint learning approach with auxiliary cross-modal relation detection for multi-modal aspect-level sentiment analysis (MALSA). |
Xincheng Ju; Dong Zhang; Rong Xiao; Junhui Li; Shoushan Li; Min Zhang; Guodong Zhou; | |
361 | Solving Aspect Category Sentiment Analysis As A Text Generation Task Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We consider a more direct way of making use of pre-trained language models, by casting the ACSA tasks into natural language generation tasks, using natural language sentences to represent the output. |
Jian Liu; Zhiyang Teng; Leyang Cui; Hanmeng Liu; Yue Zhang; | |
362 | Semantics-Preserved Data Augmentation for Aspect-Based Sentiment Analysis Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we propose a semantics-preservation data augmentation approach by considering the importance of each word in a textual sequence according to the related aspects and sentiments. |
Ting-Wei Hsu; Chung-Chi Chen; Hen-Hsen Huang; Hsin-Hsi Chen; | |
363 | The Effect of Round-Trip Translation on Fairness in Sentiment Analysis Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We explore the impact of round-trip translation on the demographic parity of sentiment classifiers and show how round-trip translation consistently improves classification fairness at test time (reducing up to 47% of between-group gaps). |
Jonathan Christiansen; Mathias Gammelgaard; Anders S?gaard; | |
364 | CHoRaL: Collecting Humor Reaction Labels from Millions of Social Media Users Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We propose CHoRaL, a framework to generate perceived humor labels on Facebook posts, using the naturally available user reactions to these posts with no manual annotation needed. |
Zixiaofan Yang; Shayan Hooshmand; Julia Hirschberg; | |
365 | CSDS: A Fine-Grained Chinese Dataset for Customer Service Dialogue Summarization Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: Therefore, in this paper, we introduce a novel Chinese dataset for Customer Service Dialogue Summarization (CSDS). |
Haitao Lin; Liqun Ma; Junnan Zhu; Lu Xiang; Yu Zhou; Jiajun Zhang; Chengqing Zong; | |
366 | CodRED: A Cross-Document Relation Extraction Dataset for Acquiring Knowledge in The Wild Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this work, we present the problem of cross-document RE, making an initial step towards knowledge acquisition in the wild. |
Yuan Yao; Jiaju Du; Yankai Lin; Peng Li; Zhiyuan Liu; Jie Zhou; Maosong Sun; | |
367 | Building and Evaluating Open-Domain Dialogue Corpora with Clarifying Questions Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: To address the problem of ‘asking clarifying questions in open-domain dialogues’: (1) we collect and release a new dataset focused on open-domain single- and multi-turn conversations, (2) we benchmark several state-of-the-art neural baselines, and (3) we propose a pipeline consisting of offline and online steps for evaluating the quality of clarifying questions in various dialogues. |
Mohammad Aliannejadi; Julia Kiseleva; Aleksandr Chuklin; Jeff Dalton; Mikhail Burtsev; | |
368 | We Need to Talk About Train-dev-test Splits Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We propose to use a tune-set when developing neural network methods, which can be used for model picking so that comparing the different versions of a new model can safely be done on the development data. |
Rob van der Goot; | |
369 | PhoMT: A High-Quality and Large-Scale Benchmark Dataset for Vietnamese-English Machine Translation Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We introduce a high-quality and large-scale Vietnamese-English parallel dataset of 3.02M sentence pairs, which is 2.9M pairs larger than the benchmark Vietnamese-English machine translation corpus IWSLT15. |
Long Doan; Linh The Nguyen; Nguyen Luong Tran; Thai Hoang; Dat Quoc Nguyen; | code |
370 | Lying Through One’s Teeth: A Study on Verbal Leakage Cues Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we study verbal leakage cues to understand the effect of the data construction method on their significance, and examine the relationship between such cues and models’ validity. |
Min-Hsuan Yeh; Lun-Wei Ku; | |
371 | Multi-granularity Textual Adversarial Attack with Behavior Cloning Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: To address such problems, in this paper we propose MAYA, a Multi-grAnularitY Attack model to effectively generate high-quality adversarial samples with fewer queries to victim models. |
Yangyi Chen; Jin Su; Wei Wei; | code |
372 | All Bark and No Bite: Rogue Dimensions in Transformer Language Models Obscure Representational Quality Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this work, we call into question the informativity of such measures for contextualized language models. |
William Timkey; Marten van Schijndel; | |
373 | Incorporating Residual and Normalization Layers Into Analysis of Masked Language Models Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this study, we extended the scope of the analysis of Transformers from solely the attention patterns to the whole attention block, i.e., multi-head attention, residual connection, and layer normalization. |
Goro Kobayashi; Tatsuki Kuribayashi; Sho Yokoi; Kentaro Inui; | |
374 | Mind The Style of Text! Adversarial and Backdoor Attacks Based on Text Style Transfer Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we make the first attempt to conduct adversarial and backdoor attacks based on text style transfer, which is aimed at altering the style of a sentence while preserving its meaning. |
Fanchao Qi; Yangyi Chen; Xurui Zhang; Mukai Li; Zhiyuan Liu; Maosong Sun; | code |
375 | Sociolectal Analysis of Pretrained Language Models Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: Using data from English cloze tests, in which subjects also self-reported their gender, age, education, and race, we examine performance differences of pretrained language models across demographic groups, defined by these (protected) attributes. |
Sheng Zhang; Xin Zhang; Weiming Zhang; Anders S?gaard; | |
376 | Examining Cross-lingual Contextual Embeddings with Orthogonal Structural Probes Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We evaluate syntactic (UD) and lexical (WordNet) structural information encoded inmBERT’s contextual representations for nine diverse languages. |
Tomasz Limisiewicz; David Marecek; | |
377 | Are Transformers A Modern Version of ELIZA? Observations on French Object Verb Agreement Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We take a critical look at this line of research by showing that it is possible to achieve high accuracy on this agreement task with simple surface heuristics, indicating a possible flaw in our assessment of neural networks’ syntactic ability. |
Bingzhi Li; Guillaume Wisniewski; Benoit Crabb?; | |
378 | Fine-grained Entity Typing Via Label Reasoning Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we argue that the implicitly entailed extrinsic and intrinsic dependencies between labels can provide critical knowledge to tackle the above challenges. |
Qing Liu; Hongyu Lin; Xinyan Xiao; Xianpei Han; Le Sun; Hua Wu; | |
379 | Enhanced Language Representation with Label Knowledge for Span Extraction Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: To address those problems, we introduce a fresh paradigm to integrate label knowledge and further propose a novel model to explicitly and efficiently integrate label knowledge into text representations. |
Pan Yang; Xin Cong; Zhenyu Sun; Xingwu Liu; | |
380 | PRIDE: Predicting Relationships in Conversations Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: To infer speakers’ relationships from dialogues we propose PRIDE, a neural multi-label classifier, based on BERT and Transformer for creating a conversation representation. |
Anna Tigunova; Paramita Mirza; Andrew Yates; Gerhard Weikum; | |
381 | Extracting Fine-Grained Knowledge Graphs of Scientific Claims: Dataset and Transformer-Based Results Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: Instead we focus on the subtleties of how experimental associations are presented by building SciClaim, a dataset of scientific claims drawn from Social and Behavior Science (SBS), PubMed, and CORD-19 papers. |
Ian Magnusson; Scott Friedman; | |
382 | Sequential Cross-Document Coreference Resolution Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this work we propose a new model that extends the efficient sequential prediction paradigm for coreference resolution to cross-document settings and achieves competitive results for both entity and event coreference while providing strong evidence of the efficacy of both sequential models and higher-order inference in cross-document settings. |
Emily Allaway; Shuai Wang; Miguel Ballesteros; | |
383 | Mixture-of-Partitions: Infusing Large Biomedical Knowledge Graphs Into BERT Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we proposed Mixture-of-Partitions (MoP), an infusion approach that can handle a very large knowledge graph (KG) by partitioning it into smaller sub-graphs and infusing their specific knowledge into various BERT models using lightweight adapters. |
Zaiqiao Meng; Fangyu Liu; Thomas Clark; Ehsan Shareghi; Nigel Collier; | |
384 | Filling The Gaps in Ancient Akkadian Texts: A Masked Language Modelling Approach Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We present models which complete missing text given transliterations of ancient Mesopotamian documents, originally written on cuneiform clay tablets (2500 BCE – 100 CE). |
Koren Lazar; Benny Saret; Asaf Yehudai; Wayne Horowitz; Nathan Wasserman; Gabriel Stanovsky; | |
385 | AVocaDo: Strategy for Adapting Vocabulary to Downstream Domain Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We propose to consider the vocabulary as an optimizable parameter, allowing us to update the vocabulary by expanding it with domain specific vocabulary based on a tokenization statistic. |
Jimin Hong; TaeHee Kim; Hyesu Lim; Jaegul Choo; | |
386 | Can We Improve Model Robustness Through Secondary Attribute Counterfactuals? Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper we study how and why modeling counterfactuals over multiple attributes can go significantly further in improving model performance. |
Ananth Balashankar; Xuezhi Wang; Ben Packer; Nithum Thain; Ed Chi; Alex Beutel; | |
387 | Long-Range Modeling of Source Code Files with EWASH: Extended Window Access By Syntax Hierarchy Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: While there are many efforts to extend the context window, we introduce an architecture-independent approach for leveraging the syntactic hierarchies of source code for incorporating entire file-level context into a fixed-length window. |
Colin Clement; Shuai Lu; Xiaoyu Liu; Michele Tufano; Dawn Drain; Nan Duan; Neel Sundaresan; Alexey Svyatkovskiy; | |
388 | Can Language Models Be Biomedical Knowledge Bases? Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: To this end, we create the BioLAMA benchmark, which is comprised of 49K biomedical factual knowledge triples for probing biomedical LMs. |
Mujeen Sung; Jinhyuk Lee; Sean Yi; Minji Jeon; Sungdong Kim; Jaewoo Kang; | |
389 | LayoutReader: Pre-training of Text and Layout for Reading Order Detection Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: Therefore, in an automated manner, we construct ReadingBank, a benchmark dataset that contains reading order, text, and layout information for 500,000 document images covering a wide spectrum of document types. |
Zilong Wang; Yiheng Xu; Lei Cui; Jingbo Shang; Furu Wei; | code |
390 | Region Under Discussion for Visual Dialog Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper we define what it means for a visual question to require dialog history and we release a subset of the Guesswhat?! |
Mauricio Mazuecos; Franco M. Luque; Jorge S?nchez; Hern?n Maina; Thomas Vadora; Luciana Benotti; | |
391 | Learning Grounded Word Meaning Representations on Similarity Graphs Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: This paper introduces a novel approach to learn visually grounded meaning representations of words as low-dimensional node embeddings on an underlying graph hierarchy. |
Mariella Dimiccoli; Herwig Wendt; Pau Batlle Franch; | |
392 | WhyAct: Identifying Action Reasons in Lifestyle Vlogs Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We introduce and make publicly available the WhyAct dataset, consisting of 1,077 visual actions manually annotated with their reasons. |
Oana Ignat; Santiago Castro; Hanwen Miao; Weiji Li; Rada Mihalcea; | |
393 | Genre As Weak Supervision for Cross-lingual Dependency Parsing Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: Specifically, we project treebank-level genre information to the finer-grained sentence level, with the goal to amplify information implicitly stored in unsupervised contextualized representations. |
Max M?ller-Eberstein; Rob van der Goot; Barbara Plank; | |
394 | On The Relation Between Syntactic Divergence and Zero-Shot Performance Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We explore the link between the extent to which syntactic relations are preserved in translation and the ease of correctly constructing a parse tree in a zero-shot setting. |
Ofir Arviv; Dmitry Nikolaev; Taelin Karidi; Omri Abend; | |
395 | Improved Latent Tree Induction with Distant Supervision Via Span Constraints Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this work, we present a technique that uses distant supervision in the form of span constraints (i.e. phrase bracketing) to improve performance in unsupervised constituency parsing. |
Zhiyang Xu; Andrew Drozdov; Jay Yoon Lee; Tim O?Gorman; Subendhu Rongali; Dylan Finkbeiner; Shilpa Suresh; Mohit Iyyer; Andrew McCallum; | |
396 | Aligning Multidimensional Worldviews and Discovering Ideological Differences Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: Extending the ability of word embedding models to capture the semantic and cultural characteristics of their training corpora, we propose a novel method for discovering the multifaceted ideological and worldview characteristics of communities. |
Jeremiah Milbauer; Adarsh Mathew; James Evans; | |
397 | Just Say No: Analyzing The Stance of Neural Dialogue Generation in Offensive Contexts Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: To better understand the dynamics of contextually offensive language, we investigate the stance of dialogue model responses in offensive Reddit conversations. |
Ashutosh Baheti; Maarten Sap; Alan Ritter; Mark Riedl; | |
398 | Multi-Modal Open-Domain Dialogue Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: With the goal of getting humans to engage in multi-modal dialogue, we investigate combining components from state-of-the-art open-domain dialogue agents with those from state-of-the-art vision models. |
Kurt Shuster; Eric Michael Smith; Da Ju; Jason Weston; | |
399 | A Label-Aware BERT Attention Network for Zero-Shot Multi-Intent Detection in Spoken Language Understanding Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: Motivated by the above, we propose a Label-Aware BERT Attention Network (LABAN) for zero-shot multi-intent detection. |
Ting-Wei Wu; Ruolin Su; Biing Juang; | |
400 | Zero-Shot Dialogue Disentanglement By Self-Supervised Entangled Response Selection Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we are the first to propose a zero-shot dialogue disentanglement solution. |
Ta-Chung Chi; Alexander Rudnicky; | |
401 | SIMMC 2.0: A Task-oriented Dialog Dataset for Immersive Multimodal Conversations Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: To overcome, we present a new dataset for Situated and Interactive Multimodal Conversations, SIMMC 2.0, which includes 11K task-oriented user<->assistant dialogs (117K utterances) in the shopping domain, grounded in immersive and photo-realistic scenes. |
Satwik Kottur; Seungwhan Moon; Alborz Geramifard; Babak Damavandi; | |
402 | RAST: Domain-Robust Dialogue Rewriting As Sequence Tagging Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We address this robustness issue by proposing a novel sequence-tagging-based model so that the search space is significantly reduced, yet the core of this task is still well covered. |
Jie Hao; Linfeng Song; Liwei Wang; Kun Xu; Zhaopeng Tu; Dong Yu; | |
403 | MRF-Chat: Improving Dialogue with Markov Random Fields Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this work, we propose a novel probabilistic approach using Markov Random Fields (MRF) to augment existing deep-learning methods for improved next utterance prediction. |
Ishaan Grover; Matthew Huggins; Cynthia Breazeal; Hae Won Park; | |
404 | Dialogue State Tracking with A Language Model Using Schema-Driven Prompting Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: Here, we introduce a new variation of the language modeling approach that uses schema-driven prompting to provide task-aware history encoding that is used for both categorical and non-categorical slots. |
Chia-Hsuan Lee; Hao Cheng; Mari Ostendorf; | code |
405 | Signed Coreference Resolution Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we: (1) introduce Signed Coreference Resolution (SCR), a new challenge for coreference modeling and Sign Language Processing; (2) collect an annotated corpus of German Sign Language with gold labels for coreference together with an annotation software for the task; (3) explore features of hand gesture, iconicity, and spatial situated properties and move forward to propose a set of linguistically informed heuristics and unsupervised models for the task; (4) put forward several proposals about ways to address the complexities of this challenge effectively. |
Kayo Yin; Kenneth DeHaan; Malihe Alikhani; | |
406 | Consistent Accelerated Inference Via Confident Adaptive Transformers Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this work, we present CATs – Confident Adaptive Transformers – in which we simultaneously increase computational efficiency, while guaranteeing a specifiable degree of consistency with the original model with high confidence. |
Tal Schuster; Adam Fisch; Tommi Jaakkola; Regina Barzilay; | |
407 | Improving and Simplifying Pattern Exploiting Training Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we focus on few-shot learning without any unlabeled data and introduce ADAPET, which modifies PET’s objective to provide denser supervision during fine-tuning. |
Derek Tam; Rakesh R. Menon; Mohit Bansal; Shashank Srivastava; Colin Raffel; | |
408 | Unsupervised Data Augmentation with Naive Augmentation and Without Unlabeled Data Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we re-examine UDA and demonstrate its efficacy on several sequential tasks. |
David Lowell; Brian Howard; Zachary C. Lipton; Byron Wallace; | |
409 | Pre-train or Annotate? Domain Adaptation with A Constrained Budget Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we study domain adaptation under budget constraints, and approach it as a customer choice problem between data annotation and pre-training. |
Fan Bai; Alan Ritter; Wei Xu; | |
410 | Lawyers Are Dishonest? Quantifying Representational Harms in Commonsense Knowledge Resources Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: Here we focus on two widely used CSKBs, ConceptNet and GenericsKB, and establish the presence of bias in the form of two types of representational harms, overgeneralization of polarized perceptions and representation disparity across different demographic groups in both CSKBs. |
Ninareh Mehrabi; Pei Zhou; Fred Morstatter; Jay Pujara; Xiang Ren; Aram Galstyan; | |
411 | OSCaR: Orthogonal Subspace Correction and Rectification of Biases in Word Embeddings Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: To address this challenge, we propose OSCaR (Orthogonal Subspace Correction and Rectification), a bias-mitigating method that focuses on disentangling biased associations between concepts instead of removing concepts wholesale. |
Sunipa Dev; Tao Li; Jeff M Phillips; Vivek Srikumar; | |
412 | Sentence-Permuted Paragraph Generation Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: Our idea is permuting the sentence orders to improve the content diversity of multi-sentence paragraph. |
Wenhao Yu; Chenguang Zhu; Tong Zhao; Zhichun Guo; Meng Jiang; | |
413 | Extract, Denoise and Enforce: Evaluating and Improving Concept Preservation for Text-to-Text Generation Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we present a systematic analysis that studies whether current seq2seq models, especially pre-trained language models, are good enough for preserving important input concepts and to what extent explicitly guiding generation with the concepts as lexical constraints is beneficial. |
Yuning Mao; Wenchang Ma; Deren Lei; Jiawei Han; Xiang Ren; | code |
414 | Paraphrase Generation: A Survey of The State of The Art Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: This paper focuses on paraphrase generation,which is a widely studied natural language generation task in NLP. |
Jianing Zhou; Suma Bhat; | |
415 | Exposure Bias Versus Self-Recovery: Are Distortions Really Incremental for Autoregressive Text Generation? Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this work, we focus on the task of open-ended language generation, propose metrics to quantify the impact of exposure bias in the aspects of quality, diversity, and consistency. |
Tianxing He; Jingzhao Zhang; Zhiming Zhou; James Glass; | |
416 | Generating Self-Contained and Summary-Centric Question Answer Pairs Via Differentiable Reward Imitation Learning Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: Motivated by suggested question generation in conversational news recommendation systems, we propose a model for generating question-answer pairs (QA pairs) with self-contained, summary-centric questions and length-constrained, article-summarizing answers. |
Li Zhou; Kevin Small; Yong Zhang; Sandeep Atluri; | |
417 | Unsupervised Paraphrasing with Pretrained Language Models Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: To address this drawback, we adopt a transfer learning approach and propose a training pipeline that enables pre-trained language models to generate high-quality paraphrases in an unsupervised setting. |
Tong Niu; Semih Yavuz; Yingbo Zhou; Nitish Shirish Keskar; Huan Wang; Caiming Xiong; | |
418 | Profanity-Avoiding Training Framework for Seq2seq Models with Certified Robustness Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: To address this problem, we propose a training framework with certified robustness to eliminate the causes that trigger the generation of profanity. |
Hengtong Zhang; Tianhang Zheng; Yaliang Li; Jing Gao; Lu Su; Bo Li; | |
419 | Journalistic Guidelines Aware News Image Captioning Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this work, we propose a new approach to this task, motivated by caption guidelines that journalists follow. |
Xuewen Yang; Svebor Karaman; Joel Tetreault; Alejandro Jaimes; | |
420 | AESOP: Paraphrase Generation with Adaptive Syntactic Control Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We propose to control paraphrase generation through carefully chosen target syntactic structures to generate more proper and higher quality paraphrases. |
Jiao Sun; Xuezhe Ma; Nanyun Peng; | |
421 | Refocusing on Relevance: Personalization in NLG Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this work, we argue that NLG systems in general should place a much higher level of emphasis on making use of additional context, and suggest that relevance (as used in Information Retrieval) be thought of as a crucial tool for designing user-oriented text-generating tasks. |
Shiran Dudy; Steven Bedrick; Bonnie Webber; | |
422 | The Future Is Not One-dimensional: Complex Event Schema Induction By Graph Modeling for Event Prediction Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We introduce a new concept of Temporal Complex Event Schema: a graph-based schema representation that encompasses events, arguments, temporal connections and argument relations. |
Manling Li; Sha Li; Zhenhailong Wang; Lifu Huang; Kyunghyun Cho; Heng Ji; Jiawei Han; Clare Voss; | |
423 | Learning Constraints and Descriptive Segmentation for Subevent Detection Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: To bridge the two tasks together, we propose an approach to learning and enforcing constraints that capture dependencies between subevent detection and EventSeg prediction, as well as guiding the model to make globally consistent inference. |
Haoyu Wang; Hongming Zhang; Muhao Chen; Dan Roth; | |
424 | ChemNER: Fine-Grained Chemistry Named Entity Recognition with Ontology-Guided Distant Supervision Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We propose ChemNER, an ontology-guided, distantly-supervised method for fine-grained chemistry NER to tackle these challenges. |
Xuan Wang; Vivian Hu; Xiangchen Song; Shweta Garg; Jinfeng Xiao; Jiawei Han; | |
425 | Moving on from OntoNotes: Coreference Resolution Model Transfer Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We aim to quantify transferability of coref models based on the number of annotated documents available in the target dataset. |
Patrick Xia; Benjamin Van Durme; | |
426 | Document-level Entity-based Extraction As Template Generation Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: To address this issue, we propose a generative framework for two document-level EE tasks: role-filler entity extraction (REE) and relation extraction (RE). |
Kung-Hsiang Huang; Sam Tang; Nanyun Peng; | |
427 | Learning Prototype Representations Across Few-Shot Tasks for Event Detection Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We propose to model the relations between training tasks in episodic few-shot learning by introducing cross-task prototypes. |
Viet Lai; Franck Dernoncourt; Thien Huu Nguyen; | code |
428 | Lifelong Event Detection with Knowledge Transfer Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In reality, the ontology of interest may change over time, adding emergent new types or more fine-grained subtypes. We propose a new lifelong learning framework to address this challenge. |
Pengfei Yu; Heng Ji; Prem Natarajan; | |
429 | Modular Self-Supervision for Document-Level Relation Extraction Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we propose decomposing document-level relation extraction into relation detection and argument resolution, taking inspiration from Davidsonian semantics. |
Sheng Zhang; Cliff Wong; Naoto Usuyama; Sarthak Jain; Tristan Naumann; Hoifung Poon; | |
430 | Unsupervised Paraphrasing Consistency Training for Low Resource Named Entity Recognition Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we explore the use of paraphrasing as a more principled data augmentation scheme for NER unsupervised consistency training. |
Rui Wang; Ricardo Henao; | |
431 | Fine-grained Entity Typing Without Knowledge Base Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: Under this setting, we propose a two-step framework to train FET models. |
Jing Qian; Yibin Liu; Lemao Liu; Yangming Li; Haiyun Jiang; Haisong Zhang; Shuming Shi; | |
432 | Adversarial Attack Against Cross-lingual Knowledge Graph Alignment Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: This paper proposes an adversarial attack model with two novel attack techniques to perturb the KG structure and degrade the quality of deep cross-lingual entity alignment. |
Zeru Zhang; Zijie Zhang; Yang Zhou; Lingfei Wu; Sixing Wu; Xiaoying Han; Dejing Dou; Tianshi Che; Da Yan; | |
433 | Towards Realistic Few-Shot Relation Extraction Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper we take a deeper look at the efficacy of strong few-shot classification models in the more common relation extraction setting, and show that typical few-shot evaluation metrics obscure a wide variability in performance across relations. |
Sam Brody; Sichao Wu; Adrian Benton; | |
434 | Data Augmentation for Cross-Domain Named Entity Recognition Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this work, we take this research direction to the opposite and study cross-domain data augmentation for the NER task. |
Shuguang Chen; Gustavo Aguilar; Leonardo Neves; Thamar Solorio; | |
435 | Incorporating Medical Knowledge in BERT for Clinical Relation Extraction Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: To solve these issues, in this research, we conduct a comprehensive examination of different techniques to add medical knowledge into a pre-trained BERT model for clinical relation extraction. |
Arpita Roy; Shimei Pan; | |
436 | ECONET: Effective Continual Pretraining of Language Models for Event Temporal Reasoning Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We present a continual pre-training approach that equips PTLMs with targeted knowledge about event temporal relations. |
Rujun Han; Xiang Ren; Nanyun Peng; | |
437 | Learning from Noisy Labels for Entity-Centric Information Extraction Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: Motivated by such properties, we propose a simple co-regularization framework for entity-centric information extraction, which consists of several neural models with identical structures but different parameter initialization. |
Wenxuan Zhou; Muhao Chen; | |
438 | Extracting Material Property Measurement Data from Scientific Articles Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this work, we describe a methodology for developing an automatic property extraction framework using material solubility as the target property. |
Gihan Panapitiya; Fred Parks; Jonathan Sepulveda; Emily Saldanha; | |
439 | Modeling Document-Level Context for Event Detection Via Important Context Selection Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: To address this issue, we propose a novel method to model document-level context for ED that dynamically selects relevant sentences in the document for the event prediction of the target sentence. |
Amir Pouran Ben Veyseh; Minh Van Nguyen; Nghia Ngo Trung; Bonan Min; Thien Huu Nguyen; | |
440 | Crosslingual Transfer Learning for Relation and Event Extraction Via Word Category and Class Alignments Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: To address this issue, we propose a novel crosslingual alignment method that leverages class information of REE tasks for representation learning. |
Minh Van Nguyen; Tuan Ngo Nguyen; Bonan Min; Thien Huu Nguyen; | |
441 | Corpus-based Open-Domain Event Type Induction Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: This work presents a corpus-based open-domain event type induction method that automatically discovers a set of event types from a given corpus. |
Jiaming Shen; Yunyi Zhang; Heng Ji; Jiawei Han; | |
442 | PDALN: Progressive Domain Adaptation Over A Pre-trained Model for Low-Resource Cross-Domain Named Entity Recognition Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: To address these challenges, we propose a progressive domain adaptation Knowledge Distillation (KD) approach – PDALN. |
Tao Zhang; Congying Xia; Philip S. Yu; Zhiwei Liu; Shu Zhao; | |
443 | Multi-Vector Attention Models for Deep Re-ranking Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we present a lightweight architecture that explores this joint cost vs. accuracy trade-off based on multi-vector attention (MVA). |
Giulio Zhou; Jacob Devlin; | |
444 | Toward Deconfounding The Effect of Entity Demographics for Question Answering Accuracy Related Papers Related Patents Related Grants Related Orgs Related Experts Details Abstract: The goal of question answering (QA) is to answer _any_ question. However, major QA datasets have skewed distributions over gender, profession, and nationality. Despite that skew, … |
Maharshi Gor; Kellie Webster; Jordan Boyd-Graber; | |
445 | Exploring Strategies for Generalizable Commonsense Reasoning with Pre-trained Models Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we investigate what models learn from commonsense reasoning datasets. |
Kaixin Ma; Filip Ilievski; Jonathan Francis; Satoru Ozaki; Eric Nyberg; Alessandro Oltramari; | |
446 | Transformer Feed-Forward Layers Are Key-Value Memories Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We show that feed-forward layers in transformer-based language models operate as key-value memories, where each key correlates with textual patterns in the training examples, and each value induces a distribution over the output vocabulary. |
Mor Geva; Roei Schuster; Jonathan Berant; Omer Levy; | |
447 | Connecting Attributions and QA Model Behavior on Realistic Counterfactuals Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: This paper investigates how well different attribution techniques align with this assumption on realistic counterfactuals in the case of reading comprehension (RC). |
Xi Ye; Rohan Nair; Greg Durrett; | |
448 | How Do Neural Sequence Models Generalize? Local and Global Cues for Out-of-Distribution Prediction Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We show that RNN and transformer language models exhibit structured, consistent generalization in out-of-distribution contexts. |
D. Anthony Bau; Jacob Andreas; | |
449 | Comparing Text Representations: A Theory-Driven Approach Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We adapt general tools from computational learning theory to fit the specific characteristics of text datasets and present a method to evaluate the compatibility between representations and tasks. |
Gregory Yauney; David Mimno; | |
450 | Human Rationales As Attribution Priors for Explainable Stance Detection Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this work, we present a method for imparting human-like rationalization to a stance detection model using crowdsourced annotations on a small fraction of the training data. |
Sahil Jayaram; Emily Allaway; | |
451 | The Stem Cell Hypothesis: Dilemma Behind Multi-Task Learning with Transformer Encoders Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: Based on this finding, we propose the Stem Cell Hypothesis to reveal the existence of attention heads naturally talented for many tasks that cannot be jointly trained to create adequate embeddings for all of those tasks. |
Han He; Jinho D. Choi; | |
452 | Text Counterfactuals Via Latent Optimization and Shapley-Guided Search Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: Given a textual input and a classification model, we aim to minimally alter the text to change the model’s prediction. |
Xiaoli Fern; Quintin Pope; | |
453 | Average Approximates First Principal Component- An Empirical Analysis on Representations from Neural Language Models Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we present an empirical property of these representations-average approximates first principal component. |
Zihan Wang; Chengyu Dong; Jingbo Shang; | |
454 | Controlled Evaluation of Grammatical Knowledge in Mandarin Chinese Language Models Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: However, it remains unclear if such an inductive bias would also improve language models’ ability to learn grammatical dependencies in typologically different languages. Here we investigate this question in Mandarin Chinese, which has a logographic, largely syllable-based writing system; different word order; and sparser morphology than English. |
Yiwen Wang; Jennifer Hu; Roger Levy; Peng Qian; | |
455 | GradTS: A Gradient-Based Automatic Auxiliary Task Selection Method Based on Transformer Networks Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: This paper presents GradTS, an automatic auxiliary task selection method based on gradient calculation in Transformer-based models. |
Weicheng Ma; Renze Lou; Kai Zhang; Lili Wang; Soroush Vosoughi; | |
456 | NegatER: Unsupervised Discovery of Negatives in Commonsense Knowledge Bases Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: As a first step toward the latter, this paper proposes NegatER, a framework that ranks potential negatives in commonsense KBs using a contextual language model (LM). |
Tara Safavi; Jing Zhu; Danai Koutra; | |
457 | Instance-adaptive Training with Noise-robust Losses Against Noisy Labels Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: This work proposes novel instance-adaptive training frameworks to change single dataset-wise hyperparameters of noise resistance in such losses to be instance-wise. |
Lifeng Jin; Linfeng Song; Kun Xu; Dong Yu; | |
458 | Distributionally Robust Multilingual Machine Translation Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we propose a new learning objective for MNMT based on distributionally robust optimization, which minimizes the worst-case expected loss over the set of language pairs. |
Chunting Zhou; Daniel Levy; Xian Li; Marjan Ghazvininejad; Graham Neubig; | |
459 | Model Selection for Cross-lingual Transfer Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we show that it is possible to select consistently better models when small amounts of annotated data are available in auxiliary pivot languages. |
Yang Chen; Alan Ritter; | |
460 | Continual Few-Shot Learning for Text Classification Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this work, we propose a continual few-shot learning (CFL) task, in which a system is challenged with a difficult phenomenon and asked to learn to correct mistakes with only a few (10 to 15) training examples. |
Ramakanth Pasunuru; Veselin Stoyanov; Mohit Bansal; | |
461 | Efficient Nearest Neighbor Language Models Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we take the recently proposed k-nearest neighbors language model as an example, exploring methods to improve its efficiency along various dimensions. |
Junxian He; Graham Neubig; Taylor Berg-Kirkpatrick; | |
462 | STraTA: Self-Training with Task Augmentation for Better Few-shot Learning Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: To address this shortcoming, we propose STraTA, which stands for Self-Training with Task Augmentation, an approach that builds on two key ideas for effective leverage of unlabeled data. |
Tu Vu; Minh-Thang Luong; Quoc Le; Grady Simon; Mohit Iyyer; | |
463 | TADPOLE: Task ADapted Pre-Training Via AnOmaLy DEtection Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we address the problem for the case when the downstream corpus is too small for additional pre-training. |
Vivek Madan; Ashish Khetan; Zohar Karnin; | |
464 | Gradient-based Adversarial Attacks Against Text Transformers Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We propose the first general-purpose gradient-based adversarial attack against transformer models. |
Chuan Guo; Alexandre Sablayrolles; Herv? J?gou; Douwe Kiela; | |
465 | Do Transformer Modifications Transfer Across Implementations and Applications? Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we comprehensively evaluate many of these modifications in a shared experimental setting that covers most of the common uses of the Transformer in natural language processing. |
Sharan Narang; Hyung Won Chung; Yi Tay; Liam Fedus; Thibault Fevry; Michael Matena; Karishma Malkan; Noah Fiedel; Noam Shazeer; Zhenzhong Lan; Yanqi Zhou; Wei Li; Nan Ding; Jake Marcus; Adam Roberts; Colin Raffel; | |
466 | Paired Examples As Indirect Supervision in Latent Decision Models Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this work, we introduce a way to leverage paired examples that provide stronger cues for learning latent decisions. |
Nitish Gupta; Sameer Singh; Matt Gardner; Dan Roth; | |
467 | Pairwise Supervised Contrastive Learning of Sentence Representations Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we propose PairSupCon, an instance discrimination based approach aiming to bridge semantic entailment and contradiction understanding with high-level categorical concept encoding. |
Dejiao Zhang; Shang-Wen Li; Wei Xiao; Henghui Zhu; Ramesh Nallapati; Andrew O. Arnold; Bing Xiang; | |
468 | Muppet: Massive Multi-task Representations with Pre-Finetuning Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We propose pre-finetuning, an additional large-scale learning stage between language model pre-training and fine-tuning. |
Armen Aghajanyan; Anchit Gupta; Akshat Shrivastava; Xilun Chen; Luke Zettlemoyer; Sonal Gupta; | |
469 | Diverse Distributions of Self-Supervised Tasks for Meta-Learning in NLP Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this work, we aim to provide task distributions for meta-learning by considering self-supervised tasks automatically proposed from unlabeled text, to enable large-scale meta-learning in NLP. |
Trapit Bansal; Karthick Prasad Gunasekaran; Tong Wang; Tsendsuren Munkhdalai; Andrew McCallum; | |
470 | A Simple and Effective Method To Eliminate The Self Language Bias in Multilingual Representations Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We explore this problem from a novel angle of geometric algebra and semantic space. |
Ziyi Yang; Yinfei Yang; Daniel Cer; Eric Darve; | |
471 | A Massively Multilingual Analysis of Cross-linguality in Shared Embedding Space Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: Here, we investigate the linguistic and non-linguistic factors affecting sentence-level alignment in cross-lingual pretrained language models for 101 languages and 5,050 language pairs. |
Alexander Jones; William Yang Wang; Kyle Mahowald; | |
472 | Frustratingly Simple But Surprisingly Strong: Using Language-Independent Features for Zero-shot Cross-lingual Semantic Parsing Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: To this end, we experiment with six Discourse Representation Structure (DRS) semantic parsers in English, and generalize them to Italian, German and Dutch, where there are only a small number of manually annotated parses available. |
Jingfeng Yang; Federico Fancellu; Bonnie Webber; Diyi Yang; | |
473 | Improving Simultaneous Translation By Incorporating Pseudo-References with Fewer Reorderings Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We propose a novel method that rewrites the target side of existing full-sentence corpora into simultaneous-style translation. |
Junkun Chen; Renjie Zheng; Atsuhito Kita; Mingbo Ma; Liang Huang; | |
474 | Classification-based Quality Estimation: Small and Efficient Models for Real-world Applications Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this work, we evaluate several model compression techniques for QE and find that, despite their popularity in other NLP tasks, they lead to poor performance in this regression setting. |
Shuo Sun; Ahmed El-Kishky; Vishrav Chaudhary; James Cross; Lucia Specia; Francisco Guzm?n; | |
475 | A Large-Scale Study of Machine Translation in Turkic Languages Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we provide the first large-scale case study of the practical application of MT in the Turkic language family in order to realize the gains of NMT for Turkic languages under high-resource to extremely low-resource scenarios. |
Jamshidbek Mirzakhalov; Anoop Babu; Duygu Ataman; Sherzod Kariev; Francis Tyers; Otabek Abduraufov; Mammad Hajili; Sardana Ivanova; Abror Khaytbaev; Antonio Laverghetta Jr.; Bekhzodbek Moydinboyev; Esra Onal; Shaxnoza Pulatova; Ahsan Wahab; Orhan Firat; Sriram Chellappan; | |
476 | Analyzing The Surprising Variability in Word Embedding Stability Across Languages Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: To gain further insight into word embeddings, we explore their stability (e.g., overlap between the nearest neighbors of a word in different embedding spaces) in diverse languages. |
Laura Burdick; Jonathan K. Kummerfeld; Rada Mihalcea; | |
477 | Rule-based Morphological Inflection Improves Neural Terminology Translation Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we introduce a modular framework for incorporating lemma constraints in neural MT (NMT) in which linguistic knowledge and diverse types of NMT models can be flexibly applied. |
Weijia Xu; Marine Carpuat; | |
478 | Data and Parameter Scaling Laws for Neural Machine Translation Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We observe that the development cross-entropy loss of supervised neural machine translation models scales like a power law with the amount of training data and the number of non-embedding parameters in the model. |
Mitchell A Gordon; Kevin Duh; Jared Kaplan; | |
479 | Good-Enough Example Extrapolation Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: To operationalize this question, I propose a simple data augmentation protocol called good-enough example extrapolation (GE3). |
Jason Wei; | |
480 | Learning to Selectively Learn for Weakly-supervised Paraphrase Generation Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this work, we go beyond the existing paradigms and propose a novel approach to generate high-quality paraphrases with data of weak supervision. |
Kaize Ding; Dingcheng Li; Alexander Hanbo Li; Xing Fan; Chenlei Guo; Yang Liu; Huan Liu; | |
481 | Effective Convolutional Attention Network for Multi-label Clinical Document Classification Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we present an effective convolutional attention network for the MLDC problem with a focus on medical code prediction from clinical documents. |
Yang Liu; Hua Cheng; Russell Klopfer; Matthew R. Gormley; Thomas Schaaf; | |
482 | Contrastive Code Representation Learning Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We propose ContraCode: a contrastive pre-training task that learns code functionality, not form. |
Paras Jain; Ajay Jain; Tianjun Zhang; Pieter Abbeel; Joseph Gonzalez; Ion Stoica; | code |
483 | IGA: An Intent-Guided Authoring Assistant Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We leverage advances in language modeling to build an interactive writing assistant that generates and rephrases text according to fine-grained author specifications. |
Simeng Sun; Wenlong Zhao; Varun Manjunatha; Rajiv Jain; Vlad Morariu; Franck Dernoncourt; Balaji Vasan Srinivasan; Mohit Iyyer; | |
484 | Math Word Problem Generation with Mathematical Consistency and Problem Context Constraints Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we develop a novel MWP generation approach that leverages i) pre-trained language models and a context keyword selection model to improve the language quality of generated MWPs and ii) an equation consistency constraint for math equations to improve the mathematical validity of the generated MWPs. |
Zichao Wang; Andrew Lan; Richard Baraniuk; | |
485 | Navigating The Kaleidoscope of COVID-19 Misinformation Using Deep Learning Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: By conducting a systematic investigation, we show that: (i) the deep Transformer-based pre-trained models, utilized via the mixed-domain transfer learning, are only good at capturing the local context, thus exhibits poor generalization, and (ii) a combination of shallow network-based domain-specific models and convolutional neural networks can efficiently extract local as well as global context directly from the target data in a hierarchical fashion, enabling it to offer a more generalizable solution. |
Yuanzhi Chen; Mohammad Hasan; | |
486 | Detecting Health Advice in Medical Research Literature Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: This study fills the gap by developing and validating an NLP-based prediction model for identifying health advice in research publications. |
Yingya Li; Jun Wang; Bei Yu; | |
487 | A Semantic Feature-Wise Transformation Relation Network for Automatic Short Answer Grading Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We propose a Semantic Feature-wise transformation Relation Network (SFRN) that exploits the multiple components of ASAG datasets more effectively. |
Zhaohui Li; Yajur Tomar; Rebecca J. Passonneau; | |
488 | Evaluating Scholarly Impact: Towards Content-Aware Bibliometrics Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We present approaches to estimate content-aware bibliometrics to quantitatively measure the scholarly impact of a publication. |
Saurav Manchanda; George Karypis; | |
489 | A Scalable Framework for Learning From Implicit User Feedback to Improve Natural Language Understanding in Large-Scale Conversational AI Systems Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We propose a scalable and automatic approach for improving NLU in a large-scale conversational AI system by leveraging implicit user feedback, with an insight that user interaction data and dialog context have rich information embedded from which user satisfaction and intention can be inferred. |
Sunghyun Park; Han Li; Ameen Patel; Sidharth Mudgal; Sungjin Lee; Young-Bum Kim; Spyros Matsoukas; Ruhi Sarikaya; | |
490 | Summarize-then-Answer: Generating Concise Explanations for Multi-hop Reading Comprehension Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: Instead, we advocate for an abstractive approach, where we propose to generate a question-focused, abstractive summary of input paragraphs and then feed it to an RC system. |
Naoya Inoue; Harsh Trivedi; Steven Sinha; Niranjan Balasubramanian; Kentaro Inui; | |
491 | FewshotQA: A Simple Framework for Few-shot Learning of Question Answering Tasks Using Pre-trained Text-to-text Models Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: To address this, we propose a simple fine-tuning framework that leverages pre-trained text-to-text models and is directly aligned with their pre-training framework. |
Rakesh Chada; Pradeep Natarajan; | |
492 | Multi-stage Training with Improved Negative Contrast for Neural Passage Retrieval Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We propose a multi-stage framework comprising of pre-training with synthetic data, fine-tuning with labeled data, and negative sampling at both stages. |
Jing Lu; Gustavo Hernandez Abrego; Ji Ma; Jianmo Ni; Yinfei Yang; | |
493 | Perhaps PTLMs Should Go to School – A Task to Assess Open Book and Closed Book QA Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: Our goal is to deliver a new task and leaderboard to stimulate research on question answering and pre-trained language models (PTLMs) to understand a significant instructional document, e.g., an introductory college textbook or a manual. |
Manuel Ciosici; Joe Cecil; Dong-Ho Lee; Alex Hedges; Marjorie Freedman; Ralph Weischedel; | |
494 | ReasonBERT: Pre-trained to Reason with Distant Supervision Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We present ReasonBert, a pre-training method that augments language models with the ability to reason over long-range relations and multiple, possibly hybrid contexts. |
Xiang Deng; Yu Su; Alyssa Lees; You Wu; Cong Yu; Huan Sun; | |
495 | Single-dataset Experts for Multi-dataset Question Answering Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: Our approach is to model multi-dataset question answering with an ensemble of single-dataset experts, by training a collection of lightweight, dataset-specific adapter modules (Houlsby et al., 2019) that share an underlying Transformer model. |
Dan Friedman; Ben Dodge; Danqi Chen; | |
496 | Simple Entity-Centric Questions Challenge Dense Retrievers Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We investigate this issue and uncover that dense retrievers can only generalize to common entities unless the question pattern is explicitly observed during training. |
Christopher Sciavolino; Zexuan Zhong; Jinhyuk Lee; Danqi Chen; | |
497 | Mitigating False-Negative Contexts in Multi-document Question Answering with Retrieval Marginalization Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We develop a new parameterization of set-valued retrieval that handles unanswerable queries, and we show that marginalizing over this set during training allows a model to mitigate false negatives in supporting evidence annotations. |
Ansong Ni; Matt Gardner; Pradeep Dasigi; | |
498 | MultiDoc2Dial: Modeling Dialogues Grounded in Multiple Documents Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this work, we aim to address more realistic scenarios where a goal-oriented information-seeking conversation involves multiple topics, and hence is grounded on different documents. |
Song Feng; Siva Sankalp Patel; Hui Wan; Sachindra Joshi; | |
499 | GupShup: Summarizing Open-Domain Code-Switched Conversations Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: Towards this objective, we introduce the task of abstractive summarization of Hindi-English (Hi-En) code-switched conversations. |
Laiba Mehnaz; Debanjan Mahata; Rakesh Gosangi; Uma Sushmitha Gunturi; Riya Jain; Gauri Gupta; Amardeep Kumar; Isabelle G. Lee; Anish Acharya; Rajiv Ratn Shah; | |
500 | BiSECT: Learning to Split and Rephrase Sentences with Bitexts Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We introduce a novel dataset and a new model for this ‘split and rephrase’ task. |
Joongwon Kim; Mounica Maddela; Reno Kriz; Wei Xu; Chris Callison-Burch; | |
501 | Data Collection Vs. Knowledge Graph Completion: What Is Needed to Improve Coverage? Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: This survey/position paper discusses ways to improve coverage of resources such as WordNet. |
Kenneth Church; Yuchen Bian; | |
502 | Universal Sentence Representation Learning with Conditional Masked Language Model Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: This paper presents a novel training method, Conditional Masked Language Modeling (CMLM), to effectively learn sentence representations on large scale unlabeled corpora. |
Ziyi Yang; Yinfei Yang; Daniel Cer; Jax Law; Eric Darve; | |
503 | On The Benefit of Syntactic Supervision for Cross-lingual Transfer in Semantic Role Labeling Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this work, we perform an empirical exploration of the helpfulness of syntactic supervision for crosslingual SRL within a simple multitask learning scheme. |
Zhisong Zhang; Emma Strubell; Eduard Hovy; | |
504 | Implicit Premise Generation with Discourse-aware Commonsense Knowledge Models Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We tackle the task of generating the implicit premise in an enthymeme, which requires not only an understanding of the stated conclusion and premise but also additional inferences that could depend on commonsense knowledge. |
Tuhin Chakrabarty; Aadit Trivedi; Smaranda Muresan; | |
505 | Inducing Transformer’s Compositional Generalization Ability Via Auxiliary Sequence Prediction Tasks Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: Motivated by the failure of a Transformer model on the SCAN compositionality challenge (Lake and Baroni, 2018), which requires parsing a command into actions, we propose two auxiliary sequence prediction tasks as additional training supervision. |
Yichen Jiang; Mohit Bansal; | |
506 | Flexible Generation of Natural Language Deductions Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we describe ParaPattern, a method for building models to generate deductive inferences from diverse natural language inputs without direct human supervision. |
Kaj Bostrom; Xinyu Zhao; Swarat Chaudhuri; Greg Durrett; | |
507 | Structure-aware Fine-tuning of Sequence-to-sequence Transformers for Transition-based AMR Parsing Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this work we explore the integration of general pre-trained sequence-to-sequence language models and a structure-aware transition-based approach. |
Jiawei Zhou; Tahira Naseem; Ram?n Fernandez Astudillo; Young-Suk Lee; Radu Florian; Salim Roukos; | |
508 | Think About It! Improving Defeasible Reasoning By First Modeling The Question Scenario Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: Our research goal asks whether neural models can similarly benefit from envisioning the question scenario before answering a defeasible query. |
Aman Madaan; Niket Tandon; Dheeraj Rajagopal; Peter Clark; Yiming Yang; Eduard Hovy; | |
509 | Open Aspect Target Sentiment Classification with Natural Language Prompts Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: To address this, we propose simple approaches that better solve ATSC with natural language prompts, enabling the task under zero-shot cases and enhancing supervised settings, especially for few-shot cases. |
Ronald Seoh; Ian Birle; Mrinal Tak; Haw-Shiuan Chang; Brian Pinette; Alfred Hough; | |
510 | Does BERT Learn As Humans Perceive? Understanding Linguistic Styles Through Lexica Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this study, we investigate lexicon usages across styles throughout two lenses: human perception and machine word importance, since words differ in the strength of the stylistic cues that they provide. |
Shirley Hayati; Dongyeop Kang; Lyle Ungar; | |
511 | Improving Stance Detection with Multi-Dataset Learning and Knowledge Distillation Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: Second, we investigate the knowledge distillation in stance detection and observe that transferring knowledge from a teacher model to a student model can be beneficial in our proposed training settings. |
Yingjie Li; Chenye Zhao; Cornelia Caragea; | |
512 | Discovering The Unknown Knowns: Turning Implicit Knowledge in The Dataset Into Explicit Training Examples for Visual Question Answering Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: Building upon these insights, we present a simple data augmentation pipeline SimpleAug to turn this known knowledge into training examples for VQA. |
Jihyung Kil; Cheng Zhang; Dong Xuan; Wei-Lun Chao; | code |
513 | Improving Pre-trained Vision-and-Language Embeddings for Phrase Grounding Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: To this end, we propose a method to extract matched phrase-region pairs from pre-trained vision-and-language embeddings and propose four fine-tuning objectives to improve the model phrase grounding ability using image-caption data without any supervised grounding signals. |
Zi-Yi Dou; Nanyun Peng; | |
514 | Sequential Randomized Smoothing for Adversarially Robust Speech Recognition Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: Our paper overcomes some of these challenges by leveraging speech-specific tools like enhancement and ROVER voting to design an ASR model that is robust to perturbations. |
Raphael Olivier; Bhiksha Raj; | |
515 | Hitting Your MARQ: Multimodal ARgument Quality Assessment in Long Debate Video Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we study argument quality assessment in a multimodal context, and experiment on DBATES, a publicly available dataset of long debate videos. |
Md Kamrul Hasan; James Spann; Masum Hasan; Md Saiful Islam; Kurtis Haut; Rada Mihalcea; Ehsan Hoque; | |
516 | Mind The Context: The Impact of Contextualization in Neural Module Networks for Grounding Visual Referring Expressions Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this work, we address these limitations and evaluate the impact of contextual clues in improving the performance of NMN models. |
Arjun Akula; Spandana Gella; Keze Wang; Song-Chun Zhu; Siva Reddy; | |
517 | Weakly-Supervised Visual-Retriever-Reader for Knowledge-based Question Answering Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We introduce various ways to retrieve knowledge using text and images and two reader styles: classification and extraction. |
Man Luo; Yankai Zeng; Pratyay Banerjee; Chitta Baral; | code |
518 | NDH-Full: Learning and Evaluating Navigational Agents on Full-Length Dialogue Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we explore the Navigation from Dialogue History (NDH) task, which is based on the Cooperative Vision-and-Dialogue Navigation (CVDN) dataset, and present a state-of-the-art model which is built upon Vision-Language transformers. |
Hyounghun Kim; Jialu Li; Mohit Bansal; | |
519 | Timeline Summarization Based on Event Graph Compression Via Time-Aware Optimal Transport Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: Following a different route, we propose to represent the news articles as an event-graph, thus the summarization becomes compressing the whole graph to its salient sub-graph. |
Manling Li; Tengfei Ma; Mo Yu; Lingfei Wu; Tian Gao; Heng Ji; Kathleen McKeown; | |
520 | StreamHover: Livestream Transcript Summarization and Annotation Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we present StreamHover, a framework for annotating and summarizing livestream transcripts. |
Sangwoo Cho; Franck Dernoncourt; Tim Ganter; Trung Bui; Nedim Lipka; Walter Chang; Hailin Jin; Jonathan Brandt; Hassan Foroosh; Fei Liu; | |
521 | Cross-Register Projection for Headline Part of Speech Tagging Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this work, we automatically annotate news headlines with POS tags by projecting predicted tags from corresponding sentences in news bodies. |
Adrian Benton; Hanyang Li; Igor Malioutov; | |
522 | Editing Factual Knowledge in Language Models Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We present KnowledgeEditor, a method which can be used to edit this knowledge and, thus, fix ‘bugs’ or unexpected predictions without the need for expensive re-training or fine-tuning. |
Nicola De Cao; Wilker Aziz; Ivan Titov; | code |
523 | Sparse Attention with Linear Units Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this work, we introduce a novel, simple method for achieving sparsity in attention: we replace the softmax activation with a , and show that sparsity naturally emerges from such a formulation. |
Biao Zhang; Ivan Titov; Rico Sennrich; | |
524 | Knowledge Base Completion Meets Transfer Learning Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this work, we introduce the first approach for transfer of knowledge from one collection of facts to another without the need for entity or relation matching. |
Vid Kocijan; Thomas Lukasiewicz; | |
525 | SPECTRA: Sparse Structured Text Rationalization Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we present a unified framework for deterministic extraction of structured explanations via constrained inference on a factor graph, forming a differentiable layer. |
Nuno M. Guerreiro; Andr? F. T. Martins; | |
526 | Towards Zero-Shot Knowledge Distillation for Natural Language Processing Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We present, to the best of our knowledge, the first work on Zero-shot Knowledge Distillation for NLP, where the student learns from the much larger teacher without any task specific data. |
Ahmad Rashid; Vasileios Lioutas; Abbas Ghaddar; Mehdi Rezagholizadeh; | |
527 | Adversarial Regularization As Stackelberg Game: An Unrolled Optimization Approach Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: To address this issue, we propose Stackelberg Adversarial Regularization (SALT), which formulates adversarial regularization as a Stackelberg game. |
Simiao Zuo; Chen Liang; Haoming Jiang; Xiaodong Liu; Pengcheng He; Jianfeng Gao; Weizhu Chen; Tuo Zhao; | |
528 | Aspect-Controllable Opinion Summarization Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we propose an approach that allows the generation of customized summaries based on aspect queries (e.g., describing the location and room of a hotel). |
Reinald Kim Amplayo; Stefanos Angelidis; Mirella Lapata; | |
529 | QuestEval: Summarization Asks for Fact-based Evaluation Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we extend previous approaches and propose a unified framework, named QuestEval. |
Thomas Scialom; Paul-Alexis Dray; Sylvain Lamprier; Benjamin Piwowarski; Jacopo Staiano; Alex Wang; Patrick Gallinari; | |
530 | Simple Conversational Data Augmentation for Semi-supervised Abstractive Dialogue Summarization Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: To reduce the dependence on labeled summaries, in this work, we present a simple yet effective set of Conversational Data Augmentation (CODA) methods for semi-supervised abstractive conversation summarization, such as random swapping/deletion to perturb the discourse relations inside conversations, dialogue-acts-guided insertion to interrupt the development of conversations, and conditional-generation-based substitution to substitute utterances with their paraphrases generated based on the conversation context. |
Jiaao Chen; Diyi Yang; | |
531 | Finding A Balanced Degree of Automation for Summary Evaluation Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this work, we propose flexible semiautomatic to automatic summary evaluation metrics, following the Pyramid human evaluation method. |
Shiyue Zhang; Mohit Bansal; | |
532 | CLIFF: Contrastive Learning for Improving Faithfulness and Factuality in Abstractive Summarization Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We study generating abstractive summaries that are faithful and factually consistent with the given articles. |
Shuyang Cao; Lu Wang; | |
533 | Multilingual Unsupervised Neural Machine Translation with Denoising Adapters Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper we propose instead to use _denoising adapters_, adapter layers with a denoising objective, on top of pre-trained mBART-50. |
Ahmet ?st?n; Alexandre Berard; Laurent Besacier; Matthias Gall?; | |
534 | BERT, MBERT, or BiBERT? A Study on Contextualized Embeddings for Neural Machine Translation Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we demonstrate that simply using the output (contextualized embeddings) of a tailored and suitable bilingual pre-trained language model (dubbed BiBERT) as the input of the NMT encoder achieves state-of-the-art translation performance. |
Haoran Xu; Benjamin Van Durme; Kenton Murray; | |
535 | Controlling Machine Translation for Multiple Attributes with Additive Interventions Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We address these problems by introducing vector-valued interventions which allow for fine-grained control over multiple attributes simultaneously via a weighted linear combination of the corresponding vectors. |
Andrea Schioppa; David Vilar; Artem Sokolov; Katja Filippova; | |
536 | A Generative Framework for Simultaneous Machine Translation Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We propose a generative framework for simultaneous machine translation. |
Yishu Miao; Phil Blunsom; Lucia Specia; | |
537 | It Is Not As Good As You Think! Evaluating Simultaneous Machine Translation on Interpretation Data Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In the absence of interpretation training data, we propose a translation-to-interpretation (T2I) style transfer method which allows converting existing offline translations into interpretation-style data, leading to up-to 2.8 BLEU improvement. |
Jinming Zhao; Philip Arthur; Gholamreza Haffari; Trevor Cohn; Ehsan Shareghi; | |
538 | Boosting Cross-Lingual Transfer Via Self-Learning with Uncertainty Estimation Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this work, we propose a self-learning framework that further utilizes unlabeled data of target languages, combined with uncertainty estimation in the process to select high-quality silver labels. |
Liyan Xu; Xuchao Zhang; Xujiang Zhao; Haifeng Chen; Feng Chen; Jinho D. Choi; | |
539 | Levenshtein Training for Word-level Quality Estimation Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We propose a novel scheme to use the Levenshtein Transformer to perform the task of word-level quality estimation. |
Shuoyang Ding; Marcin Junczys-Dowmunt; Matt Post; Philipp Koehn; | |
540 | Interactive Machine Comprehension with Dynamic Knowledge Graphs Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We describe methods that dynamically build and update these graphs during information gathering, as well as neural models to encode graph representations in RL agents. |
Xingdi Yuan; | |
541 | Residual Adapters for Parameter-Efficient ASR Adaptation to Atypical and Accented Speech Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We show that by adding a relatively small number of extra parameters to the encoder layers via so-called residual adapter, we can achieve similar adaptation gains compared to model fine-tuning, while only updating a tiny fraction (less than 0.5%) of the model parameters. |
Katrin Tomanek; Vicky Zayats; Dirk Padfield; Kara Vaillancourt; Fadi Biadsy; | |
542 | Visual News: Benchmark and Challenges in News Image Captioning Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We propose Visual News Captioner, an entity-aware model for the task of news image captioning. |
Fuxiao Liu; Yinghan Wang; Tianlu Wang; Vicente Ordonez; | |
543 | Integrating Visuospatial, Linguistic, and Commonsense Structure Into Story Visualization Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we first explore the use of constituency parse trees using a Transformer-based recurrent architecture for encoding structured input. Second, we augment the structured input with commonsense information and study the impact of this external knowledge on the generation of visual story. Third, we also incorporate visual structure via bounding boxes and dense captioning to provide feedback about the characters/objects in generated images within a dual learning setup. |
Adyasha Maharana; Mohit Bansal; | |
544 | VideoCLIP: Contrastive Pre-training for Zero-shot Video-Text Understanding Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We present VideoCLIP, a contrastive approach to pre-train a unified model for zero-shot video and text understanding, without using any labels on downstream tasks. |
Hu Xu; Gargi Ghosh; Po-Yao Huang; Dmytro Okhonko; Armen Aghajanyan; Florian Metze; Luke Zettlemoyer; Christoph Feichtenhofer; | code |
545 | NewsCLIPpings: Automatic Generation of Out-of-Context Multimodal Media Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We introduce several strategies for automatically retrieving convincing images for a given caption, capturing cases with inconsistent entities or semantic context. |
Grace Luo; Trevor Darrell; Anna Rohrbach; | |
546 | Powering Comparative Classification with Sentiment Analysis Via Domain Adaptive Knowledge Transfer Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We propose Sentiment Analysis Enhanced COmparative Network (SAECON) which improves CPC accuracy with a sentiment analyzer that learns sentiments to individual entities via domain adaptive knowledge transfer. |
Zeyu Li; Yilong Qin; Zihan Liu; Wei Wang; | |
547 | Tribrid: Stance Classification with Neural Inconsistency Detection Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: To improve the performance, we present a new neural architecture where the input also includes automatically generated negated perspectives over a given claim. |
Song Yang; Jacopo Urbani; | |
548 | SYSML: StYlometry with Structure and Multitask Learning: Implications for Darknet Forum Migrant Analysis Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We develop a novel stylometry-based multitask learning approach for natural language and model interactions using graph embeddings to construct low-dimensional representations of short episodes of user activity for authorship attribution. |
Pranav Maneriker; Yuntian He; Srinivasan Parthasarathy; | |
549 | Few-Shot Emotion Recognition in Conversation with Sequential Prototypical Networks Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this work, we place ourselves in the scope of a live chat customer service in which we want to detect emotions and their evolution in the conversation flow. |
Ga?l Guibon; Matthieu Labeau; H?l?ne Flamein; Luce Lefeuvre; Chlo? Clavel; | |
550 | CLASSIC: Continual and Contrastive Learning of Aspect Sentiment Classification Tasks Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: The key novelty is a contrastive continual learning method that enables both knowledge transfer across tasks and knowledge distillation from old tasks to the new task, which eliminates the need for task ids in testing. |
Zixuan Ke; Bing Liu; Hu Xu; Lei Shu; | |
551 | Implicit Sentiment Analysis with Event-centered Text Representation Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we focus on event-centric implicit sentiment analysis that utilizes the sentiment-aware event contained in a sentence to infer its sentiment polarity. |
Deyu Zhou; Jianan Wang; Linhai Zhang; Yulan He; | |
552 | SimCSE: Simple Contrastive Learning of Sentence Embeddings Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: This paper presents SimCSE, a simple contrastive learning framework that greatly advances the state-of-the-art sentence embeddings. |
Tianyu Gao; Xingcheng Yao; Danqi Chen; | |
553 | When Is Wall A Pared and When A Muro?: Extracting Rules Governing Lexical Selection Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this work, we present a method for automatically identifying fine-grained lexical distinctions, and extracting rules explaining these distinctions in a human- and machine-readable format. |
Aditi Chaudhary; Kayo Yin; Antonios Anastasopoulos; Graham Neubig; | |
554 | Aligning Actions Across Recipe Graphs Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We present a novel and fully-parsed English recipe corpus, ARA (Aligned Recipe Actions), which annotates correspondences between individual actions across similar recipes with the goal of capturing information implicit for accurate recipe understanding. |
Lucia Donatelli; Theresa Schmidt; Debanjali Biswas; Arne K?hn; Fangzhou Zhai; Alexander Koller; | |
555 | Generating Datasets with Pretrained Language Models Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we show how PLMs can be leveraged to obtain high-quality sentence embeddings without the need for labeled data, finetuning or modifications to the pretraining objective: We utilize the generative abilities of large and high-performing PLMs to generate entire datasets of labeled text pairs from scratch, which we then use for finetuning much smaller and more efficient models. |
Timo Schick; Hinrich Sch?tze; | |
556 | Continuous Entailment Patterns for Lexical Inference in Context Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: Contrasting patterns where a token can be any continuous vector from those where a discrete choice between vocabulary elements has to be made, we call our method CONtinous pAtterNs (CONAN). |
Martin Schmitt; Hinrich Sch?tze; | |
557 | Numeracy Enhances The Literacy of Language Models Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: This paper studies the effect of using six different number encoders on the task of masked word prediction (MWP), as a proxy for evaluating literacy. |
Avijit Thawani; Jay Pujara; Filip Ilievski; | code |
558 | Students Who Study Together Learn Better: On The Importance of Collective Knowledge Distillation for Domain Transfer in Fact Verification Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We propose Group Learning, a knowledge and model distillation approach for fact verification in which multiple student models have access to different delexicalized views of the data, but are encouraged to learn from each other through pair-wise consistency losses. |
Mitch Paul Mithun; Sandeep Suntwal; Mihai Surdeanu; | |
559 | MultiEURLEX – A Multi-lingual and Multi-label Legal Document Classification Dataset for Zero-shot Cross-lingual Transfer Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We introduce MULTI-EURLEX, a new multilingual dataset for topic classification of legal documents. |
Ilias Chalkidis; Manos Fergadiotis; Ion Androutsopoulos; | |
560 | Joint Passage Ranking for Diverse Multi-Answer Retrieval Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we introduce JPR, a joint passage retrieval model focusing on reranking. |
Sewon Min; Kenton Lee; Ming-Wei Chang; Kristina Toutanova; Hannaneh Hajishirzi; | |
561 | Generative Context Pair Selection for Multi-hop Question Answering Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We propose a generative context selection model for multi-hop QA that reasons about how the given question could have been generated given a context pair and not just independent contexts. |
Dheeru Dua; Cicero Nogueira dos Santos; Patrick Ng; Ben Athiwaratkun; Bing Xiang; Matt Gardner; Sameer Singh; | |
562 | Synthetic Data Augmentation for Zero-Shot Cross-Lingual Question Answering Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We propose a method to improve the Cross-lingual Question Answering performance without requiring additional annotated data, leveraging Question Generation models to produce synthetic samples in a cross-lingual fashion. |
Arij Riabi; Thomas Scialom; Rachel Keraron; Beno?t Sagot; Djam? Seddah; Jacopo Staiano; | |
563 | Have You Seen That Number? Investigating Extrapolation in Question Answering Models Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: Our work rigorously tests state-of-the-art models on DROP, a numerical MRC dataset, to see if they can handle passages that contain out-of-range numbers. |
Jeonghwan Kim; Giwon Hong; Kyung-min Kim; Junmo Kang; Sung-Hyon Myaeng; | |
564 | Surface Form Competition: Why The Highest Probability Answer Isn’t Always Right Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We introduce Domain Conditional Pointwise Mutual Information, an alternative scoring function that directly compensates for surface form competition by simply reweighing each option according to its a priori likelihood within the context of a specific task. |
Ari Holtzman; Peter West; Vered Shwartz; Yejin Choi; Luke Zettlemoyer; | |
565 | Entity-Based Knowledge Conflicts in Question Answering Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: To understand how models use these sources together, we formalize the problem of knowledge conflicts, where the contextual information contradicts the learned information. |
Shayne Longpre; Kartik Perisetla; Anthony Chen; Nikhil Ramesh; Chris DuBois; Sameer Singh; | |
566 | Back-Training Excels Self-Training at Unsupervised Domain Adaptation of Question Generation and Passage Retrieval Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this work, we introduce back-training, an alternative to self-training for unsupervised domain adaptation (UDA). |
Devang Kulshreshtha; Robert Belfer; Iulian Vlad Serban; Siva Reddy; | |
567 | DWUG: A Large Resource of Diachronic Word Usage Graphs in Four Languages Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we describe the creation of the largest resource of graded contextualized, diachronic word meaning annotation in four different languages, based on 100,000 human semantic proximity judgments. |
Dominik Schlechtweg; Nina Tahmasebi; Simon Hengchen; Haim Dubossarsky; Barbara McGillivray; | |
568 | I Wish I Would Have Loved This One, But I Didn’t – A Multilingual Dataset for Counterfactual Detection in Product Review Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We consider the problem of counterfactual detection (CFD) in product reviews. |
James O?Neill; Polina Rozenshtein; Ryuichi Kiryo; Motoko Kubota; Danushka Bollegala; | |
569 | Does It Capture STEL? A Modular, Similarity-based Linguistic Style Evaluation Framework Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We propose the modular, fine-grained and content-controlled similarity-based STyle EvaLuation framework (STEL) to test the performance of any model that can compare two sentences on style. |
Anna Wegmann; Dong Nguyen; | |
570 | Evaluating The Morphosyntactic Well-formedness of Generated Texts Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we propose L’AMBRE – a metric to evaluate the morphosyntactic well-formedness of text using its dependency parse and morphosyntactic rules of the language. |
Adithya Pratapa; Antonios Anastasopoulos; Shruti Rijhwani; Aditi Chaudhary; David R. Mortensen; Graham Neubig; Yulia Tsvetkov; | |
571 | AM2iCo: Evaluating Word Meaning in Context Across Low-Resource Languages with Adversarial Examples Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In order to address these gaps, we present AM2iCo (Adversarial and Multilingual Meaning in Context), a wide-coverage cross-lingual and multilingual evaluation set; it aims to faithfully assess the ability of state-of-the-art (SotA) representation models to understand the identity of word meaning in cross-lingual contexts for 14 language pairs. |
Qianchu Liu; Edoardo Maria Ponti; Diana McCarthy; Ivan Vulic; Anna Korhonen; | |
572 | CrossFit: A Few-shot Learning Challenge for Cross-task Generalization in NLP Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we explore whether and how such cross-task generalization ability can be acquired, and further applied to build better few-shot learners across diverse NLP tasks. |
Qinyuan Ye; Bill Yuchen Lin; Xiang Ren; | |
573 | On The Influence of Masking Policies in Intermediate Pre-training Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we perform a large-scale empirical study to investigate the effect of various masking policies in intermediate pre-training with nine selected tasks across three categories. |
Qinyuan Ye; Belinda Z. Li; Sinong Wang; Benjamin Bolte; Hao Ma; Wen-tau Yih; Xiang Ren; Madian Khabsa; | |
574 | ValNorm Quantifies Semantics to Reveal Consistent Valence Biases Across Languages and Over Centuries Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: By extending methods that quantify human-like biases in word embeddings, we introduce ValNorm, a novel intrinsic evaluation task and method to quantify the valence dimension of affect in human-rated word sets from social psychology. |
Autumn Toney; Aylin Caliskan; | |
575 | Perturbation CheckLists for Evaluating NLG Evaluation Metrics Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: Given this situation, we propose CheckLists for better design and evaluation of automatic metrics. |
Ananya B. Sai; Tanay Dixit; Dev Yashpal Sheth; Sreyas Mohan; Mitesh M. Khapra; | code |
576 | Robust Open-Vocabulary Translation from Visual Text Representations Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: Motivated by the robustness of human language processing, we propose the use of visual text representations, which dispense with a finite set of text embeddings in favor of continuous vocabularies created by processing visually rendered text with sliding windows. |
Elizabeth Salesky; David Etter; Matt Post; | |
577 | Don’t Go Far Off: An Empirical Study on Neural Poetry Translation Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We present an empirical investigation for poetry translation along several dimensions: 1) size and style of training data (poetic vs. non-poetic), including a zero-shot setup; 2) bilingual vs. multilingual learning; and 3) language-family-specific models vs. mixed-language-family models. |
Tuhin Chakrabarty; Arkadiy Saakyan; Smaranda Muresan; | |
578 | Improving Multilingual Translation By Representation and Gradient Regularization Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: To address this issue, we propose a joint approach to regularize NMT models at both representation-level and gradient-level. |
Yilin Yang; Akiko Eriguchi; Alexandre Muzio; Prasad Tadepalli; Stefan Lee; Hany Hassan; | |
579 | Learning Kernel-Smoothed Machine Translation with Retrieved Examples Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this work, we propose to learn Kernel-Smoothed Translation with Example Retrieval (KSTER), an effective approach to adapt neural machine translation models online. |
Qingnan Jiang; Mingxuan Wang; Jun Cao; Shanbo Cheng; Shujian Huang; Lei Li; | code |
580 | Uncertainty-Aware Balancing for Multilingual and Multi-Domain Neural Machine Translation Training Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this work, we propose an approach, MultiUAT, that dynamically adjusts the training data usage based on the model’s uncertainty on a small set of trusted clean data for multi-corpus machine translation. |
Minghao Wu; Yitong Li; Meng Zhang; Liangyou Li; Gholamreza Haffari; Qun Liu; | |
581 | Universal Simultaneous Machine Translation with Mixture-of-Experts Wait-k Policy Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we propose a universal SiMT model with Mixture-of-Experts Wait-k Policy to achieve the best translation quality under arbitrary latency with only one trained model. |
Shaolei Zhang; Yang Feng; | |
582 | How Much Coffee Was Consumed During EMNLP 2019? Fermi Problems: A New Reasoning Challenge for AI Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: To help advance AI systems towards such capabilities, we propose a new reasoning challenge, namely Fermi Problems (FPs), which are questions whose answers can only be approximately estimated because their precise computation is either impractical or impossible. |
Ashwin Kalyan; Abhinav Kumar; Arjun Chandrasekaran; Ashish Sabharwal; Peter Clark; | |
583 | Will This Question Be Answered? Question Filtering Via Answer Model Distillation for Efficient Question Answering Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper we propose a novel approach towards improving the efficiency of Question Answering (QA) systems by filtering out questions that will not be answered by them. |
Siddhant Garg; Alessandro Moschitti; | |
584 | Learning with Instance Bundles for Reading Comprehension Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: Drawing on ideas from contrastive estimation, we introduce several new supervision losses that compare question-answer scores across multiple related instances. |
Dheeru Dua; Pradeep Dasigi; Sameer Singh; Matt Gardner; | |
585 | Explaining Answers with Entailment Trees Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: Our goal, in the context of open-domain textual question-answering (QA), is to explain answers by showing the line of reasoning from what is known to the answer, rather than simply showing a fragment of textual evidence (a rationale). |
Bhavana Dalvi; Peter Jansen; Oyvind Tafjord; Zhengnan Xie; Hannah Smith; Leighanna Pipatanangkura; Peter Clark; | |
586 | SituatedQA: Incorporating Extra-Linguistic Contexts Into QA Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: To study this challenge, we introduce SituatedQA, an open-retrieval QA dataset where systems must produce the correct answer to a question given the temporal or geographical context. |
Michael Zhang; Eunsol Choi; | code |
587 | ConvAbuse: Data, Analysis, and Benchmarks for Nuanced Detection in Conversational AI Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We present the first English corpus study on abusive language towards three conversational AI systems gathered ‘in the wild’: an open-domain social bot, a rule-based chatbot, and a task-based system. |
Amanda Cercas Curry; Gavin Abercrombie; Verena Rieser; | |
588 | Conversational Multi-Hop Reasoning with Neural Commonsense Knowledge and Symbolic Logic Rules Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we propose a zero-shot commonsense reasoning system for conversational agents in an attempt to achieve this. |
Forough Arabshahi; Jennifer Lee; Antoine Bosselut; Yejin Choi; Tom Mitchell; | |
589 | Towards Automatic Evaluation of Dialog Systems: A Model-Free Off-Policy Evaluation Approach Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: To bridge such a gap, we propose a new framework named ENIGMA for estimating human evaluation scores based on recent advances of off-policy evaluation in reinforcement learning. |
Haoming Jiang; Bo Dai; Mengjiao Yang; Tuo Zhao; Wei Wei; | |
590 | Continual Learning in Task-Oriented Dialogue Systems Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we propose a first-ever continual learning benchmark for task-oriented dialogue systems with 37 domains to be learned continuously in both modularized and end-to-end learning settings. |
Andrea Madotto; Zhaojiang Lin; Zhenpeng Zhou; Seungwhan Moon; Paul Crook; Bing Liu; Zhou Yu; Eunjoon Cho; Pascale Fung; Zhiguang Wang; | |
591 | Multilingual and Cross-Lingual Intent Detection from Spoken Data Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We present a systematic study on multilingual and cross-lingual intent detection (ID) from spoken data. |
Daniela Gerz; Pei-Hao Su; Razvan Kusztos; Avishek Mondal; Michal Lis; Eshan Singhal; Nikola Mrk?ic; Tsung-Hsien Wen; Ivan Vulic; | |
592 | Investigating Robustness of Dialog Models to Popular Figurative Language Constructs Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this work, we analyze the performance of existing dialog models in situations where the input dialog context exhibits use of figurative language. |
Harsh Jhamtani; Varun Gangal; Eduard Hovy; Taylor Berg-Kirkpatrick; | |
593 | Effective Sequence-to-Sequence Dialogue State Tracking Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we study this problem from the perspectives of pre-training objectives as well as the formats of context representations. |
Jeffrey Zhao; Mahdis Mahdieh; Ye Zhang; Yuan Cao; Yonghui Wu; | |
594 | MS^2: Multi-Document Summarization of Medical Studies Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In support of this goal, we release MS^2 (Multi-Document Summarization of Medical Studies), a dataset of over 470k documents and 20K summaries derived from the scientific literature. |
Jay DeYoung; Iz Beltagy; Madeleine van Zuylen; Bailey Kuehl; Lucy Wang; | |
595 | CLIPScore: A Reference-free Evaluation Metric for Image Captioning Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we report the surprising empirical finding that CLIP (Radford et al., 2021), a cross-modal model pretrained on 400M image+caption pairs from the web, can be used for robust automatic evaluation of image captioning without the need for references. |
Jack Hessel; Ari Holtzman; Maxwell Forbes; Ronan Le Bras; Yejin Choi; | |
596 | On The Challenges of Evaluating Compositional Explanations in Multi-Hop Inference: Relevance, Completeness, and Expert Ratings Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: To address this, we construct a large corpus of 126k domain-expert (science teacher) relevance ratings that augment a corpus of explanations to standardized science exam questions, discovering 80k additional relevant facts not rated as gold. |
Peter Jansen; Kelly J. Smith; Dan Moreno; Huitzilin Ortiz; | |
597 | ESTER: A Machine Reading Comprehension Dataset for Reasoning About Event Semantic Relations Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: To facilitate these tasks, we introduce **ESTER**, a comprehensive machine reading comprehension (MRC) dataset for Event Semantic Relation Reasoning. |
Rujun Han; I-Hung Hsu; Jiao Sun; Julia Baylon; Qiang Ning; Dan Roth; Nanyun Peng; | |
598 | RICA: Evaluating Robust Inference Capabilities Based on Commonsense Axioms Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In the pursuit of advancing fluid human-AI communication, we propose a new challenge, RICA: Robust Inference using Commonsense Axioms, that evaluates robust commonsense inference despite textual perturbations. |
Pei Zhou; Rahul Khanna; Seyeon Lee; Bill Yuchen Lin; Daniel Ho; Jay Pujara; Xiang Ren; | |
599 | Compression, Transduction, and Creation: A Unified Framework for Evaluating Natural Language Generation Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we propose a unifying perspective based on the nature of information change in NLG tasks, including compression (e.g., summarization), transduction (e.g., text rewriting), and creation (e.g., dialog). |
Mingkai Deng; Bowen Tan; Zhengzhong Liu; Eric Xing; Zhiting Hu; | |
600 | MATE: Multi-view Attention for Table Transformer Efficiency Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: This work presents a sparse-attention Transformer architecture for modeling documents that contain large tables. |
Julian Eisenschlos; Maharshi Gor; Thomas M?ller; William Cohen; | |
601 | Learning with Different Amounts of Annotation: From Zero to Many Labels Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: Extending a MixUp data augmentation framework, we propose a learning algorithm that can learn from training examples with different amount of annotation (with zero, one, or multiple labels). |
Shujian Zhang; Chengyue Gong; Eunsol Choi; | |
602 | When Attention Meets Fast Recurrence: Training Language Models with Reduced Compute Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this work, we present SRU++, a highly-efficient architecture that combines fast recurrence and attention for sequence modeling. |
Tao Lei; | |
603 | Universal-KD: Attention-based Output-Grounded Intermediate Layer Knowledge Distillation Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: To tackle the aforementioned problems all together, we propose Universal-KD to match intermediate layers of the teacher and the student in the output space (by adding pseudo classifiers on intermediate layers) via the attention-based layer projection. |
Yimeng Wu; Mehdi Rezagholizadeh; Abbas Ghaddar; Md Akmal Haidar; Ali Ghodsi; | |
604 | Highly Parallel Autoregressive Entity Linking with Discriminative Correction Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this work, we propose a very efficient approach that parallelizes autoregressive linking across all potential mentions and relies on a shallow and efficient decoder. |
Nicola De Cao; Wilker Aziz; Ivan Titov; | code |
605 | Word-Level Coreference Resolution Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We propose instead to consider coreference links between individual words rather than word spans and then reconstruct the word spans. |
Vladimir Dobrovolskii; | |
606 | A Secure and Efficient Federated Learning Framework for NLP Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this work, we consider the problem of designing secure and efficient federated learning (FL) frameworks for NLP. |
Chenghong Wang; Jieren Deng; Xianrui Meng; Yijue Wang; Ji Li; Sheng Lin; Shuo Han; Fei Miao; Sanguthevar Rajasekaran; Caiwen Ding; | |
607 | Controllable Semantic Parsing Via Retrieval Augmentation Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: To this end, we propose ControllAble Semantic Parser via Exemplar Retrieval (CASPER). |
Panupong Pasupat; Yuan Zhang; Kelvin Guu; | |
608 | Constrained Language Models Yield Few-Shot Semantic Parsers Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: To bridge the gap, we use language models to paraphrase inputs into a controlled sublanguage resembling English that can be automatically mapped to a target meaning representation. |
Richard Shin; Christopher Lin; Sam Thomson; Charles Chen; Subhro Roy; Emmanouil Antonios Platanios; Adam Pauls; Dan Klein; Jason Eisner; Benjamin Van Durme; | |
609 | ExplaGraphs: An Explanation Graph Generation Task for Structured Commonsense Reasoning Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this work, we present ExplaGraphs, a new generative and structured commonsense-reasoning task (and an associated dataset) of explanation graph generation for stance prediction. |
Swarnadeep Saha; Prateek Yadav; Lisa Bauer; Mohit Bansal; | |
610 | Connect-the-Dots: Bridging Semantics Between Words and Definitions Via Aligning Word Sense Inventories Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: To address the data sparsity problem and generalize the model to be independent of one predefined inventory, we propose a gloss alignment algorithm that can align definition sentences (glosses) with the same meaning from different sense inventories to collect rich lexical knowledge. |
Wenlin Yao; Xiaoman Pan; Lifeng Jin; Jianshu Chen; Dian Yu; Dong Yu; | |
611 | LM-Critic: Language Models for Unsupervised Grammatical Error Correction Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this work, we show how to leverage a pretrained language model (LM) in defining an LM-Critic, which judges a sentence to be grammatical if the LM assigns it a higher probability than its local perturbations. |
Michihiro Yasunaga; Jure Leskovec; Percy Liang; | |
612 | Language-agnostic Representation from Multilingual Sentence Encoders for Cross-lingual Similarity Estimation Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We propose a method to distill a language-agnostic meaning embedding from a multilingual sentence encoder. |
Nattapong Tiyajamorn; Tomoyuki Kajiwara; Yuki Arase; Makoto Onizuka; | |
613 | Classifying Dyads for Militarized Conflict Analysis Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: The aim of this work is to compare these two causes in terms of how they correlate with conflict between two entities. |
Niklas Stoehr; Lucas Torroba Hennigen; Samin Ahbab; Robert West; Ryan Cotterell; | |
614 | Point-of-Interest Type Prediction Using Text and Images Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: This paper presents a study on POI type prediction using multimodal information from text and images available at posting time. |
Danae S?nchez Villegas; Nikolaos Aletras; | |
615 | Come Hither or Go Away? Recognising Pre-electoral Coalition Signals in The News Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we introduce the task of political coalition signal prediction from text, that is, the task of recognizing from the news coverage leading up to an election the (un)willingness of political parties to form a government coalition. |
Ines Rehbein; Simone Paolo Ponzetto; Anna Adendorf; Oke Bahnsen; Lukas Stoetzer; Heiner Stuckenschmidt; | |
616 | #HowYouTagTweets: Learning User Hashtagging Preferences Via Personalized Topic Attention Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: To help people find the topics they want to discuss, this paper characterizes a user’s hashtagging preferences via predicting how likely they will post with a hashtag. |
Yuji Zhang; Yubo Zhang; Chunpu Xu; Jing Li; Ziyan Jiang; Baolin Peng; | |
617 | Learning Neural Templates for Recommender Dialogue System Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: To tackle these challenges, we introduce a novel framework called NTRD for recommender dialogue system that can decouple the dialogue generation from the item recommendation. |
Zujie Liang; Huang Hu; Can Xu; Jian Miao; Yingying He; Yining Chen; Xiubo Geng; Fan Liang; Daxin Jiang; | code |
618 | Proxy Indicators for The Quality of Open-domain Dialogues Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: This work investigates using a deep-learning model trained on the General Language Understanding Evaluation (GLUE) benchmark to serve as a quality indication of open-domain dialogues. |
Rostislav Nedelchev; Jens Lehmann; Ricardo Usbeck; | |
619 | Q2: Evaluating Factual Consistency in Knowledge-Grounded Dialogues Via Question Generation and Question Answering Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: Inspired by recent work on evaluating factual consistency in abstractive summarization, we propose an automatic evaluation metric for factual consistency in knowledge-grounded dialogue using automatic question generation and question answering. |
Or Honovich; Leshem Choshen; Roee Aharoni; Ella Neeman; Idan Szpektor; Omri Abend; | |
620 | Knowledge-Aware Graph-Enhanced GPT-2 for Dialogue State Tracking Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We present a novel hybrid architecture that augments GPT-2 with representations derived from Graph Attention Networks in such a way to allow causal, sequential prediction of slot values. |
Weizhe Lin; Bo-Hsiang Tseng; Bill Byrne; | |
621 | A Collaborative Multi-agent Reinforcement Learning Framework for Dialog Action Decomposition Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: To solve these problems, we model the dialog policy learning problem with a novel multi-agent framework, in which each part of the action is led by a different agent. |
Huimin Wang; Kam-Fai Wong; | |
622 | Zero-Shot Dialogue State Tracking Via Cross-Task Transfer Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this work, we propose to transfer the cross-task knowledge from general question answering (QA) corpora for the zero-shot DST task. |
Zhaojiang Lin; Bing Liu; Andrea Madotto; Seungwhan Moon; Zhenpeng Zhou; Paul Crook; Zhiguang Wang; Zhou Yu; Eunjoon Cho; Rajen Subba; Pascale Fung; | |
623 | Uncertainty Measures in Neural Belief Tracking and The Effects on Dialogue Policy Performance Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We propose the use of different uncertainty measures in neural belief tracking. |
Carel van Niekerk; Andrey Malinin; Christian Geishauser; Michael Heck; Hsien-chin Lin; Nurul Lubis; Shutong Feng; Milica Gasic; | |
624 | Dynamic Forecasting of Conversation Derailment Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: Automatically forecasting derailment in public online conversations provides an opportunity to take early action to moderate it. Previous work in this space is limited, and we extend it in several ways. |
Yova Kementchedjhieva; Anders S?gaard; | |
625 | A Semantic Filter Based on Relations for Knowledge Graph Completion Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: Inspired by this fact, this paper designs Semantic Filter Based on Relations(SFBR) to extract the required attributes of the entities. Then the rationality of triples is compared under these extracted attributes through the traditional embedding models. |
Zongwei Liang; Junan Yang; Hui Liu; Keju Huang; | |
626 | AdapterDrop: On The Efficiency of Adapters in Transformers Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we propose AdapterDrop, removing adapters from lower transformer layers during training and inference, which incorporates concepts from all three directions. |
Andreas R?ckl?; Gregor Geigle; Max Glockner; Tilman Beck; Jonas Pfeiffer; Nils Reimers; Iryna Gurevych; | |
627 | Understanding and Overcoming The Challenges of Efficient Transformer Quantization Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this work, we explore quantization for transformers. |
Yelysei Bondarenko; Markus Nagel; Tijmen Blankevoort; | code |
628 | CAPE: Context-Aware Private Embeddings for Private Language Learning Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: To ameliorate this issue, we propose Context-Aware Private Embeddings (CAPE), a novel approach which combines differential privacy and adversarial learning to preserve privacy during training of embeddings. |
Richard Plant; Dimitra Gkatzia; Valerio Giuffrida; | |
629 | Text Detoxification Using Large Pre-trained Neural Models Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We present two novel unsupervised methods for eliminating toxicity in text. |
David Dale; Anton Voronov; Daryna Dementieva; Varvara Logacheva; Olga Kozlova; Nikita Semenov; Alexander Panchenko; | |
630 | Document-Level Text Simplification: Dataset, Criteria and Baseline Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we define and investigate a new task of document-level text simplification, which aims to simplify a document consisting of multiple sentences. |
Renliang Sun; Hanqi Jin; Xiaojun Wan; | |
631 | A Bag of Tricks for Dialogue Summarization Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this work, we explore four different challenges of the task: handling and differentiating parts of the dialogue belonging to multiple speakers, negation understanding, reasoning about the situation, and informal language understanding. |
Muhammad Khalifa; Miguel Ballesteros; Kathleen McKeown; | |
632 | Paraphrasing Compound Nominalizations Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: Our goal is to interpret nominalizations by generating clausal paraphrases. |
John Lee; Ho Hung Lim; Carol Webster; | |
633 | Data-QuestEval: A Referenceless Metric for Data-to-Text Semantic Evaluation Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: To this purpose, we propose a method to build synthetic multimodal corpora enabling to train multimodal components for a data-QuestEval metric. |
Clement Rebuffel; Thomas Scialom; Laure Soulier; Benjamin Piwowarski; Sylvain Lamprier; Jacopo Staiano; Geoffrey Scoutheeten; Patrick Gallinari; | |
634 | Low-Rank Subspaces for Unsupervised Entity Linking Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We propose a light-weight and scalable entity linking method, Eigenthemes, that relies solely on the availability of entity names and a referent knowledge base. |
Akhil Arora; Alberto Garcia-Duran; Robert West; | |
635 | TDEER: An Efficient Translating Decoding Schema for Joint Extraction of Entities and Relations Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: To address such a problem, this paper proposes a novel efficient entities and relations extraction model called TDEER, which stands for Translating Decoding Schema for Joint Extraction of Entities and Relations. |
Xianming Li; Xiaotian Luo; Chenghao Dong; Daichuan Yang; Beidi Luan; Zhen He; | code |
636 | Extracting Event Temporal Relations Via Hyperbolic Geometry Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We introduce two approaches to encode events and their temporal relations in hyperbolic spaces. |
Xingwei Tan; Gabriele Pergola; Yulan He; | |
637 | Honey or Poison? Solving The Trigger Curse in Few-shot Event Detection Via Causal Intervention Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we identify and solve the trigger curse problem in few-shot event detection (FSED) from a causal view. |
Jiawei Chen; Hongyu Lin; Xianpei Han; Le Sun; | |
638 | Back to The Basics: A Quantitative Analysis of Statistical and Graph-Based Term Weighting Schemes for Keyword Extraction Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we perform an exhaustive and large-scale empirical comparison of both statistical and graph-based term weighting methods in the context of keyword extraction. |
Asahi Ushio; Federico Liberatore; Jose Camacho-Collados; | code |
639 | Time-dependent Entity Embedding Is Not All You Need: A Re-evaluation of Temporal Knowledge Graph Completion Models Under A Unified Framework Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this work, we systematically study six temporal embedding approaches and empirically quantify their performance across a wide range of configurations with about 3000 experiments and 13159 GPU hours. |
Zhen Han; Gengyuan Zhang; Yunpu Ma; Volker Tresp; | |
640 | Matching-oriented Embedding Quantization For Ad-hoc Retrieval Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this work, we propose the Matching-oriented Product Quantization (MoPQ), where a novel objective Multinoulli Contrastive Loss (MCL) is formulated. |
Shitao Xiao; Zheng Liu; Yingxia Shao; Defu Lian; Xing Xie; | code |
641 | Efficient Mind-Map Generation Via Sequence-to-Graph and Reinforced Graph Refinement Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: To deal with the above challenges, we propose an efficient mind-map generation network that converts a document into a graph via sequence-to-graph. |
Mengting Hu; Honglei Guo; Shiwan Zhao; Hang Gao; Zhong Su; | |
642 | Deep Attention Diffusion Graph Neural Networks for Text Classification Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, a Deep Attention Diffusion Graph Neural Network (DADGNN) model is proposed to learn text representations, bridging the chasm of interaction difficulties between a word and its distant neighbors. |
Yonghao Liu; Renchu Guan; Fausto Giunchiglia; Yanchun Liang; Xiaoyue Feng; | |
643 | Balancing Methods for Multi-label Text Classification with Long-Tailed Class Distribution Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: Here, we introduce the application of balancing loss functions for multi-label text classification. |
Yi Huang; Buse Giledereli; Abdullatif K?ksal; Arzucan ?zg?r; Elif Ozkirimli; | code |
644 | Bayesian Topic Regression for Causal Inference Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: This paper presents the Bayesian Topic Regression (BTR) model that uses both text and numerical information to model an outcome variable. |
Maximilian Ahrens; Julian Ashwin; Jan-Peter Calliess; Vu Nguyen; | |
645 | Enjoy The Salience: Towards Better Transformer-based Faithful Explanations with Word Salience Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we hypothesize that salient information extracted a priori from the training data can complement the task-specific information learned by the model during fine-tuning on a downstream task. |
George Chrysostomou; Nikolaos Aletras; | |
646 | What’s in Your Head? Emergent Behaviour in Multi-Task Transformer Models Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this work, we examine the behaviour of non-target heads, that is, the output of heads when given input that belongs to a different task than the one they were trained for. |
Mor Geva; Uri Katz; Aviv Ben-Arie; Jonathan Berant; | |
647 | Don’t Search for A Search Method – Simple Heuristics Suffice for Adversarial Text Attacks Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We conclude from these results that current TextAttack benchmark tasks are too easy and constraints are too strict, preventing meaningful research on black-box adversarial text attacks. |
Nathaniel Berger; Stefan Riezler; Sebastian Ebert; Artem Sokolov; | |
648 | Adversarial Attacks on Knowledge Graph Embeddings Via Instance Attribution Methods Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We study data poisoning attacks against KGE models for link prediction. |
Peru Bhardwaj; John Kelleher; Luca Costabello; Declan O?Sullivan; | |
649 | Locke’s Holiday: Belief Bias in Machine Reading Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: I highlight a simple failure mode of state-of-the-art machine reading systems: when contexts do not align with commonly shared beliefs. |
Anders S?gaard; | |
650 | Sequence Length Is A Domain: Length-based Overfitting in Transformer Models Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We present results which suggest that the issue might also be in the mismatch between the length distributions of the training and validation data combined with the aforementioned tendency of the neural networks to overfit to the training data. |
Dusan Varis; Ondrej Bojar; | |
651 | Contrasting Human- and Machine-Generated Word-Level Adversarial Examples for Text Classification Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this work, we investigate this through the lens of human language ability. |
Maximilian Mozes; Max Bartolo; Pontus Stenetorp; Bennett Kleinberg; Lewis Griffin; | |
652 | Is Information Density Uniform in Task-Oriented Dialogues? Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we test whether, and within which contextual units this principle holds in task-oriented dialogues. |
Mario Giulianelli; Arabella Sinclair; Raquel Fern?ndez; | |
653 | On Homophony and R?nyi Entropy Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We first propose a new information-theoretic quantification of a language’s homophony: the sample R�nyi entropy. Then, we use this quantification to revisit Trott and Bergen’s claims. |
Tiago Pimentel; Clara Meister; Simone Teufel; Ryan Cotterell; | |
654 | Synthetic Textual Features for The Large-Scale Detection of Basic-level Categories in English and Mandarin Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We present the first method for the detection of BLC at scale that makes use of Rosch-style semantic features. |
Yiwen Chen; Simone Teufel; | |
655 | TimeTraveler: Reinforcement Learning for Temporal Knowledge Graph Forecasting Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: To address these challenges, we propose the first reinforcement learning method for forecasting. |
Haohai Sun; Jialun Zhong; Yunpu Ma; Zhen Han; Kun He; | |
656 | Code-switched Inspired Losses for Spoken Dialog Representations Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this work, we introduce new pretraining losses tailored to learn generic multilingual spoken dialogue representations. |
Pierre Colombo; Emile Chapuis; Matthieu Labeau; Chlo? Clavel; | |
657 | BiQUE: Biquaternionic Embeddings of Knowledge Graphs Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: To do so, we propose BiQUE, a novel model that employs biquaternions to integrate multiple geometric transformations, viz., scaling, translation, Euclidean rotation, and hyperbolic rotation. |
Jia Guo; Stanley Kok; | |
658 | Learning Neural Ordinary Equations for Forecasting Future Links on Temporal Knowledge Graphs Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: To this end, we propose a novel continuum model by extending the idea of neural ordinary differential equations (ODEs) to multi-relational graph convolutional networks. |
Zhen Han; Zifeng Ding; Yunpu Ma; Yujia Gu; Volker Tresp; | |
659 | RAP: Robustness-Aware Perturbations for Defending Against Backdoor Attacks on NLP Models Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this work, we propose an efficient online defense mechanism based on robustness-aware perturbations. |
Wenkai Yang; Yankai Lin; Peng Li; Jie Zhou; Xu Sun; | code |
660 | FAME: Feature-Based Adversarial Meta-Embeddings for Robust Input Representations Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: As an alternative to attention-based meta-embeddings, we propose feature-based adversarial meta-embeddings (FAME) with an attention function that is guided by features reflecting word-specific properties, such as shape and frequency, and show that this is beneficial to handle subword-based embeddings. |
Lukas Lange; Heike Adel; Jannik Str?tgen; Dietrich Klakow; | |
661 | A Strong Baseline for Query Efficient Attacks in A Black Box Setting Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we propose a query efficient attack strategy to generate plausible adversarial examples on text classification and entailment tasks. |
Rishabh Maheshwary; Saket Maheshwary; Vikram Pudi; | |
662 | Machine Translation Decoding Beyond Beam Search Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: Our aim is to establish whether beam search can be replaced by a more powerful metric-driven search technique. |
R?mi Leblond; Jean-Baptiste Alayrac; Laurent Sifre; Miruna Pislar; Lespiau Jean-Baptiste; Ioannis Antonoglou; Karen Simonyan; Oriol Vinyals; | |
663 | Document Graph for Neural Machine Translation Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: To address this issue, we hypothesize that a document can be represented as a graph that connects relevant contexts regardless of their distances. |
Mingzhou Xu; Liangyou Li; Derek F. Wong; Qun Liu; Lidia S. Chao; | |
664 | An Empirical Investigation of Word Alignment Supervision for Zero-Shot Multilingual Neural Machine Translation Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we investigate the benefits of an explicit alignment to language labels in Transformer-based MNMT models in the zero-shot context, by jointly training one cross attention head with word alignment supervision to stress the focus on the target language label. |
Alessandro Raganato; Ra?l V?zquez; Mathias Creutz; J?rg Tiedemann; | |
665 | Graph Algorithms for Multiparallel Word Alignment Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We present two graph algorithms for edge prediction: one inspired by recommender systems and one based on network link prediction. |
Ayyoob ImaniGooghari; Masoud Jalili Sabet; Lutfi Kerem Senel; Philipp Dufter; Fran?ois Yvon; Hinrich Sch?tze; | |
666 | Improving The Quality Trade-Off for Neural Machine Translation Multi-Domain Adaptation Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We study this problem in an adaptation setting where the goal is to preserve the existing system quality while incorporating data for domains that were not the focus of the original translation system. |
Eva Hasler; Tobias Domhan; Jonay Trenous; Ke Tran; Bill Byrne; Felix Hieber; | |
667 | Language Modeling, Lexical Translation, Reordering: The Training Process of NMT Through The Lens of Classical SMT Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this work, we look at the competences related to three core SMT components and find that during training, NMT first focuses on learning target-side language modeling, then improves translation quality approaching word-by-word translation, and finally learns more complicated reordering patterns. |
Elena Voita; Rico Sennrich; Ivan Titov; | |
668 | Effective Fine-Tuning Methods for Cross-lingual Adaptation Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this work, we propose a novel fine-tuning method based on co-training that aims to learn more generalized semantic equivalences as a complementary to multilingual language modeling using the unlabeled data in the target language. |
Tao Yu; Shafiq Joty; | |
669 | Rethinking Data Augmentation for Low-Resource Neural Machine Translation: A Multi-Task Learning Approach Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we propose to follow a completely different approach and present a multi-task DA approach in which we generate new sentence pairs with transformations, such as reversing the order of the target sentence, which produce unfluent target sentences. |
V?ctor M. S?nchez-Cartagena; Miquel Espl?-Gomis; Juan Antonio P?rez-Ortiz; Felipe S?nchez-Mart?nez; | |
670 | Wino-X: Multilingual Winograd Schemas for Commonsense Reasoning and Coreference Resolution Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: This work presents Wino-X, a parallel dataset of German, French, and Russian schemas, aligned with their English counterparts. |
Denis Emelin; Rico Sennrich; | |
671 | One Source, Two Targets: Challenges and Rewards of Dual Decoding Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we consider a stronger requirement: to jointly generate two texts so that each output side effectively depends on the other. |
Jitao Xu; Fran?ois Yvon; | |
672 | Discrete and Soft Prompting for Multilingual Models Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we show that discrete and soft prompting perform better than finetuning in multilingual cases: Crosslingual transfer and in-language training of multilingual natural language inference. |
Mengjie Zhao; Hinrich Sch?tze; | |
673 | Vision Matters When It Should: Sanity Checking Multimodal Machine Translation Models Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we present a qualitative study that examines the role of datasets in stimulating the leverage of visual modality and we propose methods to highlight the importance of visual signals in the datasets which demonstrate improvements in reliance of models on the source images. |
Jiaoda Li; Duygu Ataman; Rico Sennrich; | |
674 | Efficient Inference for Multilingual Neural Machine Translation Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this work, we consider several ways to make multilingual NMT faster at inference without degrading its quality. |
Alexandre Berard; Dain Lee; Stephane Clinchant; Kweonwoo Jung; Vassilina Nikoulina; | |
675 | Role of Language Relatedness in Multilingual Fine-tuning of Language Models: A Case Study in Indo-Aryan Languages Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We explore the impact of leveraging the relatedness of languages that belong to the same family in NLP models using multilingual fine-tuning. |
Tejas Dhamecha; Rudra Murthy; Samarth Bharadwaj; Karthik Sankaranarayanan; Pushpak Bhattacharyya; | |
676 | Comparing Feature-Engineering and Feature-Learning Approaches for Multilingual Translationese Classification Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this work, we (i) compare the traditional feature-engineering-based approach to the feature-learning-based one and (ii) analyse the neural architectures in order to investigate how well the hand-crafted features explain the variance in the neural models’ predictions. |
Daria Pylypenko; Kwabena Amponsah-Kaakyire; Koel Dutta Chowdhury; Josef van Genabith; Cristina Espa?a-Bonet; | |
677 | Multi-Sentence Resampling: A Simple Approach to Alleviate Dataset Length Bias and Beam-Search Degradation Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this work, we analyze errors that cause major quality degradation with large beams in NMT and Automatic Speech Recognition (ASR). |
Ivan Provilkov; Andrey Malinin; | |
678 | Cross-Policy Compliance Detection Via Question Answering Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper we propose to address policy compliance detection via decomposing it into question answering, where questions check whether the conditions stated in the policy apply to the scenario, and an expression tree combines the answers to obtain the label. |
Marzieh Saeidi; Majid Yazdani; Andreas Vlachos; | |
679 | Meta-LMTC: Meta-Learning for Large-Scale Multi-Label Text Classification Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, for the first time, this problem is addressed from a meta-learning perspective. |
Ran Wang; Xi?ao Su; Siyu Long; Xinyu Dai; Shujian Huang; Jiajun Chen; | |
680 | Unsupervised Multi-View Post-OCR Error Correction With Language Models Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: The goal of this study is to understand if a pretrained language model (LM) can be used in an unsupervised way to reconcile the different OCR views such that their combination contains fewer errors than each individual view. |
Harsh Gupta; Luciano Del Corro; Samuel Broscheit; Johannes Hoffart; Eliot Brenner; | |
681 | Parallel Refinements for Lexically Constrained Text Generation with BART Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: To address these challenges, we proposed Constrained BART (CBART) for lexically constrained text generation. |
Xingwei He; | |
682 | BERT-Beta: A Proactive Probabilistic Approach to Text Moderation Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this work, we explore an alternative perspective by augmenting reactive reviews with proactive forecasting. |
Fei Tan; Yifan Hu; Kevin Yen; Changwei Hu; | |
683 | STaCK: Sentence Ordering with Temporal Commonsense Knowledge Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we introduce STaCK – a framework based on graph neural networks and temporal commonsense knowledge to model global information and predict the relative order of sentences. |
Deepanway Ghosal; Navonil Majumder; Rada Mihalcea; Soujanya Poria; | code |
684 | Preventing Author Profiling Through Zero-Shot Multilingual Back-Translation Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we propose a simple, zero-shot way to effectively lower the risk of author profiling through multilingual back-translation using off-the-shelf translation models. |
David Adelani; Miaoran Zhang; Xiaoyu Shen; Ali Davody; Thomas Kleinbauer; Dietrich Klakow; | |
685 | CodeT5: Identifier-aware Unified Pre-trained Encoder-Decoder Models for Code Understanding and Generation Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We present CodeT5, a unified pre-trained encoder-decoder Transformer model that better leverages the code semantics conveyed from the developer-assigned identifiers. |
Yue Wang; Weishi Wang; Shafiq Joty; Steven C.H. Hoi; | code |
686 | Detect and Classify – Joint Span Detection and Classification for Health Outcomes Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: To address this, we propose a method that uses both word-level and sentence-level information to simultaneously perform outcome span detection and outcome type classification. |
Micheal Abaho; Danushka Bollegala; Paula Williamson; Susanna Dodd; | |
687 | Multi-Class Grammatical Error Detection for Correction: A Tale of Two Systems Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we show how a multi-class grammatical error detection (GED) system can be used to improve grammatical error correction (GEC) for English. |
Zheng Yuan; Shiva Taslimipoor; Christopher Davis; Christopher Bryant; | |
688 | Towards Zero-shot Commonsense Reasoning with Self-supervised Refinement of Language Models Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: To this end, we propose a novel self-supervised learning approach that refines the language model utilizing a set of linguistic perturbations of similar concept relationships. |
Tassilo Klein; Moin Nabi; | |
689 | To Share or Not to Share: Predicting Sets of Sources for Model Transfer Learning Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: To tackle this problem, we propose a new approach to automatically determine which and how many sources should be exploited. |
Lukas Lange; Jannik Str?tgen; Heike Adel; Dietrich Klakow; | |
690 | Self-Supervised Detection of Contextual Synonyms in A Multi-Class Setting: Phenotype Annotation Use Case Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we propose a self-supervised pre-training approach which is able to detect contextual synonyms of concepts being training on the data created by shallow matching. |
Jingqing Zhang; Luis Bolanos Trujillo; Tong Li; Ashwani Tanwar; Guilherme Freire; Xian Yang; Julia Ive; Vibhor Gupta; Yike Guo; | |
691 | ClauseRec: A Clause Recommendation Framework for AI-aided Contract Authoring Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we introduce the task of clause recommendation, as a first step to aid and accelerate the authoring of contract documents. |
Vinay Aggarwal; Aparna Garimella; Balaji Vasan Srinivasan; Anandhavelu N; Rajiv Jain; | |
692 | Finnish Dialect Identification: The Effect of Audio and Text Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We present the first approach to automatically detect the dialect of a speaker based on a dialect transcript and transcript with audio recording in a dataset consisting of 23 different dialects. |
Mika H?m?l?inen; Khalid Alnajjar; Niko Partanen; Jack Rueter; | |
693 | English Machine Reading Comprehension Datasets: A Survey Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: Our analysis reveals that Wikipedia is by far the most common data source and that there is a relative lack of why, when, and where questions across datasets. |
Daria Dzendzik; Jennifer Foster; Carl Vogel; | |
694 | Expanding End-to-End Question Answering on Differentiable Knowledge Graphs with Intersection Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we propose a model that explicitly handles multiple-entity questions by implementing a new intersection operation, which identifies the shared elements between two sets of entities. |
Priyanka Sen; Armin Oliya; Amir Saffari; | |
695 | Structured Context and High-Coverage Grammar for Conversational Question Answering Over Knowledge Graphs Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We introduce a new Logical Form (LF) grammar that can model a wide range of queries on the graph while remaining sufficiently simple to generate supervision data efficiently. |
Pierre Marion; Pawel Nowak; Francesco Piccinno; | |
696 | Improving Question Answering Model Robustness with Synthetic Adversarial Data Generation Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this work, we are the first to use synthetic adversarial data generation to make question answering models more robust to human adversaries. |
Max Bartolo; Tristan Thrush; Robin Jia; Sebastian Riedel; Pontus Stenetorp; Douwe Kiela; | |
697 | BeliefBank: Adding Memory to A Pre-Trained Language Model for A Systematic Notion of Belief Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: Our approach is to embed a PTLM in a broader system that also includes an evolving, symbolic memory of beliefs – a BeliefBank – that records but then may modify the raw PTLM answers. |
Nora Kassner; Oyvind Tafjord; Hinrich Sch?tze; Peter Clark; | |
698 | MLEC-QA: A Chinese Multi-Choice Biomedical Question Answering Dataset Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we present MLEC-QA, the largest-scale Chinese multi-choice biomedical QA dataset, collected from the National Medical Licensing Examination in China. |
Jing Li; Shangping Zhong; Kaizhi Chen; | |
699 | IndoNLG: Benchmark and Resources for Evaluating Indonesian Natural Language Generation Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: Here we introduce IndoNLG, the first benchmark to measure natural language generation (NLG) progress in three low-resource -yet widely spoken- languages of Indonesia: Indonesian, Javanese, and Sundanese. |
Samuel Cahyawijaya; Genta Indra Winata; Bryan Wilie; Karissa Vincentio; Xiaohong Li; Adhiguna Kuncoro; Sebastian Ruder; Zhi Yuan Lim; Syafri Bahar; Masayu Khodra; Ayu Purwarianti; Pascale Fung; | |
700 | Is Multi-Hop Reasoning Really Explainable? Towards Benchmarking Reasoning Interpretability Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we propose a unified framework to quantitatively evaluate the interpretability of multi-hop reasoning models so as to advance their development. |
Xin Lv; Yixin Cao; Lei Hou; Juanzi Li; Zhiyuan Liu; Yichi Zhang; Zelin Dai; | |
701 | Global Explainability of BERT-Based Evaluation Metrics By Disentangling Along Linguistic Factors Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this work, we use a simple regression based global explainability technique to disentangle metric scores along linguistic factors, including semantics, syntax, morphology, and lexical overlap. |
Marvin Kaster; Wei Zhao; Steffen Eger; | |
702 | Exploring Underexplored Limitations of Cross-Domain Text-to-SQL Generalization Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this work, we investigate the robustness of text-to-SQL models when the questions require rarely observed domain knowledge. |
Yujian Gan; Xinyun Chen; Matthew Purver; | |
703 | What Happens If You Treat Ordinal Ratings As Interval Data? Human Evaluations in NLP Are Even More Under-powered Than You Think Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: Here, we argue that there are two common factors which make this problem even worse: NLP studies usually (a) treat ordinal data as interval data and (b) operate under high variance settings while the differences they are hoping to detect are often subtle. |
David M. Howcroft; Verena Rieser; | |
704 | NeuTral Rewriter: A Rule-Based and Neural Approach to Automatic Rewriting Into Gender Neutral Alternatives Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this work, we present a rule-based and a neural approach to gender-neutral rewriting for English along with manually curated synthetic data (WinoBias+) and natural data (OpenSubtitles and Reddit) benchmarks. |
Eva Vanmassenhove; Chris Emmery; Dimitar Shterionov; | |
705 | Benchmarking Commonsense Knowledge Base Population with An Effective Evaluation Dataset Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we benchmark the CSKB population task with a new large-scale dataset by first aligning four popular CSKBs, and then presenting a high-quality human-annotated evaluation set to probe neural models’ commonsense reasoning ability. |
Tianqing Fang; Weiqi Wang; Sehyun Choi; Shibo Hao; Hongming Zhang; Yangqiu Song; Bin He; | code |
706 | Enhancing The Context Representation in Similarity-based Word Sense Disambiguation Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we investigate the contribution of both word-level and sense-level global context of an ambiguous word for disambiguation. |
Ming Wang; Jianzhang Zhang; Yinglin Wang; | |
707 | Data Augmentation with Hierarchical SQL-to-Question Generation for Cross-domain Text-to-SQL Parsing Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: This paper presents a simple yet effective data augmentation framework. |
Kun Wu; Lijie Wang; Zhenghua Li; Ao Zhang; Xinyan Xiao; Hua Wu; Min Zhang; Haifeng Wang; | |
708 | SPARQLing Database Queries from Intermediate Question Decompositions Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: Annotating a large dataset with queries is difficult as it requires query-language expertise. We reduce this burden using grounded in databases intermediate question representations. |
Irina Saparina; Anton Osokin; | code |
709 | Time-aware Graph Neural Network for Entity Alignment Between Temporal Knowledge Graphs Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we focus on the task of aligning entity pairs between TKGs and propose a novel Time-aware Entity Alignment approach based on Graph Neural Networks (TEA-GNN). |
Chengjin Xu; Fenglong Su; Jens Lehmann; | |
710 | Cross-Domain Label-Adaptive Stance Detection Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we perform an in-depth analysis of 16 stance detection datasets, and we explore the possibility for cross-domain learning from them. |
Momchil Hardalov; Arnav Arora; Preslav Nakov; Isabelle Augenstein; | |
711 | Text AutoAugment: Learning Compositional Augmentation Policy for Text Classification Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: To overcome the above limitations, we propose a framework named Text AutoAugment (TAA) to establish a compositional and learnable paradigm for data augmentation. |
Shuhuai Ren; Jinchao Zhang; Lei Li; Xu Sun; Jie Zhou; | |
712 | Distilling Relation Embeddings from Pretrained Language Models Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: To obtain relation embeddings from a pre-trained language model, we encode word pairs using a (manually or automatically generated) prompt, and we fine-tune the language model such that relationally similar word pairs yield similar output vectors. |
Asahi Ushio; Jose Camacho-Collados; Steven Schockaert; | code |
713 | Avoiding Inference Heuristics in Few-shot Prompt-based Finetuning Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this work, we demonstrate that, despite its advantages on low data regimes, finetuned prompt-based models for sentence pair classification tasks still suffer from a common pitfall of adopting inference heuristics based on lexical overlap, e.g., models incorrectly assuming a sentence pair is of the same meaning because they consist of the same set of words. |
Prasetya Utama; Nafise Sadat Moosavi; Victor Sanh; Iryna Gurevych; | |
714 | A Differentiable Relaxation of Graph Segmentation and Alignment for AMR Parsing Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: To ensure end-to-end differentiable optimization, we introduce a differentiable relaxation of the segmentation and alignment problems. |
Chunchuan Lyu; Shay B. Cohen; Ivan Titov; | |
715 | Integrating Personalized PageRank Into Neural Word Sense Disambiguation Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this work, we improve a classification model by recomputing logits as a function of both the vanilla independently produced logits and the global WordNet graph. |
Ahmed El Sheikh; Michele Bevilacqua; Roberto Navigli; | code |
716 | Cross-lingual Sentence Embedding Using Multi-Task Learning Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we propose a novel sentence embedding framework based on an unsupervised loss function for generating effective multilingual sentence embeddings, eliminating the need for parallel corpora. |
Koustava Goswami; Sourav Dutta; Haytham Assem; Theodorus Fransen; John P. McCrae; | |
717 | NB-MLM: Efficient Domain Adaptation of Masked Language Models for Sentiment Analysis Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We propose a technique for more efficient adaptation that focuses on predicting words with large weights of the Naive Bayes classifier trained for the task at hand, which are likely more relevant than the most frequent words. |
Nikolay Arefyev; Dmitrii Kharchev; Artem Shelmanov; | |
718 | Revisiting Self-training for Few-shot Learning of Language Model Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this work, we revisit the self-training technique for language model fine-tuning and present a state-of-the-art prompt-based few-shot learner, SFLM. |
Yiming Chen; Yan Zhang; Chen Zhang; Grandee Lee; Ran Cheng; Haizhou Li; | |
719 | Bridging Perception, Memory, and Inference Through Semantic Relations Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this work, we probe six popular language models for semantic relations and outline a future line of research to study how the constituent subsystems can be jointly realised and integrated. |
Johanna Bj?rklund; Adam Dahlgren Lindstr?m; Frank Drewes; | |
720 | Unimodal and Crossmodal Refinement Network for Multimodal Sequence Fusion Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, Unimodal and Crossmodal Refinement Network (UCRN) is proposed to enhance both unimodal and crossmodal representations. |
Xiaobao Guo; Adams Kong; Huan Zhou; Xianfeng Wang; Min Wang; | |
721 | YASO: A Targeted Sentiment Analysis Evaluation Dataset for Open-Domain Reviews Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: To address this gap, we present YASO – a new TSA evaluation dataset of open-domain user reviews. |
Matan Orbach; Orith Toledo-Ronen; Artem Spector; Ranit Aharonov; Yoav Katz; Noam Slonim; | code |
722 | An Empirical Study on Leveraging Position Embeddings for Target-oriented Opinion Words Extraction Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we explore a variety of text encoders based on pretrained word embeddings or language models that leverage part-of-speech and position embeddings, aiming to examine the actual contribution of each component in TOWE. |
Samuel Mensah; Kai Sun; Nikolaos Aletras; | |
723 | Improving Multimodal Fusion with Hierarchical Mutual Information Maximization for Multimodal Sentiment Analysis Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this work, we propose a framework named MultiModal InfoMax (MMIM), which hierarchically maximizes the Mutual Information (MI) in unimodal input pairs (inter-modality) and between multimodal fusion result and unimodal input in order to maintain task-related information through multimodal fusion. |
Wei Han; Hui Chen; Soujanya Poria; | |
724 | BERT4GCN: Using BERT Intermediate Layers to Augment GCN for Aspect-based Sentiment Classification Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we propose a novel model, BERT4GCN, which integrates the grammatical sequential features from the PLM of BERT, and the syntactic knowledge from dependency graphs. |
Zeguan Xiao; Jiarun Wu; Qingliang Chen; Congjian Deng; | |
725 | Does Social Pressure Drive Persuasion in Online Fora? Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: While previous research has focused on arguments between a view-holder and a persuader, we explore the premise that apart from the merits of arguments, persuasion is influenced by the ambient social community. |
Ayush Jain; Shashank Srivastava; | |
726 | Aspect Sentiment Quad Prediction As Paraphrase Generation Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this work, we introduce the Aspect Sentiment Quad Prediction (ASQP) task, aiming to jointly detect all sentiment elements in quads for a given opinionated sentence, which can reveal a more comprehensive and complete aspect-level sentiment structure. |
Wenxuan Zhang; Yang Deng; Xin Li; Yifei Yuan; Lidong Bing; Wai Lam; | |
727 | Cross-lingual Aspect-based Sentiment Analysis with Aspect Term Code-Switching Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we consider the unsupervised cross-lingual transfer for the ABSA task, where only labeled data in the source language is available and we aim at transferring its knowledge to the target language having no labeled data. |
Wenxuan Zhang; Ruidan He; Haiyun Peng; Lidong Bing; Wai Lam; | |
728 | Towards Label-Agnostic Emotion Embeddings Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: To resolve this unsatisfactory state of affairs we here propose a training scheme that learns a shared latent representation of emotion independent from different label formats, natural languages, and even disparate model architectures. |
Sven Buechel; Luise Modersohn; Udo Hahn; | |
729 | Collaborative Learning of Bidirectional Decoders for Unsupervised Text Style Transfer Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we propose a collaborative learning framework for unsupervised text style transfer using a pair of bidirectional decoders, one decoding from left to right while the other decoding from right to left. |
Yun Ma; Yangbin Chen; Xudong Mao; Qing Li; | |
730 | Exploring Non-Autoregressive Text Style Transfer Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we explore Non-AutoRegressive (NAR) decoding for unsupervised text style transfer. |
Yun Ma; Qing Li; | |
731 | PASTE: A Tagging-Free Decoding Framework Using Pointer Networks for Aspect Sentiment Triplet Extraction Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: Different from these, we present a tagging-free solution for the task, while addressing the limitations of the existing works. |
Rajdeep Mukherjee; Tapas Nayak; Yash Butala; Sourangshu Bhattacharya; Pawan Goyal; | |
732 | Adaptive Proposal Generation Network for Temporal Sentence Localization in Videos Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we propose an Adaptive Proposal Generation Network (APGN) to maintain the segment-level interaction while speeding up the efficiency. |
Daizong Liu; Xiaoye Qu; Jianfeng Dong; Pan Zhou; | |
733 | Progressively Guide to Attend: An Iterative Alignment Framework for Temporal Sentence Grounding Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we propose an Iterative Alignment Network (IA-Net) for TSG task, which iteratively interacts inter- and intra-modal features within multiple steps for more accurate grounding. |
Daizong Liu; Xiaoye Qu; Pan Zhou; | |
734 | Language Models Are Few-Shot Butlers Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We introduce a two-stage procedure to learn from a small set of demonstrations and further improve by interacting with an environment. |
Vincent Micheli; Francois Fleuret; | |
735 | R^3Net:Relation-embedded Representation Reconstruction Network for Change Captioning Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we propose a Relation-embedded Representation Reconstruction Network (R^3Net) to explicitly distinguish the real change from the large amount of clutter and irrelevant changes. |
Yunbin Tu; Liang Li; Chenggang Yan; Shengxiang Gao; Zhengtao Yu; | |
736 | Looking for Confirmations: An Effective and Human-Like Visual Dialogue Strategy Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: Inspired by the cognitive literature on information search and cross-situational word learning, we design Confirm-it, a model based on a beam search re-ranking algorithm that guides an effective goal-oriented strategy by asking questions that confirm the model’s conjecture about the referent. |
Alberto Testoni; Raffaella Bernardi; | |
737 | A Unified Speaker Adaptation Approach for ASR Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this work, we propose a unified speaker adaptation approach consisting of feature adaptation and model adaptation. |
Yingzhu Zhao; Chongjia Ni; Cheung-Chi Leung; Shafiq Joty; Eng Siong Chng; Bin Ma; | |
738 | Caption Enriched Samples for Improving Hateful Memes Detection Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: Motivated by the need to model the contrast between the image content and the overlayed text, we suggest applying an off-the-shelf image captioning tool in order to capture the first. |
Efrat Blaier; Itzik Malkiel; Lior Wolf; | |
739 | Sparsity and Sentence Structure in Encoder-Decoder Attention of Summarization Systems Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: Modified encoder architectures such as LED or LoBART use local attention patterns to address this problem for summarization. In contrast, this work focuses on the transformer’s encoder-decoder attention mechanism. |
Potsawee Manakul; Mark Gales; | |
740 | BARThez: A Skilled Pretrained French Sequence-to-Sequence Model Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this work, we introduce BARThez, the first large-scale pretrained seq2seq model for French. |
Moussa Kamal Eddine; Antoine Tixier; Michalis Vazirgiannis; | |
741 | ARMAN: Pre-training with Semantically Selecting and Reordering of Sentences for Persian Abstractive Summarization Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We propose ARMAN, a Transformer-based encoder-decoder model pre-trained with three novel objectives to address this issue. |
Alireza Salemi; Emad Kebriaei; Ghazal Neisi Minaei; Azadeh Shakery; | |
742 | Models and Datasets for Cross-Lingual Summarisation Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We present a cross-lingual summarisation corpus with long documents in a source language associated with multi-sentence summaries in a target language. |
Laura Perez-Beltrachini; Mirella Lapata; | |
743 | Learning Opinion Summarizers By Selecting Informative Reviews Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this work, we collect a large dataset of summaries paired with user reviews for over 31,000 products, enabling supervised training. |
Arthur Bra?inskas; Mirella Lapata; Ivan Titov; | |
744 | Enriching and Controlling Global Semantics for Text Summarization Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we attempt to address this issue by introducing a neural topic model empowered with normalizing flow to capture the global semantics of the document, which are then integrated into the summarization model. |
Thong Nguyen; Anh Tuan Luu; Truc Lu; Tho Quan; | |
745 | Revisiting Tri-training of Dependency Parsers Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We compare two orthogonal semi-supervised learning techniques, namely tri-training and pretrained word embeddings, in the task of dependency parsing. |
Joachim Wagner; Jennifer Foster; | |
746 | Bridge to Target Domain By Prototypical Contrastive Learning and Label Confusion: Re-explore Zero-Shot Learning for Slot Filling Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: To solve this, we propose a novel approach based on prototypical contrastive learning with a dynamic label confusion strategy for zero-shot slot filling. |
Liwen Wang; Xuefeng Li; Jiachi Liu; Keqing He; Yuanmeng Yan; Weiran Xu; | |
747 | Neuralizing Regular Expressions for Slot Filling Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we study the integration of the two approaches for the slot filling task by converting regular expressions into neural networks. |
Chengyue Jiang; Zijian Jin; Kewei Tu; | |
748 | Causal Direction of Data Collection Matters: Implications of Causal and Anticausal Learning for NLP Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this work, we argue that the causal direction of the data collection process bears nontrivial implications that can explain a number of published NLP findings, such as differences in semi-supervised learning (SSL) and domain adaptation (DA) performance across different settings. |
Zhijing Jin; Julius von K?gelgen; Jingwei Ni; Tejas Vaidhya; Ayush Kaushal; Mrinmaya Sachan; Bernhard Schoelkopf; | |
749 | Raise A Child in Large Language Model: Towards Effective and Generalizable Fine-tuning Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we propose a straightforward yet effective fine-tuning technique, Child-Tuning, which updates a subset of parameters (called child network) of large pretrained models via strategically masking out the gradients of the non-child network during the backward process. |
Runxin Xu; Fuli Luo; Zhiyuan Zhang; Chuanqi Tan; Baobao Chang; Songfang Huang; Fei Huang; | |
750 | Knowledge Graph Representation Learning Using Ordinary Differential Equations Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: To address this problem, we propose a neuro differential KGE that embeds nodes of a KG on the trajectories of Ordinary Differential Equations (ODEs). |
Mojtaba Nayyeri; Chengjin Xu; Franca Hoffmann; Mirza Mohtashim Alam; Jens Lehmann; Sahar Vahdati; | |
751 | KnowMAN: Weakly Supervised Multinomial Adversarial Networks Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We propose KnowMAN, an adversarial scheme that enables to control influence of signals associated with specific labeling functions. |
Luisa M?rz; Ehsaneddin Asgari; Fabienne Braune; Franziska Zimmermann; Benjamin Roth; | |
752 | ONION: A Simple and Effective Defense Against Textual Backdoor Attacks Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we propose a simple and effective textual backdoor defense named ONION, which is based on outlier word detection and, to the best of our knowledge, is the first method that can handle all the textual backdoor attack situations. |
Fanchao Qi; Yangyi Chen; Mukai Li; Yuan Yao; Zhiyuan Liu; Maosong Sun; | code |
753 | Value-aware Approximate Attention Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this work, we argue that research efforts should be directed towards approximating the true output of the attention sub-layer, which includes the value vectors. |
Ankit Gupta; Jonathan Berant; | |
754 | Contrastive Domain Adaptation for Question Answering Using Limited Text Corpora Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we propose a novel framework for domain adaptation called contrastive domain adaptation for QA (CAQA). |
Zhenrui Yue; Bernhard Kratzwald; Stefan Feuerriegel; | |
755 | Case-based Reasoning for Natural Language Queries Over Knowledge Bases Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We propose a neuro-symbolic CBR approach (CBR-KBQA) for question answering over large knowledge bases. |
Rajarshi Das; Manzil Zaheer; Dung Thai; Ameya Godbole; Ethan Perez; Jay Yoon Lee; Lizhen Tan; Lazaros Polymenakos; Andrew McCallum; | |
756 | Distantly-Supervised Dense Retrieval Enables Open-Domain Question Answering Without Evidence Annotation Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We introduce a novel approach (DistDR) that iteratively improves over a weak retriever by alternately finding evidence from the up-to-date model and encouraging the model to learn the most likely evidence. |
Chen Zhao; Chenyan Xiong; Jordan Boyd-Graber; Hal Daum? III; | code |
757 | What’s in A Name? Answer Equivalence For Open-Domain Question Answering Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: This work explores mining alias entities from knowledge bases and using them as additional gold answers (i.e., equivalent answers). |
Chenglei Si; Chen Zhao; Jordan Boyd-Graber; | |
758 | Evaluation Paradigms in Question Answering Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: Question answering (QA) primarily descends from two branches of research: (1) Alan Turing’s investigation of machine intelligence at Manchester University and (2) Cyril Cleverdon’s comparison of library card catalog indices at Cranfield University. This position paper names and distinguishes these paradigms. |
Pedro Rodriguez; Jordan Boyd-Graber; | |
759 | Numerical Reasoning in Machine Reading Comprehension Tasks: Are We There Yet? Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we present a controlled study on some of the top-performing model architectures for the task of numerical reasoning. |
Hadeel Al-Negheimish; Pranava Madhyastha; Alessandra Russo; | |
760 | Set Generation Networks for End-to-End Knowledge Base Population Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we formulate end-to-end KBP as a direct set generation problem, avoiding considering the order of multiple facts. |
Dianbo Sui; Chenhao Wang; Yubo Chen; Kang Liu; Jun Zhao; Wei Bi; | |
761 | Knowing False Negatives: An Adversarial Training Method for Distantly Supervised Relation Extraction Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: First, it finds out possible FN samples by heuristically leveraging the memory mechanism of deep neural networks. Then, it aligns those unlabeled data with the training data into a unified feature space by adversarial training to assign pseudo labels and further utilize the information contained in them. |
Kailong Hao; Botao Yu; Wei Hu; | |
762 | Progressive Adversarial Learning for Bootstrapping: A Case Study on Entity Set Expansion Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we propose BootstrapGAN, a new learning method for bootstrapping which jointly models the bootstrapping process and the boundary learning process in a GAN framework. |
Lingyong Yan; Xianpei Han; Le Sun; | |
763 | Uncovering Main Causalities for Long-tailed Information Extraction Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: This motivates us to propose counterfactual IE (CFIE), a novel framework that aims to uncover the main causalities behind data in the view of causal inference. |
Guoshun Nan; Jiaqi Zeng; Rui Qiao; Zhijiang Guo; Wei Lu; | |
764 | Maximal Clique Based Non-Autoregressive Open Information Extraction Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: To break this bottleneck, we propose MacroIE, a novel non-autoregressive framework for OpenIE. |
Bowen Yu; Yucheng Wang; Tingwen Liu; Hongsong Zhu; Limin Sun; Bin Wang; | |
765 | A Relation-Oriented Clustering Method for Open Relation Extraction Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this work, we propose a relation-oriented clustering model and use it to identify the novel relations in the unlabeled data. |
Jun Zhao; Tao Gui; Qi Zhang; Yaqian Zhou; | |
766 | Exploring Methods for Generating Feedback Comments for Writing Learning Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: To shed light on these points, we investigate a wider range of methods for generating many feedback comments in this study. |
Kazuaki Hanawa; Ryo Nagata; Kentaro Inui; | |
767 | A Role-Selected Sharing Network for Joint Machine-Human Chatting Handoff and Service Satisfaction Analysis Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this study, we propose a novel model, Role-Selected Sharing Network (RSSN), which integrates both dialogue satisfaction estimation and handoff prediction in one multi-task learning framework. |
Jiawei Liu; Kaisong Song; Yangyang Kang; Guoxiu He; Zhuoren Jiang; Changlong Sun; Wei Lu; Xiaozhong Liu; | |
768 | Meta Distant Transfer Learning for Pre-trained Language Models Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: Inspired by meta-learning, we propose the Meta Distant Transfer Learning (Meta-DTL) framework to learn the cross-task knowledge for PLM-based methods. |
Chengyu Wang; Haojie Pan; Minghui Qiu; Jun Huang; Fei Yang; Yin Zhang; | |
769 | UniKER: A Unified Framework for Combining Embedding and Definite Horn Rule Reasoning for Knowledge Graph Inference Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we propose a novel framework UniKER to address these challenges by restricting logical rules to be definite Horn rules, which can fully exploit the knowledge in logical rules and enable the mutual enhancement of logical rule-based reasoning and KGE in an extremely efficient way. |
Kewei Cheng; Ziqing Yang; Ming Zhang; Yizhou Sun; | |
770 | Wasserstein Selective Transfer Learning for Cross-domain Text Mining Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: To alleviate this issue, we propose a Wasserstein Selective Transfer Learning (WSTL) method. |
Lingyun Feng; Minghui Qiu; Yaliang Li; Haitao Zheng; Ying Shen; | |
771 | Jointly Learning to Repair Code and Generate Commit Message Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We propose a novel task of jointly repairing program codes and generating commit messages. |
Jiaqi Bai; Long Zhou; Ambrosio Blanco; Shujie Liu; Furu Wei; Ming Zhou; Zhoujun Li; | |
772 | Inflate and Shrink:Enriching and Reducing Interactions for Fast Text-Image Retrieval Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this work, we propose an inflating and shrinking approach to further boost the efficiency and accuracy of late-interaction methods. |
Haoliang Liu; Tan Yu; Ping Li; | |
773 | On Pursuit of Designing Multi-modal Transformer for Video Grounding Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: To this end, we reformulate video grounding as a set prediction task and propose a novel end-to-end multi-modal Transformer model, dubbed as GTR. |
Meng Cao; Long Chen; Mike Zheng Shou; Can Zhang; Yuexian Zou; | |
774 | COVR: A Test-Bed for Visually Grounded Compositional Generalization with Real Images Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this work, we propose COVR, a new test-bed for visually-grounded compositional generalization with real images. |
Ben Bogin; Shivanshu Gupta; Matt Gardner; Jonathan Berant; | |
775 | Vision-and-Language or Vision-for-Language? On Cross-Modal Influence in Multimodal Transformers Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We propose a diagnostic method based on cross-modal input ablation to assess the extent to which these models actually integrate cross-modal information. |
Stella Frank; Emanuele Bugliarello; Desmond Elliott; | |
776 | HypMix: Hyperbolic Interpolative Data Augmentation Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We propose HypMix, a novel model-, data-, and modality-agnostic interpolative data augmentation technique operating in the hyperbolic space, which captures the complex geometry of input and hidden state hierarchies better than its contemporaries. |
Ramit Sawhney; Megh Thakkar; Shivam Agarwal; Di Jin; Diyi Yang; Lucie Flek; | |
777 | Integrating Deep Event-Level and Script-Level Information for Script Event Prediction Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we propose a Transformer-based model, called MCPredictor, which integrates deep event-level and script-level information for script event prediction. |
Long Bai; Saiping Guan; Jiafeng Guo; Zixuan Li; Xiaolong Jin; Xueqi Cheng; | |
778 | QA-Align: Representing Cross-Text Content Overlap By Aligning Question-Answer Propositions Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We employ crowd-workers for constructing a dataset of QA-based alignments, and present a baseline QA alignment model trained over our dataset. |
Daniela Brook Weiss; Paul Roit; Ayal Klein; Ori Ernst; Ido Dagan; | |
779 | PICARD: Parsing Incrementally for Constrained Auto-Regressive Decoding from Language Models Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We propose PICARD (code available at https://github.com/ElementAI/picard), a method for constraining auto-regressive decoders of language models through incremental parsing. |
Torsten Scholak; Nathan Schucher; Dzmitry Bahdanau; | code |
780 | Exploiting Twitter As Source of Large Corpora of Weakly Similar Pairs for Semantic Sentence Embeddings Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We instead propose a language-independent approach to build large datasets of pairs of informal texts weakly similar, without manual human effort, exploiting Twitter’s intrinsic powerful signals of relatedness: replies and quotes of tweets. |
Marco Di Giovanni; Marco Brambilla; | code |
781 | Guilt By Association: Emotion Intensities in Lexical Representations Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this study, we consider the task of estimating word-level emotion intensity scores for specific emotions, exploring unsupervised, supervised, and finally a self-supervised method of extracting emotional associations from pretrained vectors and models. |
Shahab Raji; Gerard de Melo; | |
782 | Using Sociolinguistic Variables to Reveal Changing Attitudes Towards Sexuality and Gender Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: Here, we develop computational methods to study word choice within a sociolinguistic lexical variable-alternate words used to express the same concept-in order to test for change in the United States towards sexuality and gender. |
Sky CH-Wang; David Jurgens; | |
783 | Identifying Morality Frames in Political Tweets Using Relational Learning Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we introduce morality frames, a representation framework for organizing moral attitudes directed at different entities, and come up with a novel and high-quality annotated dataset of tweets written by US politicians. |
Shamik Roy; Maria Leonor Pacheco; Dan Goldwasser; | |
784 | Measuring Sentence-Level and Aspect-Level (Un)certainty in Science Communications Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: Here, we introduce a new study of certainty that models both the level and the aspects of certainty in scientific findings. |
Jiaxin Pei; David Jurgens; | code |
785 | Assessing The Reliability of Word Embedding Gender Bias Measures Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we assess three types of reliability of word embedding gender bias measures, namely test-retest reliability, inter-rater consistency and internal consistency. |
Yupei Du; Qixiang Fang; Dong Nguyen; | |
786 | Rumor Detection on Twitter with Claim-Guided Hierarchical Graph Attention Networks Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this study, to substantially reinforces the interaction of user opinions while alleviating the negative impact imposed by irrelevant posts, we first represent the conversation thread as an undirected interaction graph. We then present a Claim-guided Hierarchical Graph Attention Network for rumor classification, which enhances the representation learning for responsive posts considering the entire social contexts and attends over the posts that can semantically infer the target claim. |
Hongzhan Lin; Jing Ma; Mingfei Cheng; Zhiwei Yang; Liangliang Chen; Guang Chen; | |
787 | Learning Bill Similarity with Annotated and Augmented Corpora of Bills Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we overcome these limitations by proposing a 5-class classification task that closely reflects the nature of the bill generation process. |
Jiseon Kim; Elden Griggs; In Song Kim; Alice Oh; | |
788 | SWEAT: Scoring Polarization of Topics Across Different Corpora Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we propose the Sliced Word Embedding Association Test (SWEAT), a novel statistical measure to compute the relative polarization of a topical wordset across two distributional representations. |
Federico Bianchi; Marco Marelli; Paolo Nicoli; Matteo Palmonari; | |
789 | So You Think You’re Funny?: Rating The Humour Quotient in Standup Comedy Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this work, we create a multi-modal humour-annotated dataset (~40 hours) using stand-up comedy clips. |
Anirudh Mittal; Pranav Jeevan P; Prerak Gandhi; Diptesh Kanojia; Pushpak Bhattacharyya; | |
790 | Was It stated or Was It claimed: How Linguistic Bias Affects Generative Language Models Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we test whether generative language models (including GPT-2 (CITATION) are sensitive to these linguistic framing effects. |
Roma Patel; Ellie Pavlick; | |
791 | PAUSE: Positive and Annealed Unlabeled Sentence Embedding Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: To that end, we propose a generic and end-to-end approach – PAUSE (Positive and Annealed Unlabeled Sentence Embedding), capable of learning high-quality sentence embeddings from a partially labeled dataset. |
Lele Cao; Emil Larsson; Vilhelm von Ehrenheim; Dhiana Deva Cavalcanti Rocha; Anna Martin; Sonja Horn; | |
792 | A Simple Geometric Method for Cross-Lingual Linguistic Transformations with Pre-trained Autoencoders Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: For efficient learning, we investigate the use of a geometric mapping in embedding space to transform linguistic properties, without any tuning of the pre-trained sentence encoder or decoder. |
Maarten De Raedt; Fr?deric Godin; Pieter Buteneers; Chris Develder; Thomas Demeester; | |
793 | An Information-Theoretic Characterization of Morphological Fusion Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We present an information-theoretic measure, called informational fusion, to quantify the degree of fusion of a given set of morphological features in a surface form, which naturally provides such a graded scale. |
Neil Rathi; Michael Hahn; Richard Futrell; | |
794 | The Effect of Efficient Messaging and Input Variability on Neural-Agent Iterated Language Learning Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: Our simulations show that neural agents mainly strive to maintain the utterance type distribution observed during learning, instead of developing a more efficient or systematic language. |
Yuchen Lian; Arianna Bisazza; Tessa Verhoef; | |
795 | On Classifying Whether Two Texts Are on The Same Side of An Argument Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: Our evaluation shows that current state-of-the-art approaches cannot determine same side stance by considering only domain-independent linguistic similarity features, but appear to require domain knowledge and semantic inference, too. |
Erik K?rner; Gregor Wiedemann; Ahmad Dawar Hakimi; Gerhard Heyer; Martin Potthast; | |
796 | Chinese Opinion Role Labeling with Corpus Translation: A Pivot Study Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: Unlike most of the previous works focusing on the English language, in this paper, we present the first work of Chinese ORL. |
Ranran Zhen; Rui Wang; Guohong Fu; Chengguo Lv; Meishan Zhang; | |
797 | MassiveSumm: A Very Large-scale, Very Multilingual, News Summarisation Dataset Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we present a large-scale multilingual summarisation dataset containing articles in 92 languages, spread across 28.8 million articles, in more than 35 writing scripts. |
Daniel Varab; Natalie Schluter; | |
798 | AUTOSUMM: Automatic Model Creation for Text Summarization Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we propose methods to automatically create deep learning models for the tasks of extractive and abstractive text summarization. |
Sharmila Reddy Nangi; Atharv Tyagi; Jay Mundra; Sagnik Mukherjee; Raj Snehal; Niyati Chhaya; Aparna Garimella; | |
799 | Investigating The Helpfulness of Word-Level Quality Estimation for Post-Editing Machine Translation Output Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we address both research questions with real and simulated word-level QE, visualizations, and user studies, where time, subjective ratings, and quality of the final translations are assessed. |
Raksha Shenoy; Nico Herbig; Antonio Kr?ger; Josef van Genabith; | |
800 | UNKs Everywhere: Adapting Multilingual Language Models to New Scripts Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this work, we propose a series of novel data-efficient methods that enable quick and effective adaptation of pretrained multilingual models to such low-resource languages and unseen scripts. |
Jonas Pfeiffer; Ivan Vulic; Iryna Gurevych; Sebastian Ruder; | |
801 | Neural Machine Translation Quality and Post-Editing Performance Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: Across all models, we found that better MT systems indeed lead to fewer changes in the sentences in this industry setting. |
Vil?m Zouhar; Martin Popel; Ondrej Bojar; Ale? Tamchyna; | |
802 | XTREME-R: Towards More Challenging and Nuanced Multilingual Evaluation Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In order to catalyze meaningful progress, we extend XTREME to XTREME-R, which consists of an improved set of ten natural language understanding tasks, including challenging language-agnostic retrieval tasks, and covers 50 typologically diverse languages. |
Sebastian Ruder; Noah Constant; Jan Botha; Aditya Siddhant; Orhan Firat; Jinlan Fu; Pengfei Liu; Junjie Hu; Dan Garrette; Graham Neubig; Melvin Johnson; | |
803 | Contrastive Conditioning for Assessing Disambiguation in MT: A Case Study of Distilled Bias Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We propose contrastive conditioning as a reference-free black-box method for detecting disambiguation errors. |
Jannis Vamvas; Rico Sennrich; | |
804 | Measuring Association Between Labels and Free-Text Rationales Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We investigate the extent to which the labels and rationales predicted by these models are associated, a necessary property of faithful explanation. |
Sarah Wiegreffe; Ana Marasovic; Noah A. Smith; | |
805 | Discretized Integrated Gradients for Explaining Language Models Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: Here we propose Discretized Integrated Gradients (DIG), which allows effective attribution along non-linear interpolation paths. |
Soumya Sanyal; Xiang Ren; | |
806 | Putting Words in BERT’s Mouth: Navigating Contextualized Vector Spaces with Pseudowords Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We present a method for exploring regions around individual points in a contextualized vector space (particularly, BERT space), as a way to investigate how these regions correspond to word senses. |
Taelin Karidi; Yichu Zhou; Nathan Schneider; Omri Abend; Vivek Srikumar; | |
807 | Rationales for Sequential Predictions Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: Enumerating all subsets is intractable, so we propose an efficient greedy algorithm to approximate this objective. |
Keyon Vafa; Yuntian Deng; David Blei; Alexander Rush; | |
808 | FastIF: Scalable Influence Functions for Efficient Model Interpretation and Debugging Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We present FastIF, a set of simple modifications to influence functions that significantly improves their run-time. |
Han Guo; Nazneen Rajani; Peter Hase; Mohit Bansal; Caiming Xiong; | |
809 | Studying Word Order Through Iterative Shuffling Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: This view rests upon a hypothesis that has not yet been empirically tested: that word order encodes meaning essential to performing these tasks. We refute this hypothesis in many cases: in the GLUE suite and in various genres of English text, the words in a sentence or phrase can rarely be permuted to form a phrase carrying substantially different information. |
Nikolay Malkin; Sameera Lanka; Pranav Goel; Nebojsa Jojic; | |
810 | Distantly-Supervised Named Entity Recognition with Noise-Robust Learning and Language Model Augmented Self-Training Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we propose (1) a noise-robust learning scheme comprised of a new loss function and a noisy label removal step, for training NER models on distantly-labeled data, and (2) a self-training method that uses contextualized augmentations created by pre-trained language models to improve the generalization ability of the NER model. |
Yu Meng; Yunyi Zhang; Jiaxin Huang; Xuan Wang; Yu Zhang; Heng Ji; Jiawei Han; | |
811 | Open Knowledge Graphs Canonicalization Using Variational Autoencoders Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this work, we propose Canonicalizing Using Variational AutoEncoders and Side Information (CUVA), a joint model to learn both embeddings and cluster assignments in an end-to-end approach, which leads to a better vector representation for the noun and relation phrases. |
Sarthak Dash; Gaetano Rossiello; Nandana Mihindukulasooriya; Sugato Bagchi; Alfio Gliozzo; | |
812 | HittER: Hierarchical Transformers for Knowledge Graph Embeddings Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We propose HittER, a Hierarchical Transformer model to jointly learn Entity-relation composition and Relational contextualization based on a source entity’s neighborhood. |
Sanxing Chen; Xiaodong Liu; Jianfeng Gao; Jian Jiao; Ruofei Zhang; Yangfeng Ji; | |
813 | Few-Shot Named Entity Recognition: An Empirical Baseline Study Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: This paper presents an empirical study to efficiently build named entity recognition (NER) systems when a small amount of in-domain labeled data is available. |
Jiaxin Huang; Chunyuan Li; Krishan Subudhi; Damien Jose; Shobana Balakrishnan; Weizhu Chen; Baolin Peng; Jianfeng Gao; Jiawei Han; | |
814 | XLEnt: Mining A Large Cross-lingual Entity Dataset with Lexical-Semantic-Phonetic Word Alignment Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: To address this, we propose Lexical-Semantic-Phonetic Align (LSP-Align), a technique to automatically mine cross-lingual entity lexica from mined web data. |
Ahmed El-Kishky; Adithya Renduchintala; James Cross; Francisco Guzm?n; Philipp Koehn; | |
815 | Utilizing Relative Event Time to Enhance Event-Event Temporal Relation Extraction Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we propose a joint model for event-event temporal relation classification and an auxiliary task, relative event time prediction, which predicts the event time as real numbers. |
Haoyang Wen; Heng Ji; | |
816 | Separating Retention from Extraction in The Evaluation of End-to-end Relation Extraction Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper we propose two experiments confirming that retention of known facts is a key factor of performance on standard benchmarks. |
Bruno Taill?; Vincent Guigue; Geoffrey Scoutheeten; Patrick Gallinari; | |
817 | Automatic Text Evaluation Through The Lens of Wasserstein Barycenters Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: A new metric BaryScore to evaluate text generation based on deep contextualized embeddings (e.g., BERT, Roberta, ELMo) is introduced. |
Pierre Colombo; Guillaume Staerman; Chlo? Clavel; Pablo Piantanida; | |
818 | Visually Grounded Reasoning Across Languages and Cultures Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In particular, we let the selection of both concepts and images be entirely driven by native speakers, rather than scraping them automatically. |
Fangyu Liu; Emanuele Bugliarello; Edoardo Maria Ponti; Siva Reddy; Nigel Collier; Desmond Elliott; | |
819 | Back to Square One: Artifact Detection, Training and Commonsense Disentanglement in The Winograd Schema Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: This paper suggests that the apparent progress on WS may not necessarily reflect progress in commonsense reasoning. |
Yanai Elazar; Hongming Zhang; Yoav Goldberg; Dan Roth; | |
820 | Robustness Evaluation of Entity Disambiguation Using Prior Probes: The Case of Entity Overshadowing Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: To provide a more adequate evaluation benchmark, we introduce the ShadowLink dataset, which includes 16K short text snippets annotated with entity mentions. |
Vera Provatorova; Samarth Bhargav; Svitlana Vakulenko; Evangelos Kanoulas; | |
821 | IndoNLI: A Natural Language Inference Dataset for Indonesian Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We present IndoNLI, the first human-elicited NLI dataset for Indonesian. |
Rahmad Mahendra; Alham Fikri Aji; Samuel Louvan; Fahrurrozi Rahman; Clara Vania; | |
822 | Agreeing to Disagree: Annotating Offensive Language Datasets with Annotators’ Disagreement Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: Following a trend that has emerged recently, we focus on the level of agreement among annotators while selecting data to create offensive language datasets, a task involving a high level of subjectivity. |
Elisa Leonardelli; Stefano Menini; Alessio Palmero Aprosio; Marco Guerini; Sara Tonelli; | |
823 | A Root of A Problem: Optimizing Single-Root Dependency Parsing Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We describe two approaches to single-root dependency parsing that yield significant speed ups in such parsing. |
Milo? Stanojevic; Shay B. Cohen; | |
824 | Efficient Sampling of Dependency Structure Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we adapt two spanning tree sampling algorithms to faithfully sample dependency trees from a graph subject to the root constraint. |
Ran Zmigrod; Tim Vieira; Ryan Cotterell; | |
825 | Reducing Discontinuous to Continuous Parsing with Pointer Network Reordering Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: Based on that, we propose to reduce discontinuous parsing to a continuous problem, which can then be directly solved by any off-the-shelf continuous parser. |
Daniel Fern?ndez-Gonz?lez; Carlos G?mez-Rodr?guez; | |
826 | A New Representation for Span-based CCG Parsing Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: This paper proposes a new representation for CCG derivations. |
Yoshihide Kato; Shigeki Matsubara; | |
827 | What to Pre-Train On? Efficient Intermediate Task Selection Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this work, we provide a comprehensive comparison of different methods for efficiently identifying beneficial tasks for intermediate transfer learning. |
Clifton Poth; Jonas Pfeiffer; Andreas R?ckl?; Iryna Gurevych; | |
828 | PermuteFormer: Efficient Relative Position Encoding for Long Sequences Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we discuss possible ways to add relative position encoding to Performer. |
Peng Chen; | |
829 | Block Pruning For Faster Transformers Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We introduce a block pruning approach targeting both small and fast models. |
Fran?ois Lagunas; Ella Charlaix; Victor Sanh; Alexander Rush; | |
830 | Finetuning Pretrained Transformers Into RNNs Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: As many models for natural language tasks are increasingly dependent on large-scale pretrained transformers, this work presents a viable approach to improving inference efficiency without repeating the expensive pretraining process. |
Jungo Kasai; Hao Peng; Yizhe Zhang; Dani Yogatama; Gabriel Ilharco; Nikolaos Pappas; Yi Mao; Weizhu Chen; Noah A. Smith; | |
831 | How to Train BERT with An Academic Budget Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We present a recipe for pretraining a masked language model in 24 hours using a single low-end deep learning server. |
Peter Izsak; Moshe Berchansky; Omer Levy; | |
832 | Beyond Preserved Accuracy: Evaluating Loyalty and Robustness of BERT Compression Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we propose two new metrics, label loyalty and probability loyalty that measure how closely a compressed model (i.e., student) mimics the original model (i.e., teacher). |
Canwen Xu; Wangchunshu Zhou; Tao Ge; Ke Xu; Julian McAuley; Furu Wei; | |
833 | IndoBERTweet: A Pretrained Language Model for Indonesian Twitter with Effective Domain-Specific Vocabulary Initialization Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We present IndoBERTweet, the first large-scale pretrained model for Indonesian Twitter that is trained by extending a monolingually-trained Indonesian BERT model with additive domain-specific vocabulary. |
Fajri Koto; Jey Han Lau; Timothy Baldwin; | |
834 | Pushing on Text Readability Assessment: A Transformer Meets Handcrafted Linguistic Features Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We report two essential improvements in readability assessment: 1. three novel features in advanced semantics and 2. the timely evidence that traditional ML models (e.g. Random Forest, using handcrafted features) can combine with transformers (e.g. RoBERTa) to augment model performance. |
Bruce W. Lee; Yoo Sung Jang; Jason Lee; | |
835 | Types of Out-of-Distribution Texts and How to Detect Them Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We categorize these examples as exhibiting a background shift or semantic shift, and find that the two major approaches to OOD detection, calibration and density estimation (language modeling for text), have distinct behavior on these types of OOD data. |
Udit Arora; William Huang; He He; | |
836 | Self-training with Few-shot Rationalization Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: To this end, we develop a multi-task teacher-student framework based on self-training pre-trained language models with limited task-specific labels and rationales and judicious sample selection to learn from informative pseudo-labeled examples. |
Meghana Moorthy Bhat; Alessandro Sordoni; Subhabrata Mukherjee; | |
837 | MTAdam: Automatic Balancing of Multiple Training Loss Terms Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this work, we generalize the Adam optimization algorithm to handle multiple loss terms. |
Itzik Malkiel; Lior Wolf; | |
838 | Softmax Tree: An Accurate, Fast Classifier When The Number of Classes Is Large Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We propose the softmax tree, consisting of a binary tree having sparse hyperplanes at the decision nodes (which make hard, not soft, decisions) and small softmax classifiers at the leaves. |
Arman Zharmagambetov; Magzhan Gabidolla; Miguel A. Carreira-Perpinan; | |
839 | Improving Distantly-Supervised Named Entity Recognition with Self-Collaborative Denoising Learning Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: To address this issue, we propose a robust learning paradigm named Self-Collaborative Denoising Learning (SCDL), which jointly trains two teacher-student networks in a mutually-beneficial manner to iteratively perform noisy label refinery. |
Xinghua Zhang; Bowen Yu; Tingwen Liu; Zhenyu Zhang; Jiawei Sheng; Xue Mengge; Hongbo Xu; | |
840 | Multivalent Entailment Graphs for Question Answering Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We make three contributions: (1) we reinterpret the Distributional Inclusion Hypothesis to model entailment between predicates of different valencies, like DEFEAT(Biden, Trump) entails WIN(Biden); (2) we actualize this theory by learning unsupervised Multivalent Entailment Graphs of open-domain predicates; and (3) we demonstrate the capabilities of these graphs on a novel question answering task. |
Nick McKenna; Liane Guillou; Mohammad Javad Hosseini; Sander Bijl de Vroe; Mark Johnson; Mark Steedman; | |
841 | Is Everything in Order? A Simple Way to Order Sentences Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We present Reorder-BART (Re-BART) that leverages a pre-trained Transformer-based model to identify a coherent order for a given set of shuffled sentences. |
Somnath Basu Roy Chowdhury; Faeze Brahman; Snigdha Chaturvedi; | |
842 | VeeAlign: Multifaceted Context Representation Using Dual Attention for Ontology Alignment Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this work, we propose VeeAlign, a Deep Learning based model that uses a novel dual-attention mechanism to compute the contextualized representation of a concept which, in turn, is used to discover alignments. |
Vivek Iyer; Arvind Agarwal; Harshit Kumar; | code |
843 | Finding Needles in A Haystack: Sampling Structurally-diverse Training Sets from Synthetic Data for Compositional Generalization Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this work, we investigate automatic generation of synthetic utterance-program pairs for improving compositional generalization in semantic parsing. |
Inbar Oren; Jonathan Herzig; Jonathan Berant; | |
844 | GeneSis: A Generative Approach to Substitutes in Context Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: To assess these issues, we proposed GeneSis (Generating Substitutes in contexts), the first generative approach to lexical substitution. |
Caterina Lacerra; Rocco Tripodi; Roberto Navigli; | code |
845 | Semi-Supervised Exaggeration Detection of Health Science Press Releases Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: Given this, we present a formalization of and study into the problem of exaggeration detection in science communication. |
Dustin Wright; Isabelle Augenstein; | |
846 | Phrase-BERT: Improved Phrase Embeddings from BERT with An Application to Corpus Exploration Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: In this paper, we propose a contrastive fine-tuning objective that enables BERT to produce more powerful phrase embeddings. |
Shufan Wang; Laure Thompson; Mohit Iyyer; | |
847 | Detecting Contact-Induced Semantic Shifts: What Can Embedding-Based Methods Do in Practice? Related Papers Related Patents Related Grants Related Orgs Related Experts Details Highlight: We introduce a new 80-item test set and conduct both quantitative and qualitative evaluations. |
Filip Miletic; Anne Przewozny-Desriaux; Ludovic Tanguy; |