0%

【arxiv论文】 Computation and Language 2020-06-29

目录

1. Pre-training via Paraphrasing [PDF] 摘要
2. ProVe -- Self-supervised pipeline for automated product replacement and cold-starting based on neural language models [PDF] 摘要
3. What they do when in doubt: a study of inductive biases in seq2seq learners [PDF] 摘要
4. LSBert: A Simple Framework for Lexical Simplification [PDF] 摘要
5. Evaluation of Text Generation: A Survey [PDF] 摘要
6. Dialog as a Vehicle for Lifelong Learning [PDF] 摘要
7. Graph Optimal Transport for Cross-Domain Alignment [PDF] 摘要
8. THEaiTRE: Artificial Intelligence to Write a Theatre Play [PDF] 摘要
9. LPar -- A Distributed Multi Agent platform for building Polyglot, Omni Channel and Industrial grade Natural Language Interfaces [PDF] 摘要
10. TURL: Table Understanding through Representation Learning [PDF] 摘要
11. Does the Whole Exceed its Parts? The Effect of AI Explanations on Complementary Team Performance [PDF] 摘要

摘要

1. Pre-training via Paraphrasing [PDF] 返回目录
  Mike Lewis, Marjan Ghazvininejad, Gargi Ghosh, Armen Aghajanyan, Sida Wang, Luke Zettlemoyer
Abstract: We introduce MARGE, a pre-trained sequence-to-sequence model learned with an unsupervised multi-lingual multi-document paraphrasing objective. MARGE provides an alternative to the dominant masked language modeling paradigm, where we self-supervise the reconstruction of target text by retrieving a set of related texts (in many languages) and conditioning on them to maximize the likelihood of generating the original. We show it is possible to jointly learn to do retrieval and reconstruction, given only a random initialization. The objective noisily captures aspects of paraphrase, translation, multi-document summarization, and information retrieval, allowing for strong zero-shot performance on several tasks. For example, with no additional task-specific training we achieve BLEU scores of up to 35.8 for document translation. We further show that fine-tuning gives strong performance on a range of discriminative and generative tasks in many languages, making MARGE the most generally applicable pre-training method to date.
摘要:介绍玛吉,预训练序列对序列模型的无监督多语种多文档转述客观教训。玛吉提供占主导地位的蒙面语言建模范例,在这里我们通过检索一组相关的文本(在许多语言)和调理上他们能够最大限度地产生原始的可能性自我监督目标文本重建的替代品。我们证明可以共同学习做恢复和重建,只给出一个随机的初始化。意译,翻译,多文档文摘,信息检索的目的,大肆捕捉方面,允许在几个任务,强大的零射门的表现。例如,没有额外的任务,具体的培训,我们实现了高达35.8对文件翻译的BLEU分数。进一步的研究表明微调给出了一系列的歧视性和生成任务强劲的性能在许多语言,让玛吉最普遍适用的岗前培训方法的日期。

2. ProVe -- Self-supervised pipeline for automated product replacement and cold-starting based on neural language models [PDF] 返回目录
  Andrei Ionut Damian, Laurentiu Piciu, Cosmin Mihai Marinescu
Abstract: In retail vertical industries, businesses are dealing with human limitation of quickly understanding and adapting to new purchasing behaviors. Moreover, retail businesses need to overcome the human limitation of properly managing a massive selection of products/brands/categories. These limitations lead to deficiencies from both commercial (e.g. loss of sales, decrease in customer satisfaction) and operational perspective (e.g. out-of-stock, over-stock). In this paper, we propose a pipeline approach based on Natural Language Understanding, for recommending the most suitable replacements for products that are out-of-stock. Moreover, we will propose a solution for managing products that were newly introduced in a retailer's portfolio with almost no transactional history. This solution will help businesses: automatically assign the new products to the right category; recommend complementary products for cross-sell from day 1; perform sales predictions even with almost no transactional history. Finally, the vector space model resulted by applying the pipeline presented in this paper is directly used as semantic information in deep learning-based demand forecasting solutions, leading to more accurate predictions. The whole research and experimentation process have been done using real-life private transactional data, however the source code is available on this https URL
摘要:在零售垂直行业,企业所面对的是快速理解和适应新的购买行为人的限制。此外,零售企业需要克服的妥善管理的产品/品牌/类别大规模选择的人的限制。这些限制导致从商业的不足(如销售损失,客户满意度下降)和操作的角度来看(例如外的股票,在股票)。在本文中,我们提出了一种基于自然语言管线方法的理解,对推荐最合适的替代那些外的股票产品。此外,我们将提出用于管理都是新的零售商的投资组合几乎没有任何交易历史推出的产品解决方案。该解决方案将帮助企业:自动分配的新产品,以正确的类别;推荐配套产品为1天的交叉销售;进行销售预测,甚至几乎没有交易的历史。最后,向量空间模型,导致通过应用本文提出的管道直接用作深学习型需求预测解决方案的语义信息,从而更准确的预测。整个研究和实验过程中一直在使用真实生活的私人交易数据做了,但是源代码都可以在此HTTPS URL

3. What they do when in doubt: a study of inductive biases in seq2seq learners [PDF] 返回目录
  Eugene Kharitonov, Rahma Chaabouni
Abstract: Sequence-to-sequence (seq2seq) learners are widely used, but we still have only limited knowledge about what inductive biases shape the way they generalize. We address that by investigating how popular seq2seq learners generalize in tasks that have high ambiguity in the training data. We use SCAN and three new tasks to study learners' preferences for memorization, arithmetic, hierarchical, and compositional reasoning. Further, we connect to Solomonoff's theory of induction and propose to use description length as a principled and sensitive measure of inductive biases. In our experimental study, we find that LSTM-based learners can learn to perform counting, addition, and multiplication by a constant from a single training example. Furthermore, Transformer and LSTM-based learners show a bias toward the hierarchical induction over the linear one, while CNN-based learners prefer the opposite. On the SCAN dataset, we find that CNN-based, and, to a lesser degree, Transformer- and LSTM-based learners have a preference for compositional generalization over memorization. Finally, across all our experiments, description length proved to be a sensitive measure of inductive biases.
摘要:序列到序列(seq2seq)学习者被广泛使用,但我们还是有什么感应偏见塑造他们推广的方式仅限于知识。我们解决通过调查seq2seq学习者有训练数据的高不确定性的任务有多受欢迎一概而论。我们使用扫描和三个新任务,研究学生的喜好记忆,运算,层次和组合推理。此外,我们连接到感应Solomonoff的理论,并提出要使用说明长度感性偏见的原则性和敏感的措施。在我们的实验研究中,我们发现基于LSTM学习者可以学习常数从单一的训练例子来进行计数,加法和乘法。此外,变压器和基于LSTM学习者显示朝向分级感应在所述线性一个偏压,而基于CNN学习者喜欢相反。在扫描数据集,我们发现,基于CNN和,程度较轻,变压器 - 基于LSTM学习者有过记忆的组成概括的偏好。最后,在我们所有的实验中,描述长度被证明是感应式偏见的一个敏感指标。

4. LSBert: A Simple Framework for Lexical Simplification [PDF] 返回目录
  Jipeng Qiang, Yun Li, Yi Zhu, Yunhao Yuan, Xindong Wu
Abstract: Lexical simplification (LS) aims to replace complex words in a given sentence with their simpler alternatives of equivalent meaning, to simplify the sentence. Recently unsupervised lexical simplification approaches only rely on the complex word itself regardless of the given sentence to generate candidate substitutions, which will inevitably produce a large number of spurious candidates. In this paper, we propose a lexical simplification framework LSBert based on pretrained representation model Bert, that is capable of (1) making use of the wider context when both detecting the words in need of simplification and generating substitue candidates, and (2) taking five high-quality features into account for ranking candidates, including Bert prediction order, Bert-based language model, and the paraphrase database PPDB, in addition to the word frequency and word similarity commonly used in other LS methods. We show that our system outputs lexical simplifications that are grammatically correct and semantically appropriate, and obtains obvious improvement compared with these baselines, outperforming the state-of-the-art by 29.8 Accuracy points on three well-known benchmarks.
摘要:词汇简化(LS)的目标与他们简单的等价含义的替代品取代复杂的单词在句子给出,简化句子。最近无人监督的词汇简化办法只能依靠复杂的文字本身,无论给出的句子来产生候选替代,这将不可避免地产生大量虚假的候选人。在本文中,我们提出了一种基于预训练的表示模型伯特,其能够(1)利用所述更广泛的范围的当两个检测需要简化的话,并产生substitue候选,以及(2)拍摄的词汇简化框架LSBert五个高品质的功能考虑在内进行排名候选人,包括伯特预测顺序,基于伯特-语言模型,并且复述数据库PPDB,除了在其它LS方法中常用的词频率和字相似。我们证明了我们是正确的语法和语义适当,并与这些基线相比得到明显改善,29.8精度点上三个著名的基准跑赢国家的最先进的系统输出词汇简化。

5. Evaluation of Text Generation: A Survey [PDF] 返回目录
  Asli Celikyilmaz, Elizabeth Clark, Jianfeng Gao
Abstract: The paper surveys evaluation methods of natural language generation (NLG) systems that have been developed in the last few years. We group NLG evaluation methods into three categories: (1) human-centric evaluation metrics, (2) automatic metrics that require no training, and (3) machine-learned metrics. For each category, we discuss the progress that has been made and the challenges still being faced, with a focus on the evaluation of recently proposed NLG tasks and neural NLG models. We then present two case studies of automatic text summarization and long text generation, and conclude the paper by proposing future research directions.
摘要:已在过去几年中已开发自然语言生成(NLG)系统的纸调查评价方法。我们组NLG的评价方法分为三类:(1)以人为中心的评价指标,(2)不需要培训,以及(3)机器学习的指标自动指标。对于每个类别中,我们讨论已经取得和挑战仍然面临,重点对最近提出的NLG任务和神经NLG模型评估的进展情况。我们自动文摘和长文本生成的,然后现在两个案例研究,并通过提出今后的研究方向总结的纸。

6. Dialog as a Vehicle for Lifelong Learning [PDF] 返回目录
  Aishwarya Padmakumar, Raymond J. Mooney
Abstract: Dialog systems research has primarily been focused around two main types of applications - task-oriented dialog systems that learn to use clarification to aid in understanding a goal, and open-ended dialog systems that are expected to carry out unconstrained "chit chat" conversations. However, dialog interactions can also be used to obtain various types of knowledge that can be used to improve an underlying language understanding system, or other machine learning systems that the dialog acts over. In this position paper, we present the problem of designing dialog systems that enable lifelong learning as an important challenge problem, in particular for applications involving physically situated robots. We include examples of prior work in this direction, and discuss challenges that remain to be addressed.
摘要:对话系统的研究主要集中围绕两个主要类型的应用程序 - 学会使用说明为了帮助理解目标面向任务的对话系统,以及开放式按照预期进行不受约束的“闲聊”对话系统交谈。然而,对话互动也可以用来获得可用于该对话框上的作用,提高底层语言理解系统,或其它机器学习系统的各类知识。在此立场文件中,我们提出在设计对话系统,使终身学习作为一项重要的挑战的问题,特别是涉及物理位于机器人应用的问题。我们包括在这个方向前工作的例子,并讨论有待解决的挑战。

7. Graph Optimal Transport for Cross-Domain Alignment [PDF] 返回目录
  Liqun Chen, Zhe Gan, Yu Cheng, Linjie Li, Lawrence Carin, Jingjing Liu
Abstract: Cross-domain alignment between two sets of entities (e.g., objects in an image, words in a sentence) is fundamental to both computer vision and natural language processing. Existing methods mainly focus on designing advanced attention mechanisms to simulate soft alignment, with no training signals to explicitly encourage alignment. The learned attention matrices are also dense and lacks interpretability. We propose Graph Optimal Transport (GOT), a principled framework that germinates from recent advances in Optimal Transport (OT). In GOT, cross-domain alignment is formulated as a graph matching problem, by representing entities into a dynamically-constructed graph. Two types of OT distances are considered: (i) Wasserstein distance (WD) for node (entity) matching; and (ii) Gromov-Wasserstein distance (GWD) for edge (structure) matching. Both WD and GWD can be incorporated into existing neural network models, effectively acting as a drop-in regularizer. The inferred transport plan also yields sparse and self-normalized alignment, enhancing the interpretability of the learned model. Experiments show consistent outperformance of GOT over baselines across a wide range of tasks, including image-text retrieval, visual question answering, image captioning, machine translation, and text summarization.
摘要:两组实体之间的交叉域对准(例如,图像中的对象,在一个句子话)是既计算机视觉和自然语言处理的基础。现有的方法主要集中在设计先进注意机制,模拟软校准,没有训练信号明确鼓励对齐。博学的注意矩阵也密集和缺乏可解释性。我们建议图最优运输(GOT),一个原则性的框架,从优化交通运输(OT)的最新进展发芽。在GOT,跨域对准配制成图匹配问题,由代表实体成动态构建的曲线图。两种类型的OT的距离被认为是:(ⅰ)瓦瑟斯坦距离(WD)为节点(实体)匹配;和(ii)用于边缘(结构)匹配格罗莫夫-瓦瑟斯坦距离(GWD)。无论WD和GWD可以被纳入现有的神经网络模型,有效地充当一个下拉正则。推断运输计划也得到稀疏,自正则化调整,增强学习模型的可解释性。实验证明了在广泛的任务,包括图像,文本检索,可视化问答,图像字幕,机器翻译和文本摘要基线GOT的持续跑赢大市。

8. THEaiTRE: Artificial Intelligence to Write a Theatre Play [PDF] 返回目录
  Rudolf Rosa, Ondřej Dušek, Tom Kocmi, David Mareček, Tomáš Musil, Patrícia Schmidtová, Dominik Jurko, Ondřej Bojar, Daniel Hrbek, David Košťák, Martina Kinská, Josef Doležal, Klára Vosecká
Abstract: We present THEaiTRE, a starting project aimed at automatic generation of theatre play scripts. This paper reviews related work and drafts an approach we intend to follow. We plan to adopt generative neural language models and hierarchical generation approaches, supported by summarization and machine translation methods, and complemented with a human-in-the-loop approach.
摘要:我们提出THEaiTRE,起始项目,旨在自动生成剧场戏剧的脚本。本文回顾相关的工作,并起草我们打算遵循的方法。我们计划采用生成神经语言模型和层次生成方法,通过汇总和机器翻译方法的支持,并与人合的环办法补充。

9. LPar -- A Distributed Multi Agent platform for building Polyglot, Omni Channel and Industrial grade Natural Language Interfaces [PDF] 返回目录
  Pranav Sharma
Abstract: The goal of serving and delighting customers in a personal and near human like manner is very high on automation agendas of most Enterprises. Last few years, have seen huge progress in Natural Language Processing domain which has led to deployments of conversational agents in many enterprises. Most of the current industrial deployments tend to use Monolithic Single Agent designs that model the entire knowledge and skill of the Domain. While this approach is one of the fastest to market, the monolithic design makes it very hard to scale beyond a point. There are also challenges in seamlessly leveraging many tools offered by sub fields of Natural Language Processing and Information Retrieval in a single solution. The sub fields that can be leveraged to provide relevant information are, Question and Answer system, Abstractive Summarization, Semantic Search, Knowledge Graph etc. Current deployments also tend to be very dependent on the underlying Conversational AI platform (open source or commercial) , which is a challenge as this is a fast evolving space and no one platform can be considered future proof even in medium term of 3-4 years. Lately,there is also work done to build multi agent solutions that tend to leverage a concept of master agent. While this has shown promise, this approach still makes the master agent in itself difficult to scale. To address these challenges, we introduce LPar, a distributed multi agent platform for large scale industrial deployment of polyglot, diverse and inter-operable agents. The asynchronous design of LPar supports dynamically expandable domain. We also introduce multiple strategies available in the LPar system to elect the most suitable agent to service a customer query.
摘要:在个人和附近的人喜欢的方式提供服务和取悦客户的目标是大多数企业的自动化议程非常高。过去几年中,已经看到了在自然语言处理领域的巨大进步已导致会话代理的部署,在许多企业。目前大多数工业部署的倾向于使用单片单代理公司设计模型域的全部知识和技能。虽然这种方法是最快的市场之一,整体设计,使得它很难扩展到超过一个点。有在利用无缝地通过自然语言处理和信息检索的子场在一个单一的解决方案提供了许多工具,也带来了挑战。可以用来提供相关信息的子字段,问答系统,写意总结,语义搜索,知识图等。目前的部署也往往是非常依赖的基本会话AI平台(开源或商业),其上是一个挑战,因为这是一个快速发展的空间,并没有一个平台,甚至可以在3 - 4年的中期被认为是未来的证明。最近,也有做构建多代理解决方案,往往能够利用主代理的概念作品。虽然这已经显示出的承诺,这种做法还是让本身主代理难以形成规模。为了应对这些挑战,我们引入LPAR,分布式多代理平台,为大型工业部署多语种的,多样化和可互操作的代理。 LPAR的异步设计支持动态可扩展域。我们还介绍了在LPAR系统提供多种策略来选择最合适的代理服务一个客户的查询。

10. TURL: Table Understanding through Representation Learning [PDF] 返回目录
  Xiang Deng, Huan Sun, Alyssa Lees, You Wu, Cong Yu
Abstract: Relational tables on the Web store a vast amount of knowledge. Owing to the wealth of such tables, there has been tremendous progress on a variety of tasks in the area of table understanding. However, existing work generally relies on heavily-engineered task specific features and model architectures. In this paper, we present TURL, a novel framework that introduces the pre-training/finetuning paradigm to relational Web tables. During pre-training, our framework learns deep contextualized representations on relational tables in an unsupervised manner. Its universal model design with pre-trained representations can be applied to a wide range of tasks with minimal task-specific fine-tuning. Specifically, we propose a structure-aware Transformer encoder to model the row-column structure of relational tables, and present a new Masked Entity Recovery (MER) objective for pre-training to capture the semantics and knowledge in large-scale unlabeled data. We systematically evaluate TURL with a benchmark consisting of 6 different tasks for table understanding (e.g., relation extraction, cell filling). We show that TURL generalizes well to all tasks and substantially outperforms existing methods in almost all instances.
摘要:在网络上关系表存储知识的大量。由于这种表的财富,出现了多种在表了解该地区的任务的巨大进步。然而,现有的工作通常依赖于大量工程任务特定的功能和模式的架构。在本文中,我们目前TURL,一个新的框架,引入了前培训/微调模式到关系网表。在训练前,我们的框架获悉深上下文化关系表表示在无人监督的方式。它与预训练表示通用模型的设计可应用于广泛的用最少的特定任务的微调任务。具体来说,我们建议的结构感知的转换器编码器到关系表的行列结构模型,并提出了新的蒙面实体恢复(MER),客观上为前培训捕捉语义和知识的大型标签数据。我们系统地评价TURL用由对表的理解(例如,关系提取,细胞填充)6个不同的任务的基准。我们发现,TURL推广以及所有任务,并大幅优于在几乎所有情况下,现有的方法。

11. Does the Whole Exceed its Parts? The Effect of AI Explanations on Complementary Team Performance [PDF] 返回目录
  Gagan Bansal, Tongshuang Wu, Joyce Zhu, Raymond Fok, Besmira Nushi, Ece Kamar, Marco Tulio Ribeiro, Daniel S. Weld
Abstract: Increasingly, organizations are pairing humans with AI systems to improve decision-making and reducing costs. Proponents of human-centered AI argue that team performance can even further improve when the AI model explains its recommendations. However, a careful analysis of existing literature reveals that prior studies observed improvements due to explanations only when the AI, alone, outperformed both the human and the best human-AI team. This raises an important question: can explanations lead to complementary performance, i.e., with accuracy higher than both the human and the AI working alone? We address this question by devising comprehensive studies on human-AI teaming, where participants solve a task with help from an AI system without explanations and from one with varying types of AI explanation support. We carefully controlled to ensure comparable human and AI accuracy across experiments on three NLP datasets (two for sentiment analysis and one for question answering). While we found complementary improvements from AI augmentation, they were not increased by state-of-the-art explanations compared to simpler strategies, such as displaying the AI's confidence. We show that explanations increase the chance that humans will accept the AI's recommendation regardless of whether the AI is correct. While this clarifies the gains in team performance from explanations in prior work, it poses new challenges for human-centered AI: how can we best design systems to produce complementary performance? Can we develop explanatory approaches that help humans decide whether and when to trust AI input?
摘要:越来越多的企业配对与AI系统人类提高决策和降低成本。以人为本的AI的支持者认为,球队的表现甚至可以进一步提高,当AI模型解释了其建议。然而,现有的文献进行仔细分析发现,以前的研究中得到了改善,由于只有当AI,独自一人,跑赢人类与人类最好的-AI团队解释。这就提出了一个重要问题:能解释导致互补性能,即,具有精度高于两个人与AI单独工作?我们通过对人体-AI制定全面研究的合作,与会者解决由AI系统的帮助任务没有解释,并从一个与不同类型的AI解释支持解决这个问题。我们严格控制,以确保整个实验比较的人类和人工智能准确性三个NLP数据集(两个情感分析和一个问答)。虽然我们发现从AI增强互补性的改进,相对于简单的策略,如显示AI的信心,他们不会增加国家的最先进的解释。我们发现,解释增加,人类将接受AI的建议无论AI是否正确的机会。虽然这种澄清在之前的工作中的解释球队表现的收益,它构成了以人为本的AI新的挑战:怎样才能最好的设计系统,以产生互补的表现?我们可以开发解释方法是帮助人们决定是否以及何时相信AI输入?

注:中文为机器翻译结果!