diff --git a/bin/latex_to_unicode.py b/bin/latex_to_unicode.py index 4049c3a13a..c35c668b4d 100755 --- a/bin/latex_to_unicode.py +++ b/bin/latex_to_unicode.py @@ -23,6 +23,13 @@ # Entry(r'\textsc', None, 'sc', 'unary', False), # Entry(r'\sc', None, 'sc', 'setter', False), Entry(r"\url", None, "url", "unary", True), + Entry(r"\footnote", None, "fn", "unary", False), + Entry(r"\href", None, "delete", "unary", False), + # \cite and friends. There are a lot more, but these are the most common. + Entry(r"\cite", None, "cite", "unary", False), + Entry(r"\newcite", None, "cite", "unary", False), + Entry(r"\citet", None, "cite", "unary", False), + Entry(r"\citep", None, "cite", "unary", False), Entry(r"\fixedcase", None, "fixed-case", "unary", False), # for our internal use Entry(r"", None, None, "trivial", False), ] @@ -185,7 +192,7 @@ def latex_to_unicode(s): s = s.replace(r"\}", "}") def repl(s): - logging.warning("discarding control sequence {}".format(s.group(0))) + logging.warning(f"discarding control sequence '{s.group(0)}' from '{s.string}'") return "" s = re.sub(r"\\[A-Za-z]+ |\\.", repl, s) @@ -223,6 +230,27 @@ def visit(node): return node +def remove_notes(node): + def visit(node, outparent): + if isinstance(node, str): + outparent.append(node) + elif isinstance(node, list): + if openers[node[0].rstrip()].tag in ["fn", "delete"]: + return # without copying + elif openers[node[0].rstrip()].tag == "cite": + outparent.append("(CITATION)") + else: + outnode = [] + for child in node: + visit(child, outnode) + outparent.append(outnode) + + outroot = [] + visit(node, outroot) + assert len(outroot) == 1 + return outroot[0] + + def append_text(xnode, text): if len(xnode) == 0: xnode.text = (xnode.text or "") + text @@ -293,6 +321,7 @@ def latex_to_xml(s, fixed_case=False, trivial_math=False): tree = find_fixed_case(tree, conservative=True) if trivial_math: tree = flatten_trivial_math(tree) + tree = remove_notes(tree) tree = latextree_to_xml(tree) tree = xml_to_unicode(tree) return tree diff --git a/bin/normalize_anth.py b/bin/normalize_anth.py index f90ca251a7..c6a756ad9e 100755 --- a/bin/normalize_anth.py +++ b/bin/normalize_anth.py @@ -174,7 +174,11 @@ def normalize(oldnode, informat): ) ) oldtext = "".join(oldnode.itertext()) - newnode = latex_to_xml(oldtext, trivial_math=True, fixed_case=True) + newnode = latex_to_xml( + oldtext, + trivial_math=True, + fixed_case=oldnode.tag in ["title", "booktitle"], + ) newnode.tag = oldnode.tag newnode.attrib.update(oldnode.attrib) replace_node(oldnode, newnode) diff --git a/data/xml/D19.xml b/data/xml/D19.xml index ec70cd199b..a40ebc0c3b 100644 --- a/data/xml/D19.xml +++ b/data/xml/D19.xml @@ -543,7 +543,7 @@ KevinZhou TengchaoLv 476–485 - Most of the current effective methods for text classification tasks are based on large-scale labeled data and a great number of parameters, but when the supervised training data are few and difficult to be collected, these models are not available. In this work, we propose a hierarchical attention prototypical networks (HAPN) for few-shot text classification. We design the feature level, word level, and instance level multi cross attention for our model to enhance the expressive ability of semantic space, so it can highlight or weaken the importance of the features, words, and instances separately. We verify the effectiveness of our model on two standard benchmark few-shot text classification datasets – FewRel and CSID, and achieve the state-of-the-art performance. The visualization of hierarchical attention layers illustrates that our model can capture more important features, words, and instances. In addition, our attention mechanism increases support set augmentability and accelerates convergence speed in the training stage. + Most of the current effective methods for text classification tasks are based on large-scale labeled data and a great number of parameters, but when the supervised training data are few and difficult to be collected, these models are not available. In this work, we propose a hierarchical attention prototypical networks (HAPN) for few-shot text classification. We design the feature level, word level, and instance level multi cross attention for our model to enhance the expressive ability of semantic space, so it can highlight or weaken the importance of the features, words, and instances separately. We verify the effectiveness of our model on two standard benchmark few-shot text classification datasets—FewRel and CSID, and achieve the state-of-the-art performance. The visualization of hierarchical attention layers illustrates that our model can capture more important features, words, and instances. In addition, our attention mechanism increases support set augmentability and accelerates convergence speed in the training stage. D19-1045 10.18653/v1/D19-1045 @@ -874,7 +874,7 @@ Towards Linear Time Neural Machine Translation with Capsule Networks MingxuanWang 803–812 - In this study, we first investigate a novel capsule network with dynamic routing for linear time Neural Machine Translation (NMT), referred as CapsNMT. CapsNMT uses an aggregation mechanism to map the source sentence into a matrix with pre-determined size, and then applys a deep LSTM network to decode the target sequence from the source representation. Unlike the previous work sutskever2014sequence to store the source sentence with a passive and bottom-up way, the dynamic routing policy encodes the source sentence with an iterative process to decide the credit attribution between nodes from lower and higher layers. CapsNMT has two core properties: it runs in time that is linear in the length of the sequences and provides a more flexible way to aggregate the part-whole information of the source sentence. On WMT14 English-German task and a larger WMT14 English-French task, CapsNMT achieves comparable results with the Transformer system. To the best of our knowledge, this is the first work that capsule networks have been empirically investigated for sequence to sequence problems. + In this study, we first investigate a novel capsule network with dynamic routing for linear time Neural Machine Translation (NMT), referred as CapsNMT. CapsNMT uses an aggregation mechanism to map the source sentence into a matrix with pre-determined size, and then applys a deep LSTM network to decode the target sequence from the source representation. Unlike the previous work (CITATION) to store the source sentence with a passive and bottom-up way, the dynamic routing policy encodes the source sentence with an iterative process to decide the credit attribution between nodes from lower and higher layers. CapsNMT has two core properties: it runs in time that is linear in the length of the sequences and provides a more flexible way to aggregate the part-whole information of the source sentence. On WMT14 English-German task and a larger WMT14 English-French task, CapsNMT achieves comparable results with the Transformer system. To the best of our knowledge, this is the first work that capsule networks have been empirically investigated for sequence to sequence problems. D19-1074 10.18653/v1/D19-1074 @@ -1004,7 +1004,7 @@ Xin-YuDai JiajunChen 931–941 - Previous studies have shown that neural machine translation (NMT) models can benefit from explicitly modeling translated () and untranslated () source contents as recurrent states zheng2018modeling. However, this less interpretable recurrent process hinders its power to model the dynamic updating of and contents during decoding. In this paper, we propose to model the dynamic principles by explicitly separating source words into groups of translated and untranslated contents through parts-to-wholes assignment. The assignment is learned through a novel variant of routing-by-agreement mechanism sabour2017dynamic, namely Guided Dynamic Routing, where the translating status at each decoding step guides the routing process to assign each source word to its associated group (i.e., translated or untranslated content) represented by a capsule, enabling translation to be made from holistic context. Experiments show that our approach achieves substantial improvements over both Rnmt and Transformer by producing more adequate translations. Extensive analysis demonstrates that our method is highly interpretable, which is able to recognize the translated and untranslated contents as expected. + Previous studies have shown that neural machine translation (NMT) models can benefit from explicitly modeling translated () and untranslated () source contents as recurrent states (CITATION). However, this less interpretable recurrent process hinders its power to model the dynamic updating of and contents during decoding. In this paper, we propose to model the dynamic principles by explicitly separating source words into groups of translated and untranslated contents through parts-to-wholes assignment. The assignment is learned through a novel variant of routing-by-agreement mechanism (CITATION), namely Guided Dynamic Routing, where the translating status at each decoding step guides the routing process to assign each source word to its associated group (i.e., translated or untranslated content) represented by a capsule, enabling translation to be made from holistic context. Experiments show that our approach achieves substantial improvements over both Rnmt and Transformer by producing more adequate translations. Extensive analysis demonstrates that our method is highly interpretable, which is able to recognize the translated and untranslated contents as expected. D19-1086 D19-1086.Attachment.pdf 10.18653/v1/D19-1086 @@ -2849,7 +2849,7 @@ ChandraBhagavatula YejinChoi 2391–2401 - Understanding narratives requires reading between the lines, which in turn, requires interpreting the likely causes and effects of events, even when they are not mentioned explicitly. In this paper, we introduce Cosmos QA, a large-scale dataset of 35,600 problems that require commonsense-based reading comprehension, formulated as multiple-choice questions. In stark contrast to most existing reading comprehension datasets where the questions focus on factual and literal understanding of the context paragraph, our dataset focuses on reading between the lines over a diverse collection of people’s everyday narratives, asking such questions as “what might be the possible reason of ...?”, or “what would have happened if ...” that require reasoning beyond the exact text spans in the context. To establish baseline performances on Cosmos QA, we experiment with several state-of-the-art neural architectures for reading comprehension, and also propose a new architecture that improves over the competitive baselines. Experimental results demonstrate a significant gap between machine (68.4%) and human performance (94%), pointing to avenues for future research on commonsense machine comprehension. Dataset, code and leaderboard is publicly available at https://wilburone.github.io/cosmos. + Understanding narratives requires reading between the lines, which in turn, requires interpreting the likely causes and effects of events, even when they are not mentioned explicitly. In this paper, we introduce Cosmos QA, a large-scale dataset of 35,600 problems that require commonsense-based reading comprehension, formulated as multiple-choice questions. In stark contrast to most existing reading comprehension datasets where the questions focus on factual and literal understanding of the context paragraph, our dataset focuses on reading between the lines over a diverse collection of people’s everyday narratives, asking such questions as “what might be the possible reason of ...?", or “what would have happened if ..." that require reasoning beyond the exact text spans in the context. To establish baseline performances on Cosmos QA, we experiment with several state-of-the-art neural architectures for reading comprehension, and also propose a new architecture that improves over the competitive baselines. Experimental results demonstrate a significant gap between machine (68.4%) and human performance (94%), pointing to avenues for future research on commonsense machine comprehension. Dataset, code and leaderboard is publicly available at https://wilburone.github.io/cosmos. D19-1243 D19-1243.Attachment.pdf 10.18653/v1/D19-1243 @@ -3449,7 +3449,7 @@ ShafiqJoty IrinaTemnikova PreslavNakov - 2964–2973 + 2964–2975 The ongoing neural revolution in machine translation has made it easier to model larger contexts beyond the sentence-level, which can potentially help resolve some discourse-level ambiguities such as pronominal anaphora, thus enabling better translations. Unfortunately, even when the resulting improvements are seen as substantial by humans, they remain virtually unnoticed by traditional automatic evaluation measures like BLEU, as only a few words end up being affected. Thus, specialized evaluation measures are needed. With this aim in mind, we contribute an extensive, targeted dataset that can be used as a test suite for pronoun translation, covering multiple source languages and different pronoun errors drawn from real system translations, for English. We further propose an evaluation measure to differentiate good and bad pronoun translations. We also conduct a user study to report correlations with human judgments. D19-1294 D19-1294.Attachment.zip @@ -3459,7 +3459,7 @@ A Regularization Approach for Incorporating Event Knowledge and Coreference Relations into Neural Discourse Parsing ZeyuDai RuihongHuang - 2974–2985 + 2976–2987 We argue that external commonsense knowledge and linguistic constraints need to be incorporated into neural network models for mitigating data sparsity issues and further improving the performance of discourse parsing. Realizing that external knowledge and linguistic constraints may not always apply in understanding a particular context, we propose a regularization approach that tightly integrates these constraints with contexts for deriving word representations. Meanwhile, it balances attentions over contexts and constraints through adding a regularization term into the objective function. Experiments show that our knowledge regularization approach outperforms all previous systems on the benchmark dataset PDTB for discourse parsing. D19-1295 10.18653/v1/D19-1295 @@ -3467,7 +3467,7 @@ Weakly Supervised Multilingual Causality Extraction from <fixed-case>W</fixed-case>ikipedia ChikaraHashimoto - 2986–2997 + 2988–2999 We present a method for extracting causality knowledge from Wikipedia, such as Protectionism -> Trade war, where the cause and effect entities correspond to Wikipedia articles. Such causality knowledge is easy to verify by reading corresponding Wikipedia articles, to translate to multiple languages through Wikidata, and to connect to knowledge bases derived from Wikipedia. Our method exploits Wikipedia article sections that describe causality and the redundancy stemming from the multilinguality of Wikipedia. Experiments showed that our method achieved precision and recall above 98% and 64%, respectively. In particular, it could extract causalities whose cause and effect were written distantly in a Wikipedia article. We have released the code and data for further research. D19-1296 10.18653/v1/D19-1296 @@ -3478,7 +3478,7 @@ XuepengWang DaweiYin ChengqingZong - 2998–3008 + 3000–3010 Review summarization aims to generate a condensed summary for a review or multiple reviews. Existing review summarization systems mainly generate summary only based on review content and neglect the authors’ attributes (e.g., gender, age, and occupation). In fact, when summarizing a review, users with different attributes usually pay attention to specific aspects and have their own word-using habits or writing styles. Therefore, we propose an Attribute-aware Sequence Network (ASN) to take the aforementioned users’ characteristics into account, which includes three modules: an attribute encoder encodes the attribute preferences over the words; an attribute-aware review encoder adopts an attribute-based selective mechanism to select the important information of a review; and an attribute-aware summary decoder incorporates attribute embedding and attribute-specific word-using habits into word prediction. To validate our model, we collect a new dataset TripAtt, comprising 495,440 attribute-review-summary triplets with three kinds of attribute information: gender, age, and travel status. Extensive experiments show that ASN achieves state-of-the-art performance on review summarization in both auto-metric ROUGE and human evaluation. D19-1297 10.18653/v1/D19-1297 @@ -3487,7 +3487,7 @@ Extractive Summarization of Long Documents by Combining Global and Local Context WenXiao GiuseppeCarenini - 3009–3019 + 3011–3021 In this paper, we propose a novel neural single-document extractive summarization model for long documents, incorporating both the global context of the whole document and the local context within the current topic. We evaluate the model on two datasets of scientific papers , Pubmed and arXiv, where it outperforms previous work, both extractive and abstractive models, on ROUGE-1, ROUGE-2 and METEOR scores. We also show that, consistently with our goal, the benefits of our method become stronger as we apply it to longer documents. Rather surprisingly, an ablation study indicates that the benefits of our model seem to come exclusively from modeling the local context, even for the longest documents. D19-1298 D19-1298.Attachment.zip @@ -3501,7 +3501,7 @@ FengJiang BingQin Chin-YewLin - 3020–3030 + 3022–3032 Recent neural models for data-to-text generation rely on massive parallel pairs of data and text to learn the writing knowledge. They often assume that writing knowledge can be acquired from the training data alone. However, when people are writing, they not only rely on the data but also consider related knowledge. In this paper, we enhance neural data-to-text models with external knowledge in a simple but effective way to improve the fidelity of generated text. Besides relying on parallel data and text as in previous work, our model attends to relevant external knowledge, encoded as a temporary memory, and combines this knowledge with the context representation of data before generating words. This allows the model to infer relevant facts which are not explicitly stated in the data table from an external knowledge source. Experimental results on twenty-one Wikipedia infobox-to-text datasets show our model, KBAtt, consistently improves a state-of-the-art model on most of the datasets. In addition, to quantify when and why external knowledge is effective, we design a metric, KBGain, which shows a strong correlation with the observed performance boost. This result demonstrates the relevance of external knowledge and sparseness of original data are the main factors affecting system performance. D19-1299 10.18653/v1/D19-1299 @@ -3514,7 +3514,7 @@ FeiyangPan MinYang QingHe - 3031–3041 + 3033–3043 In this work, we re-examine the problem of extractive text summarization for long documents. We observe that the process of extracting summarization of human can be divided into two stages: 1) a rough reading stage to look for sketched information, and 2) a subsequent careful reading stage to select key sentences to form the summary. By simulating such a two-stage process, we propose a novel approach for extractive summarization. We formulate the problem as a contextual-bandit problem and solve it with policy gradient. We adopt a convolutional neural network to encode gist of paragraphs for rough reading, and a decision making policy with an adapted termination mechanism for careful reading. Experiments on the CNN and DailyMail datasets show that our proposed method can provide high-quality summaries with varied length, and significantly outperform the state-of-the-art extractive methods in terms of ROUGE metrics. D19-1300 D19-1300.Attachment.zip @@ -3528,7 +3528,7 @@ MinZhang WeihuaLuo YueZhang - 3042–3051 + 3044–3053 We propose a contrastive attention mechanism to extend the sequence-to-sequence framework for abstractive sentence summarization task, which aims to generate a brief summary of a given source sentence. The proposed contrastive attention mechanism accommodates two categories of attention: one is the conventional attention that attends to relevant parts of the source sentence, the other is the opponent attention that attends to irrelevant or less relevant parts of the source sentence. Both attentions are trained in an opposite way so that the contribution from the conventional attention is encouraged and the contribution from the opponent attention is discouraged through a novel softmax and softmin functionality. Experiments on benchmark datasets show that, the proposed contrastive attention mechanism is more focused on the relevant parts for the summary than the conventional attention mechanism, and greatly advances the state-of-the-art performance on the abstractive sentence summarization task. We release the code at https://github.com/travel-go/ Abstractive-Text-Summarization. D19-1301 10.18653/v1/D19-1301 @@ -3542,7 +3542,7 @@ JiajunZhang ShaonanWang ChengqingZong - 3052–3062 + 3054–3064 Cross-lingual summarization (CLS) is the task to produce a summary in one particular language for a source document in a different language. Existing methods simply divide this task into two steps: summarization and translation, leading to the problem of error propagation. To handle that, we present an end-to-end CLS framework, which we refer to as Neural Cross-Lingual Summarization (NCLS), for the first time. Moreover, we propose to further improve NCLS by incorporating two related tasks, monolingual summarization and machine translation, into the training process of CLS under multi-task learning. Due to the lack of supervised CLS data, we propose a round-trip translation strategy to acquire two high-quality large-scale CLS datasets based on existing monolingual summarization datasets. Experimental results have shown that our NCLS achieves remarkable improvement over traditional pipeline methods on both English-to-Chinese and Chinese-to-English CLS human-corrected test sets. In addition, NCLS with multi-task learning can further significantly improve the quality of generated summaries. We make our dataset and code publicly available here: http://www.nlpr.ia.ac.cn/cip/dataset.htm. D19-1302 10.18653/v1/D19-1302 @@ -3553,7 +3553,7 @@ Chien-ShengWu AndreaMadotto PascaleFung - 3063–3073 + 3065–3075 Sensational headlines are headlines that capture people’s attention and generate reader interest. Conventional abstractive headline generation methods, unlike human writers, do not optimize for maximal reader attention. In this paper, we propose a model that generates sensational headlines without labeled data. We first train a sensationalism scorer by classifying online headlines with many comments (“clickbait”) against a baseline of headlines generated from a summarization model. The score from the sensationalism scorer is used as the reward for a reinforcement learner. However, maximizing the noisy sensationalism reward will generate unnatural phrases instead of sensational headlines. To effectively leverage this noisy reward, we propose a novel loss function, Auto-tuned Reinforcement Learning (ARL), to dynamically balance reinforcement learning (RL) with maximum likelihood estimation (MLE). Human evaluation shows that 60.8% of samples generated by our model are sensational, which is significantly better than the Pointer-Gen baseline and other RL models. D19-1303 10.18653/v1/D19-1303 @@ -3564,7 +3564,7 @@ YangGao HeyanHuang YuxiangZhou - 3074–3083 + 3076–3085 A quality abstractive summary should not only copy salient source texts as summaries but should also tend to generate new conceptual words to express concrete details. Inspired by the popular pointer generator sequence-to-sequence model, this paper presents a concept pointer network for improving these aspects of abstractive summarization. The network leverages knowledge-based, context-aware conceptualizations to derive an extended set of candidate concepts. The model then points to the most appropriate choice using both the concept set and original source text. This joint approach generates abstractive summaries with higher-level semantic concepts. The training model is also optimized in a way that adapts to different data, which is based on a novel method of distant-supervised learning guided by reference summaries and testing set. Overall, the proposed approach provides statistically significant improvements over several state-of-the-art models on both the DUC-2004 and Gigaword datasets. A human evaluation of the model’s abstractive abilities also supports the quality of the summaries produced within this framework. D19-1304 D19-1304.Attachment.zip @@ -3574,7 +3574,7 @@ Surface Realisation Using Full Delexicalisation AnastasiaShimorina ClaireGardent - 3084–3094 + 3086–3096 Surface realisation (SR) maps a meaning representation to a sentence and can be viewed as consisting of three subtasks: word ordering, morphological inflection and contraction generation (e.g., clitic attachment in Portuguese or elision in French). We propose a modular approach to surface realisation which models each of these components separately, and evaluate our approach on the 10 languages covered by the SR’18 Surface Realisation Shared Task shallow track. We provide a detailed evaluation of how word order, morphological realisation and contractions are handled by the model and an analysis of the differences in word ordering performance across languages. D19-1305 D19-1305.Attachment.zip @@ -3587,7 +3587,7 @@ JonasMueller NicholasMatthews EnricoSantus - 3095–3107 + 3097–3109 Text attribute transfer aims to automatically rewrite sentences such that they possess certain linguistic attributes, while simultaneously preserving their semantic content. This task remains challenging due to a lack of supervised parallel data. Existing approaches try to explicitly disentangle content and attribute information, but this is difficult and often results in poor content-preservation and ungrammaticality. In contrast, we propose a simpler approach, Iterative Matching and Translation (IMaT), which: (1) constructs a pseudo-parallel corpus by aligning a subset of semantically similar sentences from the source and the target corpora; (2) applies a standard sequence-to-sequence model to learn the attribute transfer; (3) iteratively improves the learned transfer function by refining imperfections in the alignment. In sentiment modification and formality transfer tasks, our method outperforms complex state-of-the-art systems by a large margin. As an auxiliary contribution, we produce a publicly-available test set with human-generated transfer references. D19-1306 10.18653/v1/D19-1306 @@ -3600,7 +3600,7 @@ OriShapira IdoDagan IrynaGurevych - 3108–3118 + 3110–3120 Reinforcement Learning (RL)based document summarisation systems yield state-of-the-art performance in terms of ROUGE scores, because they directly use ROUGE as the rewards during training. However, summaries with high ROUGE scores often receive low human judgement. To find a better reward function that can guide RL to generate human-appealing summaries, we learn a reward function from human ratings on 2,500 summaries. Our reward function only takes the document and system summary as input. Hence, once trained, it can be used to train RL based summarisation systems without using any reference summaries. We show that our learned rewards have significantly higher correlation with human ratings than previous approaches. Human evaluation experiments show that, compared to the state-of-the-art supervised-learning systems and ROUGE-as-rewards RL summarisation systems, the RL systems using our learned rewards during training generate summaries with higher human ratings. The learned reward function and our source code are available at https://github.com/yg211/summary-reward-no-reference. D19-1307 D19-1307.Attachment.pdf @@ -3611,7 +3611,7 @@ JaeminCho MinjoonSeo HannanehHajishirzi - 3119–3129 + 3121–3131 Generating diverse sequences is important in many NLP applications such as question generation or summarization that exhibit semantically one-to-many relationships between source and the target sequences. We present a method to explicitly separate diversification from generation using a general plug-and-play module (called SELECTOR) that wraps around and guides an existing encoder-decoder model. The diversification stage uses a mixture of experts to sample different binary masks on the source sequence for diverse content selection. The generation stage uses a standard encoder-decoder model given each selected content from the source sequence. Due to the non-differentiable nature of discrete sampling and the lack of ground truth labels for binary mask, we leverage a proxy for ground truth mask and adopt stochastic hard-EM for training. In question generation (SQuAD) and abstractive summarization (CNN-DM), our method demonstrates significant improvements in accuracy, diversity and training efficiency, including state-of-the-art top-1 accuracy in both datasets, 6% gain in top-5 accuracy, and 3.7 times faster training over a state-of-the-art model. Our code is publicly available at https://github.com/clovaai/FocusSeq2Seq. D19-1308 10.18653/v1/D19-1308 @@ -3625,7 +3625,7 @@ WenlinWang GuoyinWang LawrenceCarin - 3130–3140 + 3132–3142 Generating high-quality paraphrases is a fundamental yet challenging natural language processing task. Despite the effectiveness of previous work based on generative models, there remain problems with exposure bias in recurrent neural networks, and often a failure to generate realistic sentences. To overcome these challenges, we propose the first end-to-end conditional generative architecture for generating paraphrases via adversarial training, which does not depend on extra linguistic information. Extensive experiments on four public datasets demonstrate the proposed method achieves state-of-the-art results, outperforming previous generative architectures on both automatic metrics (BLEU, METEOR, and TER) and human evaluations. D19-1309 10.18653/v1/D19-1309 @@ -3636,7 +3636,7 @@ XiaochengFeng BingQin TingLiu - 3141–3150 + 3143–3152 Although Seq2Seq models for table-to-text generation have achieved remarkable progress, modeling table representation in one dimension is inadequate. This is because (1) the table consists of multiple rows and columns, which means that encoding a table should not depend only on one dimensional sequence or set of records and (2) most of the tables are time series data (e.g. NBA game data, stock market data), which means that the description of the current table may be affected by its historical data. To address aforementioned problems, not only do we model each table cell considering other records in the same row, we also enrich table’s representation by modeling each table cell in context of other cells in the same column or with historical (time dimension) data respectively. In addition, we develop a table cell fusion gate to combine representations from row, column and time dimension into one dense vector according to the saliency of each dimension’s representation. We evaluated our methods on ROTOWIRE, a benchmark dataset of NBA basketball games. Both automatic and human evaluation results demonstrate the effectiveness of our model with improvement of 2.66 in BLEU over the strong baseline and outperformance of state-of-the-art model. D19-1310 10.18653/v1/D19-1310 @@ -3647,7 +3647,7 @@ AixinSun JingLi KarthikMuthuswamy - 3151–3160 + 3153–3162 In multi-document summarization, a set of documents to be summarized is assumed to be on the same topic, known as the underlying topic in this paper. That is, the underlying topic can be collectively represented by all the documents in the set. Meanwhile, different documents may cover various different subtopics and the same subtopic can be across several documents. Inspired by topic model, the underlying topic of a document set can also be viewed as a collection of different subtopics of different importance. In this paper, we propose a summarization model called STDS. The model generates the underlying topic representation from both document view and subtopic view in parallel. The learning objective is to minimize the distance between the representations learned from the two views. The contextual information is encoded through a hierarchical RNN architecture. Sentence salience is estimated in a hierarchical way with subtopic salience and relative sentence salience, by considering the contextual information. Top ranked sentences are then extracted as a summary. Note that the notion of subtopic enables us to bring in additional information (e.g. comments to news articles) that is helpful for document summarization. Experimental results show that the proposed solution outperforms state-of-the-art methods on benchmark datasets. D19-1311 10.18653/v1/D19-1311 @@ -3656,7 +3656,7 @@ Referring Expression Generation Using Entity Profiles MengCao Jackie Chi KitCheung - 3161–3170 + 3163–3172 Referring Expression Generation (REG) is the task of generating contextually appropriate references to entities. A limitation of existing REG systems is that they rely on entity-specific supervised training, which means that they cannot handle entities not seen during training. In this study, we address this in two ways. First, we propose task setups in which we specifically test a REG system’s ability to generalize to entities not seen during training. Second, we propose a profile-based deep neural network model, ProfileREG, which encodes both the local context and an external profile of the entity to generate reference realizations. Our model generates tokens by learning to choose between generating pronouns, generating from a fixed vocabulary, or copying a word from the profile. We evaluate our model on three different splits of the WebNLG dataset, and show that it outperforms competitive baselines in all settings according to automatic and human evaluations. D19-1312 10.18653/v1/D19-1312 @@ -3668,7 +3668,7 @@ WeinanZhang XinJiang YongYu - 3171–3180 + 3173–3182 Paraphrasing plays an important role in various natural language processing (NLP) tasks, such as question answering, information retrieval and sentence simplification. Recently, neural generative models have shown promising results in paraphrase generation. However, prior work mainly focused on single paraphrase generation, while ignoring the fact that diversity is essential for enhancing generalization capability and robustness of downstream applications. Few works have been done to solve diverse paraphrase generation. In this paper, we propose a novel approach with two discriminators and multiple generators to generate a variety of different paraphrases. A reinforcement learning algorithm is applied to train our model. Our experiments on two real-world datasets demonstrate that our model not only gains a significant increase in diversity but also improves generation quality over several state-of-the-art baselines. D19-1313 10.18653/v1/D19-1313 @@ -3678,7 +3678,7 @@ Leonardo F. R.Ribeiro ClaireGardent IrynaGurevych - 3181–3192 + 3183–3194 Generating text from graph-based data, such as Abstract Meaning Representation (AMR), is a challenging task due to the inherent difficulty in how to properly encode the structure of a graph with labeled edges. To address this difficulty, we propose a novel graph-to-sequence model that encodes different but complementary perspectives of the structural information contained in the AMR graph. The model learns parallel top-down and bottom-up representations of nodes capturing contrasting views of the graph. We also investigate the use of different node message passing strategies, employing different state-of-the-art graph encoders to compute node representations based on incoming and outgoing perspectives. In our experiments, we demonstrate that the dual graph representation leads to improvements in AMR-to-text generation, achieving state-of-the-art results on two AMR datasets D19-1314 D19-1314.Attachment.zip @@ -3692,7 +3692,7 @@ TomokiTaniguchi YasuhideMiura TomokoOhkuma - 3193–3203 + 3195–3205 The automated generation of information indicating the characteristics of articles such as headlines, key phrases, summaries and categories helps writers to alleviate their workload. Previous research has tackled these tasks using neural abstractive summarization and classification methods. However, the outputs may be inconsistent if they are generated individually. The purpose of our study is to generate multiple outputs consistently. We introduce a multi-task learning model with a shared encoder and multiple decoders for each task. We propose a novel loss function called hierarchical consistency loss to maintain consistency among the attention weights of the decoders. To evaluate the consistency, we employ a human evaluation. The results show that our model generates more consistent headlines, key phrases and categories. In addition, our model outperforms the baseline model on the ROUGE scores, and generates more adequate and fluent headlines. D19-1315 D19-1315.Attachment.pdf @@ -3701,7 +3701,7 @@ Toward a Task of Feedback Comment Generation for Writing Learning RyoNagata - 3204–3213 + 3206–3215 In this paper, we introduce a novel task called feedback comment generation — a task of automatically generating feedback comments such as a hint or an explanatory note for writing learning for non-native learners of English. There has been almost no work on this task nor corpus annotated with feedback comments. We have taken the first step by creating learner corpora consisting of approximately 1,900 essays where all preposition errors are manually annotated with feedback comments. We have tested three baseline methods on the dataset, showing that a simple neural retrieval-based method sets a baseline performance with an F-measure of 0.34 to 0.41. Finally, we have looked into the results to explore what modifications we need to make to achieve better performance. We also have explored problems unaddressed in this work D19-1316 10.18653/v1/D19-1316 @@ -3713,7 +3713,7 @@ LidongBing IrwinKing Michael R.Lyu - 3214–3224 + 3216–3226 Question generation (QG) is the task of generating a question from a reference sentence and a specified answer within the sentence. A major challenge in QG is to identify answer-relevant context words to finish the declarative-to-interrogative sentence transformation. Existing sequence-to-sequence neural models achieve this goal by proximity-based answer position encoding under the intuition that neighboring words of answers are of high possibility to be answer-relevant. However, such intuition may not apply to all cases especially for sentences with complex answer-relevant relations. Consequently, the performance of these models drops sharply when the relative distance between the answer fragment and other non-stop sentence words that also appear in the ground truth question increases. To address this issue, we propose a method to jointly model the unstructured sentence and the structured answer-relevant relation (extracted from the sentence in advance) for question generation. Specifically, the structured answer-relevant relation acts as the to the point context and it thus naturally helps keep the generated question to the point, while the unstructured sentence provides the full information. Extensive experiments show that to the point context helps our question generation model achieve significant improvements on several automatic evaluation metrics. Furthermore, our model is capable of generating diverse questions for a sentence which conveys multiple relations of its answer fragment. D19-1317 10.18653/v1/D19-1317 @@ -3723,7 +3723,7 @@ JuliaIve PranavaMadhyastha LuciaSpecia - 3225–3234 + 3227–3236 Most text-to-text generation tasks, for example text summarisation and text simplification, require copying words from the input to the output. We introduce Copycat, a transformer-based pointer network for such tasks which obtains competitive results in abstractive text summarisation and generates more abstractive summaries. We propose a further extension of this architecture for automatic post-editing, where generation is conditioned over two inputs (source language and machine translation), and the model is capable of deciding where to copy information from. This approach achieves competitive performance when compared to state-of-the-art automated post-editing systems. More importantly, we show that it addresses a well-known limitation of automatic post-editing - overcorrecting translations - and that our novel mechanism for copying source language words improves the results. D19-1318 10.18653/v1/D19-1318 @@ -3732,7 +3732,7 @@ Towards Controllable and Personalized Review Generation PanLi AlexanderTuzhilin - 3235–3243 + 3237–3245 In this paper, we propose a novel model RevGAN that automatically generates controllable and personalized user reviews based on the arbitrarily given sentimental and stylistic information. RevGAN utilizes the combination of three novel components, including self-attentive recursive autoencoders, conditional discriminators, and personalized decoders. We test its performance on the several real-world datasets, where our model significantly outperforms state-of-the-art generation models in terms of sentence quality, coherence, personalization, and human evaluations. We also empirically show that the generated reviews could not be easily distinguished from the organically produced reviews and that they follow the same statistical linguistics laws. D19-1319 10.18653/v1/D19-1319 @@ -3743,7 +3743,7 @@ SylvainLamprier BenjaminPiwowarski JacopoStaiano - 3244–3254 + 3246–3256 Abstractive summarization approaches based on Reinforcement Learning (RL) have recently been proposed to overcome classical likelihood maximization. RL enables to consider complex, possibly non differentiable, metrics that globally assess the quality and relevance of the generated outputs. ROUGE, the most used summarization metric, is known to suffer from bias towards lexical similarity as well as from sub-optimal accounting for fluency and readability of the generated abstracts. We thus explore and propose alternative evaluation measures: the reported human-evaluation analysis shows that the proposed metrics, based on Question Answering, favorably compare to ROUGE – with the additional property of not requiring reference summaries. Training a RL-based model on these metrics leads to improvements (both in terms of human or automated metrics) over current approaches that use ROUGE as reward. D19-1320 10.18653/v1/D19-1320 @@ -3755,7 +3755,7 @@ JiangtaoWen WenfeiXu XiaoyanZhu - 3255–3266 + 3257–3268 Existing neural methods for data-to-text generation are still struggling to produce long and diverse texts: they are insufficient to model input data dynamically during generation, to capture inter-sentence coherence, or to generate diversified expressions. To address these issues, we propose a Planning-based Hierarchical Variational Model (PHVM). Our model first plans a sequence of groups (each group is a subset of input items to be covered by a sentence) and then realizes each sentence conditioned on the planning result and the previously generated context, thereby decomposing long text generation into dependent sentence generation sub-tasks. To capture expression diversity, we devise a hierarchical latent structure where a global planning latent variable models the diversity of reasonable planning and a sequence of local latent variables controls sentence realization. Experiments show that our model outperforms state-of-the-art baselines in long and diverse text generation. D19-1321 D19-1321.Attachment.zip @@ -3766,7 +3766,7 @@ AkhileshSudhakar BhargavUpadhyay ArjunMaheswaran - 3267–3277 + 3269–3279 Text style transfer is the task of transferring the style of text having certain stylistic attributes, while preserving non-stylistic or content information. In this work we introduce the Generative Style Transformer (GST) - a new approach to rewriting sentences to a target style in the absence of parallel style corpora. GST leverages the power of both, large unsupervised pre-trained language models as well as the Transformer. GST is a part of a larger ‘Delete Retrieve Generate’ framework, in which we also propose a novel method of deleting style attributes from the source sentence by exploiting the inner workings of the Transformer. Our models outperform state-of-art systems across 5 datasets on sentiment, gender and political slant transfer. We also propose the use of the GLEU metric as an automatic metric of evaluation of style transfer, which we found to compare better with human ratings than the predominantly used BLEU score. D19-1322 10.18653/v1/D19-1322 @@ -3777,7 +3777,7 @@ LuyangHuang ZheHu LuWang - 3278–3289 + 3280–3291 Abstractive summarization systems aim to produce more coherent and concise summaries than their extractive counterparts. Popular neural models have achieved impressive results for single-document summarization, yet their outputs are often incoherent and unfaithful to the input. In this paper, we introduce SENECA, a novel System for ENtity-drivEn Coherent Abstractive summarization framework that leverages entity information to generate informative and coherent abstracts. Our framework takes a two-step approach: (1) an entity-aware content selection module first identifies salient sentences from the input, then (2) an abstract generation module conducts cross-sentence information compression and abstraction to generate the final summary, which is trained with rewards to promote coherence, conciseness, and clarity. The two components are further connected using reinforcement learning. Automatic evaluation shows that our model significantly outperforms previous state-of-the-art based on ROUGE and our proposed coherence measures on New York Times and CNN/Daily Mail datasets. Human judges further rate our system summaries as more informative and coherent than those by popular summarization models. D19-1323 D19-1323.Attachment.zip @@ -3787,7 +3787,7 @@ Neural Extractive Text Summarization with Syntactic Compression JiachengXu GregDurrett - 3290–3301 + 3292–3303 Recent neural network approaches to summarization are largely either selection-based extraction or generation-based abstraction. In this work, we present a neural model for single-document summarization based on joint extraction and syntactic compression. Our model chooses sentences from the document, identifies possible compressions based on constituency parses, and scores those compressions with a neural model to produce the final summary. For learning, we construct oracle extractive-compressive summaries, then learn both of our components jointly with this supervision. Experimental results on the CNN/Daily Mail and New York Times datasets show that our model achieves strong performance (comparable to state-of-the-art systems) as evaluated by ROUGE. Moreover, our approach outperforms an off-the-shelf compression module, and human and manual evaluation shows that our model’s output generally remains grammatical. D19-1324 D19-1324.Attachment.pdf @@ -3802,7 +3802,7 @@ ChrisBrockett BillDolan Ming-TingSun - 3302–3311 + 3304–3313 Text style transfer without parallel data has achieved some practical success. However, in the scenario where less data is available, these methods may yield poor performance. In this paper, we examine domain adaptation for text style transfer to leverage massively available data from other domains. These data may demonstrate domain shift, which impedes the benefits of utilizing such data for training. To address this challenge, we propose simple yet effective domain adaptive text style transfer models, enabling domain-adaptive information exchange. The proposed models presumably learn from the source domain to: (i) distinguish stylized information and generic content information; (ii) maximally preserve content information; and (iii) adaptively transfer the styles in a domain-aware manner. We evaluate the proposed models on two style transfer tasks (sentiment and formality) over multiple target domains where only limited non-parallel data is available. Extensive experiments demonstrate the effectiveness of the proposed model compared to the baselines. D19-1325 D19-1325.Attachment.pdf @@ -3815,8 +3815,8 @@ Mitesh M.Khapra Balaji VasanSrinivasan BalaramanRavindran - 3312–3321 - In this work, we focus on the task of Automatic Question Generation (AQG) where given a passage and an answer the task is to generate the corresponding question. It is desired that the generated question should be (i) grammatically correct (ii) answerable from the passage and (iii) specific to the given answer. An analysis of existing AQG models shows that they produce questions which do not adhere to one or more of the above-mentioned qualities. In particular, the generated questions look like an incomplete draft of the desired question with a clear scope for refinement. To alleviate this shortcoming, we propose a method which tries to mimic the human process of generating questions by first creating an initial draft and then refining it. More specifically, we propose Refine Network (RefNet) which contains two decoders. The second decoder uses a dual attention network which pays attention to both (i) the original passage and (ii) the question (initial draft) generated by the first decoder. In effect, it refines the question generated by the first decoder, thereby making it more correct and complete. We evaluate RefNet on three datasets, viz., SQuAD, HOTPOT-QA, and DROP, and show that it outperforms existing state-of-the-art methods by 7-16% on all of these datasets. Lastly, we show that we can improve the quality of the second decoder on specific metrics, such as, fluency and answerability by explicitly rewarding revisions that improve on the corresponding metric during training. The code has been made publicly available https://github.com/PrekshaNema25/RefNet-QG. + 3314–3323 + In this work, we focus on the task of Automatic Question Generation (AQG) where given a passage and an answer the task is to generate the corresponding question. It is desired that the generated question should be (i) grammatically correct (ii) answerable from the passage and (iii) specific to the given answer. An analysis of existing AQG models shows that they produce questions which do not adhere to one or more of the above-mentioned qualities. In particular, the generated questions look like an incomplete draft of the desired question with a clear scope for refinement. To alleviate this shortcoming, we propose a method which tries to mimic the human process of generating questions by first creating an initial draft and then refining it. More specifically, we propose Refine Network (RefNet) which contains two decoders. The second decoder uses a dual attention network which pays attention to both (i) the original passage and (ii) the question (initial draft) generated by the first decoder. In effect, it refines the question generated by the first decoder, thereby making it more correct and complete. We evaluate RefNet on three datasets, viz., SQuAD, HOTPOT-QA, and DROP, and show that it outperforms existing state-of-the-art methods by 7-16% on all of these datasets. Lastly, we show that we can improve the quality of the second decoder on specific metrics, such as, fluency and answerability by explicitly rewarding revisions that improve on the corresponding metric during training. The code has been made publicly available . D19-1326 D19-1326.Attachment.pdf 10.18653/v1/D19-1326 @@ -3827,7 +3827,7 @@ DongyeopKang LucasMentch EduardHovy - 3322–3333 + 3324–3335 Despite the recent developments on neural summarization systems, the underlying logic behind the improvements from the systems and its corpus-dependency remains largely unexplored. Position of sentences in the original text, for example, is a well known bias for news summarization. Following in the spirit of the claim that summarization is a combination of sub-functions, we define three sub-aspects of summarization: position, importance, and diversity and conduct an extensive analysis of the biases of each sub-aspect with respect to the domain of nine different summarization corpora (e.g., news, academic papers, meeting minutes, movie script, books, posts). We find that while position exhibits substantial bias in news articles, this is not the case, for example, with academic papers and meeting minutes. Furthermore, our empirical study shows that different types of summarization systems (e.g., neural-based) are composed of different degrees of the sub-aspects. Our study provides useful lessons regarding consideration of underlying sub-aspects when collecting a new summarization dataset or developing a new system. D19-1327 D19-1327.Attachment.pdf @@ -3838,7 +3838,7 @@ YovaKementchedjhieva MareikeHartmann AndersSøgaard - 3334–3339 + 3336–3341 The task of bilingual dictionary induction (BDI) is commonly used for intrinsic evaluation of cross-lingual word embeddings. The largest dataset for BDI was generated automatically, so its quality is dubious. We study the composition and quality of the test sets for five diverse languages from this dataset, with concerning findings: (1) a quarter of the data consists of proper nouns, which can be hardly indicative of BDI performance, and (2) there are pervasive gaps in the gold-standard targets. These issues appear to affect the ranking between cross-lingual embedding systems on individual languages, and the overall degree to which the systems differ in performance. With proper nouns removed from the data, the margin between the top two systems included in the study grows from 3.4% to 17.2%. Manual verification of the predictions, on the other hand, reveals that gaps in the gold standard targets artificially inflate the margin between the two systems on English to Bulgarian BDI from 0.1% to 6.7%. We thus suggest that future research either avoids drawing conclusions from quantitative results on this BDI dataset, or accompanies such evaluation with rigorous error analysis. D19-1328 D19-1328.Attachment.zip @@ -3849,7 +3849,7 @@ KatharinaKann KyunghyunCho Samuel R.Bowman - 3340–3347 + 3342–3349 Development sets are impractical to obtain for real low-resource languages, since using all available data for training is often more effective. However, development sets are widely used in research papers that purport to deal with low-resource natural language processing (NLP). Here, we aim to answer the following questions: Does using a development set for early stopping in the low-resource setting influence results as compared to a more realistic alternative, where the number of training epochs is tuned on development languages? And does it lead to overestimation or underestimation of performance? We repeat multiple experiments from recent work on neural models for low-resource NLP and compare results for models obtained by training with and without development sets. On average over languages, absolute accuracy differs by up to 1.4%. However, for some languages and tasks, differences are as big as 18.0% accuracy. Our results highlight the importance of realistic experimental setups in the publication of low-resource NLP research results. D19-1329 D19-1329.Attachment.zip @@ -3862,7 +3862,7 @@ LongZhou YuchenLiu ChengqingZong - 3348–3353 + 3350–3355 In this paper, we introduce a novel interactive approach to translate a source language into two different languages simultaneously and interactively. Specifically, the generation of one language relies on not only previously generated outputs by itself, but also the outputs predicted in the other language. Experimental results on IWSLT and WMT datasets demonstrate that our method can obtain significant improvements over both conventional Neural Machine Translation (NMT) model and multilingual NMT model. D19-1330 10.18653/v1/D19-1330 @@ -3871,7 +3871,7 @@ On <fixed-case>NMT</fixed-case> Search Errors and Model Errors: Cat Got Your Tongue? FelixStahlberg BillByrne - 3354–3360 + 3356–3362 We report on search errors and model errors in neural machine translation (NMT). We present an exact inference procedure for neural sequence models based on a combination of beam search and depth-first search. We use our exact search to find the global best model scores under a Transformer base model for the entire WMT15 English-German test set. Surprisingly, beam search fails to find these global best model scores in most cases, even with a very large beam size of 100. For more than 50% of the sentences, the model in fact assigns its global best score to the empty translation, revealing a massive failure of neural models in properly accounting for adequacy. We show by constraining search with a minimum translation length that at the root of the problem of empty translations lies an inherent bias towards shorter translations. We conclude that vanilla NMT in its current form requires just the right amount of beam search errors, which, from a modelling perspective, is a highly unsatisfactory conclusion indeed, as the model often prefers an empty translation. D19-1331 10.18653/v1/D19-1331 @@ -3882,7 +3882,7 @@ DanielKhashabi QiangNing DanRoth - 3361–3367 + 3363–3369 Understanding time is crucial for understanding events expressed in natural language. Because people rarely say the obvious, it is often necessary to have commonsense knowledge about various temporal aspects of events, such as duration, frequency, and temporal order. However, this important problem has so far received limited attention. This paper systematically studies this temporal commonsense problem. Specifically, we define five classes of temporal commonsense, and use crowdsourcing to develop a new dataset, MCTACO, that serves as a test set for this task. We find that the best current methods used on MCTACO are still far behind human performance, by about 20%, and discuss several directions for improvement. We hope that the new dataset and our study here can foster more future research on this topic. D19-1332 D19-1332.Attachment.zip @@ -3892,7 +3892,7 @@ <fixed-case>QAI</fixed-case>nfomax: Learning Robust Question Answering System by Mutual Information Maximization Yi-TingYeh Yun-NungChen - 3368–3373 + 3370–3375 Standard accuracy metrics indicate that modern reading comprehension systems have achieved strong performance in many question answering datasets. However, the extent these systems truly understand language remains unknown, and existing systems are not good at distinguishing distractor sentences which look related but do not answer the question. To address this problem, we propose QAInfomax as a regularizer in reading comprehension systems by maximizing mutual information among passages, a question, and its answer. QAInfomax helps regularize the model to not simply learn the superficial correlation for answering the questions. The experiments show that our proposed QAInfomax achieves the state-of-the-art performance on the benchmark Adversarial-SQuAD dataset. D19-1333 D19-1333.Attachment.zip @@ -3906,7 +3906,7 @@ LeiHou JuanziLi ZhiyuanLiu - 3374–3379 + 3376–3381 Multi-hop knowledge graph (KG) reasoning is an effective and explainable method for predicting the target entity via reasoning paths in query answering (QA) task. Most previous methods assume that every relation in KGs has enough triples for training, regardless of those few-shot relations which cannot provide sufficient triples for training robust reasoning models. In fact, the performance of existing multi-hop reasoning methods drops significantly on few-shot relations. In this paper, we propose a meta-based multi-hop reasoning method (Meta-KGR), which adopts meta-learning to learn effective meta parameters from high-frequency relations that could quickly adapt to few-shot relations. We evaluate Meta-KGR on two public datasets sampled from Freebase and NELL, and the experimental results show that Meta-KGR outperforms state-of-the-art methods in few-shot scenarios. In the future, our codes and datasets will also be available to provide more details. D19-1334 10.18653/v1/D19-1334 @@ -3918,7 +3918,7 @@ AdamTrischler KaheerSuleman Jackie Chi KitCheung - 3380–3385 + 3382–3387 Recent studies have significantly improved the state-of-the-art on common-sense reasoning (CSR) benchmarks like the Winograd Schema Challenge (WSC) and SWAG. The question we ask in this paper is whether improved performance on these benchmarks represents genuine progress towards common-sense-enabled systems. We make case studies of both benchmarks and design protocols that clarify and qualify the results of previous work by analyzing threats to the validity of previous experimental designs. Our protocols account for several properties prevalent in common-sense benchmarks including size limitations, structural regularities, and variable instance difficulty. D19-1335 D19-1335.Attachment.pdf @@ -3933,7 +3933,7 @@ BaobaoChang ZhifangSui XuSun - 3386–3391 + 3388–3393 In this paper, we focus on the task of generating a pun sentence given a pair of word senses. A major challenge for pun generation is the lack of large-scale pun corpus to guide supervised learning. To remedy this, we propose an adversarial generative network for pun generation (Pun-GAN). It consists of a generator to produce pun sentences, and a discriminator to distinguish between the generated pun sentences and the real sentences with specific word senses. The output of the discriminator is then used as a reward to train the generator via reinforcement learning, encouraging it to produce pun sentences which can support two word senses simultaneously. Experiments show that the proposed Pun-GAN can generate sentences that are more ambiguous and diverse in both automatic and human evaluation. D19-1336 10.18653/v1/D19-1336 @@ -3943,7 +3943,7 @@ WenjieZhou MinghuaZhang YunfangWu - 3392–3397 + 3394–3399 This paper explores the task of answer-aware questions generation. Based on the attention-based pointer generator model, we propose to incorporate an auxiliary task of language modeling to help question generation in a hierarchical multi-task learning structure. Our joint-learning model enables the encoder to learn a better representation of the input sequence, which will guide the decoder to generate more coherent and fluent questions. On both SQuAD and MARCO datasets, our multi-task learning model boosts the performance, achieving state-of-the-art results. Moreover, human evaluation further proves the high quality of our generated questions. D19-1337 10.18653/v1/D19-1337 @@ -3953,7 +3953,7 @@ FlorianSchmidt StephanMandt ThomasHofmann - 3398–3404 + 3400–3406 Autoregressive state transitions, where predictions are conditioned on past predictions, are the predominant choice for both deterministic and stochastic sequential models. However, autoregressive feedback exposes the evolution of the hidden state trajectory to potential biases from well-known train-test discrepancies. In this paper, we combine a latent state space model with a CRF observation model. We argue that such autoregressive observation models form an interesting middle ground that expresses local correlations on the word level but keeps the state evolution non-autoregressive. On unconditional sentence generation we show performance improvements compared to RNN and GAN baselines while avoiding some prototypical failure modes of autoregressive models. D19-1338 D19-1338.Attachment.pdf @@ -3965,7 +3965,7 @@ Kai-WeiChang PremkumarNatarajan NanyunPeng - 3405–3410 + 3407–3412 We present a systematic study of biases in natural language generation (NLG) by analyzing text generated from prompts that contain mentions of different demographic groups. In this work, we introduce the notion of the regard towards a demographic, use the varying levels of regard towards different demographics as a defining metric for bias in NLG, and analyze the extent to which sentiment scores are a relevant proxy metric for regard. To this end, we collect strategically-generated text from language models and manually annotate the text with both sentiment and regard scores. Additionally, we build an automatic regard classifier through transfer learning, so that we can analyze biases in unseen text. Together, these methods reveal the extent of the biased nature of language model generations. Our analysis provides a study of biases in NLG, bias metrics and correlated human judgments, and empirical evidence on the usefulness of our annotated dataset. D19-1339 D19-1339.Attachment.zip @@ -3977,7 +3977,7 @@ MithunPaul RebeccaSharp MihaiSurdeanu - 3411–3416 + 3413–3418 While neural networks produce state-of-the-art performance in many NLP tasks, they generally learn from lexical information, which may transfer poorly between domains. Here, we investigate the importance that a model assigns to various aspects of data while learning and making predictions, specifically, in a recognizing textual entailment (RTE) task. By inspecting the attention weights assigned by the model, we confirm that most of the weights are assigned to noun phrases. To mitigate this dependence on lexicalized information, we experiment with two strategies of masking. First, we replace named entities with their corresponding semantic tags along with a unique identifier to indicate lexical overlap between claim and evidence. Second, we similarly replace other word classes in the sentence (nouns, verbs, adjectives, and adverbs) with their super sense tags (Ciaramita and Johnson, 2003). Our results show that, while performance on the in-domain dataset remains on par with that of the model trained on fully lexicalized data, it improves considerably when tested out of domain. For example, the performance of a state-of-the-art RTE model trained on the masked Fake News Challenge (Pomerleau and Rao, 2017) data and evaluated on Fact Extraction and Verification (Thorne et al., 2018) data improved by over 10% in accuracy score compared to the fully lexicalized model. D19-1340 10.18653/v1/D19-1340 @@ -3990,7 +3990,7 @@ DanielRoberto Filizzola Ortiz EnricoSantus ReginaBarzilay - 3417–3423 + 3419–3425 Fact verification requires validating a claim in the context of evidence. We show, however, that in the popular FEVER dataset this might not necessarily be the case. Claim-only classifiers perform competitively with top evidence-aware models. In this paper, we investigate the cause of this phenomenon, identifying strong cues for predicting labels solely based on the claim, without considering any evidence. We create an evaluation set that avoids those idiosyncrasies. The performance of FEVER-trained models significantly drops when evaluated on this test set. Therefore, we introduce a regularization method which alleviates the effect of bias in the training data, obtaining improvements on the newly created test set. This work is a step towards a more sound evaluation of reasoning capabilities in fact verification models. D19-1341 D19-1341.Attachment.pdf @@ -4001,7 +4001,7 @@ XingweiTan YiCai ChangxiZhu - 3424–3429 + 3426–3431 Aspect-level sentiment classification, which is a fine-grained sentiment analysis task, has received lots of attention these years. There is a phenomenon that people express both positive and negative sentiments towards an aspect at the same time. Such opinions with conflicting sentiments, however, are ignored by existing studies, which design models based on the absence of them. We argue that the exclusion of conflict opinions is problematic, for the reason that it represents an important style of human thinking – dialectic thinking. If a real-world sentiment classification system ignores the existence of conflict opinions when it is designed, it will incorrectly mixed conflict opinions into other sentiment polarity categories in action. Existing models have problems when recognizing conflicting opinions, such as data sparsity. In this paper, we propose a multi-label classification model with dual attention mechanism to address these problems. D19-1342 D19-1342.Attachment.zip @@ -4013,7 +4013,7 @@ Liang-ChihYu K. RobertLai XuejieZhang - 3430–3435 + 3432–3437 Deep neural network models such as long short-term memory (LSTM) and tree-LSTM have been proven to be effective for sentiment analysis. However, sequential LSTM is a bias model wherein the words in the tail of a sentence are more heavily emphasized than those in the header for building sentence representations. Even tree-LSTM, with useful structural information, could not avoid the bias problem because the root node will be dominant and the nodes in the bottom of the parse tree will be less emphasized even though they may contain salient information. To overcome the bias problem, this study proposes a capsule tree-LSTM model, introducing a dynamic routing algorithm as an aggregation layer to build sentence representation by assigning different weights to nodes according to their contributions to prediction. Experiments on Stanford Sentiment Treebank (SST) for sentiment classification and EmoBank for regression show that the proposed method improved the performance of tree-LSTM and other neural network models. In addition, the deeper the tree structure, the bigger the improvement. D19-1343 10.18653/v1/D19-1343 @@ -4022,7 +4022,7 @@ A Label Informative Wide & Deep Classifier for Patents and Papers MuyaoNiu JieCai - 3436–3441 + 3438–3443 In this paper, we provide a simple and effective baseline for classifying both patents and papers to the well-established Cooperative Patent Classification (CPC). We propose a label-informative classifier based on the Wide & Deep structure, where the Wide part encodes string-level similarities between texts and labels, and the Deep part captures semantic-level similarities via non-linear transformations. Our model trains on millions of patents, and transfers to papers by developing distant-supervised training set and domain-specific features. Extensive experiments show that our model achieves comparable performance to the state-of-the-art model used in industry on both patents and papers. The output of this work should facilitate the searching, granting and filing of innovative ideas for patent examiners, attorneys and researchers. D19-1344 D19-1344.Attachment.zip @@ -4035,7 +4035,7 @@ SujianLi XiaodongZhang HoufengWang - 3442–3448 + 3444–3450 Recently, researches have explored the graph neural network (GNN) techniques on text classification, since GNN does well in handling complex structures and preserving global information. However, previous methods based on GNN are mainly faced with the practical problems of fixed corpus level graph structure which don’t support online testing and high memory consumption. To tackle the problems, we propose a new GNN based model that builds graphs for each input text with global parameters sharing instead of a single graph for the whole corpus. This method removes the burden of dependence between an individual text and entire corpus which support online testing, but still preserve global information. Besides, we build graphs by much smaller windows in the text, which not only extract more local features but also significantly reduce the edge numbers as well as memory consumption. Experiments show that our model outperforms existing models on several text classification datasets even with consuming less memory. D19-1345 10.18653/v1/D19-1345 @@ -4045,7 +4045,7 @@ AhmedSabir FrancescMoreno LluísPadró - 3449–3455 + 3451–3457 Applications such as textual entailment, plagiarism detection or document clustering rely on the notion of semantic similarity, and are usually approached with dimension reduction techniques like LDA or with embedding-based neural approaches. We present a scenario where semantic similarity is not enough, and we devise a neural approach to learn semantic relatedness. The scenario is text spotting in the wild, where a text in an image (e.g. street sign, advertisement or bus destination) must be identified and recognized. Our goal is to improve the performance of vision systems by leveraging semantic information. Our rationale is that the text to be spotted is often related to the image context in which it appears (word pairs such as Delta-airplane, or quarters-parking are not similar, but are clearly related). We show how learning a word-to-word or word-to-sentence relatedness score can improve the performance of text spotting systems up to 2.9 points, outperforming other measures in a benchmark dataset. D19-1346 10.18653/v1/D19-1346 @@ -4054,7 +4054,7 @@ Delta-training: Simple Semi-Supervised Text Classification using Pretrained Word Embeddings HwiyeolJo CeydaCinarel - 3456–3461 + 3458–3463 We propose a novel and simple method for semi-supervised text classification. The method stems from the hypothesis that a classifier with pretrained word embeddings always outperforms the same classifier with randomly initialized word embeddings, as empirically observed in NLP tasks. Our method first builds two sets of classifiers as a form of model ensemble, and then initializes their word embeddings differently: one using random, the other using pretrained word embeddings. We focus on different predictions between the two classifiers on unlabeled data while following the self-training framework. We also use early-stopping in meta-epoch to improve the performance of our method. Our method, Delta-training, outperforms the self-training and the co-training framework in 4 different text classification datasets, showing robustness against error accumulation. D19-1347 10.18653/v1/D19-1347 @@ -4063,7 +4063,7 @@ Visual Detection with Context for Document Layout Analysis CarlosSoto ShinjaeYoo - 3462–3468 + 3464–3470 We present 1) a work in progress method to visually segment key regions of scientific articles using an object detection technique augmented with contextual features, and 2) a novel dataset of region-labeled articles. A continuing challenge in scientific literature mining is the difficulty of consistently extracting high-quality text from formatted PDFs. To address this, we adapt the object-detection technique Faster R-CNN for document layout detection, incorporating contextual information that leverages the inherently localized nature of article contents to improve the region detection performance. Due to the limited availability of high-quality region-labels for scientific articles, we also contribute a novel dataset of region annotations, the first version of which covers 9 region classes and 822 article pages. Initial experimental results demonstrate a 23.9% absolute improvement in mean average precision over the baseline model by incorporating contextual features, and a processing speed 14x faster than a text-based technique. Ongoing work on further improvements is also discussed. D19-1348 D19-1348.Attachment.pdf @@ -4074,7 +4074,7 @@ LinziXing Michael J.Paul GiuseppeCarenini - 3469–3475 + 3471–3477 Probabilistic topic models such as latent Dirichlet allocation (LDA) are popularly used with Bayesian inference methods such as Gibbs sampling to learn posterior distributions over topic model parameters. We derive a novel measure of LDA topic quality using the variability of the posterior distributions. Compared to several existing baselines for automatic topic evaluation, the proposed metric achieves state-of-the-art correlations with human judgments of topic quality in experiments on three corpora. We additionally demonstrate that topic quality estimation can be further improved using a supervised estimator that combines multiple metrics. D19-1349 D19-1349.Attachment.zip @@ -4088,7 +4088,7 @@ YuZhou RuifengXu YulanHe - 3476–3481 + 3478–3483 In recent years, advances in neural variational inference have achieved many successes in text processing. Examples include neural topic models which are typically built upon variational autoencoder (VAE) with an objective of minimising the error of reconstructing original documents based on the learned latent topic vectors. However, minimising reconstruction errors does not necessarily lead to high quality topics. In this paper, we borrow the idea of reinforcement learning and incorporate topic coherence measures as reward signals to guide the learning of a VAE-based topic model. Furthermore, our proposed model is able to automatically separating background words dynamically from topic words, thus eliminating the pre-processing step of filtering infrequent and/or top frequent words, typically required for learning traditional topic models. Experimental results on the 20 Newsgroups and the NIPS datasets show superior performance both on perplexity and topic coherence measure compared to state-of-the-art neural topic models. D19-1350 10.18653/v1/D19-1350 @@ -4097,7 +4097,7 @@ Modelling Stopping Criteria for Search Results using <fixed-case>P</fixed-case>oisson Processes AlisonSneyd MarkStevenson - 3482–3487 + 3484–3489 Text retrieval systems often return large sets of documents, particularly when applied to large collections. Stopping criteria can reduce the number of these documents that need to be manually evaluated for relevance by predicting when a suitable level of recall has been achieved. In this work, a novel method for determining a stopping criterion is proposed that models the rate at which relevant documents occur using a Poisson process. This method allows a user to specify both a minimum desired level of recall to achieve and a desired probability of having achieved it. We evaluate our method on a public dataset and compare it with previous techniques for determining stopping criteria. D19-1351 10.18653/v1/D19-1351 @@ -4108,7 +4108,7 @@ WeiYang HaotianZhang JimmyLin - 3488–3494 + 3490–3496 This paper applies BERT to ad hoc document retrieval on news articles, which requires addressing two challenges: relevance judgments in existing test collections are typically provided only at the document level, and documents often exceed the length that BERT was designed to handle. Our solution is to aggregate sentence-level evidence to rank documents. Furthermore, we are able to leverage passage-level relevance judgments fortuitously available in other domains to fine-tune BERT models that are able to capture cross-domain notions of relevance, and can be directly used for ranking news articles. Our simple neural ranking models achieve state-of-the-art effectiveness on three standard test collections. D19-1352 10.18653/v1/D19-1352 @@ -4121,7 +4121,7 @@ PratikMehta W. BruceCroft ScottMiller - 3495–3500 + 3497–3502 When performing cross-language information retrieval (CLIR) for lower-resourced languages, a common approach is to retrieve over the output of machine translation (MT). However, there is no established guidance on how to optimize the resulting MT-IR system. In this paper, we examine the relationship between the performance of MT systems and both neural and term frequency-based IR models to identify how CLIR performance can be best predicted from MT quality. We explore performance at varying amounts of MT training data, byte pair encoding (BPE) merge operations, and across two IR collections and retrieval models. We find that the choice of IR collection can substantially affect the predictive power of MT tuning decisions and evaluation, potentially introducing dissociations between MT-only and overall CLIR performance. D19-1353 D19-1353.Attachment.zip @@ -4130,7 +4130,7 @@ Rotate King to get Queen: Word Relationships as Orthogonal Transformations in Embedding Space KawinEthayarajh - 3501–3506 + 3503–3508 A notable property of word embeddings is that word relationships can exist as linear substructures in the embedding space. For example, ‘gender’ corresponds to v_woman - v_man and v_queen - v_king. This, in turn, allows word analogies to be solved arithmetically: v_king - v_man + v_woman = v_queen. This property is notable because it suggests that models trained on word embeddings can easily learn such relationships as geometric translations. However, there is no evidence that models exclusively represent relationships in this manner. We document an alternative way in which downstream models might learn these relationships: orthogonal and linear transformations. For example, given a translation vector for ‘gender’, we can find an orthogonal matrix R, representing a rotation and reflection, such that R(v_king) = v_queen and R(v_man) = v_woman. Analogical reasoning using orthogonal transformations is almost as accurate as using vector arithmetic; using linear transformations is more accurate than both. Our findings suggest that these transformations can be as good a representation of word relationships as translation vectors. D19-1354 10.18653/v1/D19-1354 @@ -4141,7 +4141,7 @@ ChiSun XipengQiu XuanjingHuang - 3507–3512 + 3509–3514 Word Sense Disambiguation (WSD) aims to find the exact sense of an ambiguous word in a particular context. Traditional supervised methods rarely take into consideration the lexical resources like WordNet, which are widely utilized in knowledge-based methods. Recent studies have shown the effectiveness of incorporating gloss (sense definition) into neural networks for WSD. However, compared with traditional word expert supervised methods, they have not achieved much improvement. In this paper, we focus on how to better leverage gloss knowledge in a supervised neural WSD system. We construct context-gloss pairs and propose three BERT based models for WSD. We fine-tune the pre-trained BERT model and achieve new state-of-the-art results on WSD task. D19-1355 10.18653/v1/D19-1355 @@ -4154,7 +4154,7 @@ BeiChen Jian-GuangLou ZhoujunLi - 3513–3518 + 3515–3520 One key component in text-to-SQL is to predict the comparison relations between columns and their values. To the best of our knowledge, no existing models explicitly introduce external common knowledge to address this problem, thus their capabilities of predicting comparison relations are limited beyond training data. In this paper, we propose to leverage adjective-noun phrasing knowledge mined from the web to predict the comparison relations in text-to-SQL. Experimental results on both the original and the re-split Spider dataset show that our approach achieves significant improvement over state-of-the-art methods on comparison relation prediction. D19-1356 10.18653/v1/D19-1356 @@ -4164,7 +4164,7 @@ KokiWashio SatoshiSekine TsuneakiKato - 3519–3525 + 3521–3527 Definition modeling includes acquiring word embeddings from dictionary definitions and generating definitions of words. While the meanings of defining words are important in dictionary definitions, it is crucial to capture the lexical semantic relations between defined words and defining words. However, thus far, the utilization of such relations has not been explored for definition modeling. In this paper, we propose definition modeling methods that use lexical semantic relations. To utilize implicit semantic relations in definitions, we use unsupervisedly obtained pattern-based word-pair embeddings that represent semantic relations of word pairs. Experimental results indicate that our methods improve the performance in learning embeddings from definitions, as well as definition generation. D19-1357 D19-1357.Attachment.zip @@ -4175,7 +4175,7 @@ Kang MinYoo TaeukKim Sang-gooLee - 3526–3531 + 3528–3533 We propose a simple yet effective approach for improving Korean word representations using additional linguistic annotation (i.e. Hanja). We employ cross-lingual transfer learning in training word representations by leveraging the fact that Hanja is closely related to Chinese. We evaluate the intrinsic quality of representations learned through our approach using the word analogy and similarity tests. In addition, we demonstrate their effectiveness on several downstream tasks, including a novel Korean news headline generation task. D19-1358 D19-1358.Attachment.zip @@ -4187,7 +4187,7 @@ FedericoScozzafava FedericoMartelli RobertoNavigli - 3532–3538 + 3534–3540 Current research in knowledge-based Word Sense Disambiguation (WSD) indicates that performances depend heavily on the Lexical Knowledge Base (LKB) employed. This paper introduces SyntagNet, a novel resource consisting of manually disambiguated lexical-semantic combinations. By capturing sense distinctions evoked by syntagmatic relations, SyntagNet enables knowledge-based WSD systems to establish a new state of the art which challenges the hitherto unrivaled performances attained by supervised approaches. To the best of our knowledge, SyntagNet is the first large-scale manually-curated resource of this kind made available to the community (at http://syntagnet.org). D19-1359 10.18653/v1/D19-1359 @@ -4199,7 +4199,7 @@ JaminShin ZihanLiu PascaleFung - 3539–3545 + 3541–3547 In countries that speak multiple main languages, mixing up different languages within a conversation is commonly called code-switching. Previous works addressing this challenge mainly focused on word-level aspects such as word embeddings. However, in many cases, languages share common subwords, especially for closely related languages, but also for languages that are seemingly irrelevant. Therefore, we propose Hierarchical Meta-Embeddings (HME) that learn to combine multiple monolingual word-level and subword-level embeddings to create language-agnostic lexical representations. On the task of Named Entity Recognition for English-Spanish code-switching data, our model achieves the state-of-the-art performance in the multilingual settings. We also show that, in cross-lingual settings, our model not only leverages closely related languages, but also learns from languages with different roots. Finally, we show that combining different subunits are crucial for capturing code-switching entities. D19-1360 10.18653/v1/D19-1360 @@ -4210,7 +4210,7 @@ YingmingLi MingChen ZhongfeiZhang - 3546–3551 + 3548–3553 In this paper, we develop a novel Sparse Self-Attention Fine-tuning model (referred as SSAF) which integrates sparsity into self-attention mechanism to enhance the fine-tuning performance of BERT. In particular, sparsity is introduced into the self-attention by replacing softmax function with a controllable sparse transformation when fine-tuning with BERT. It enables us to learn a structurally sparse attention distribution, which leads to a more interpretable representation for the whole input. The proposed model is evaluated on sentiment analysis, question answering, and natural language inference tasks. The extensive experimental results across multiple datasets demonstrate its effectiveness and superiority to the baseline methods. D19-1361 10.18653/v1/D19-1361 @@ -4220,7 +4220,7 @@ LukasLange Michael A.Hedderich DietrichKlakow - 3552–3557 + 3554–3559 In low-resource settings, the performance of supervised labeling models can be improved with automatically annotated or distantly supervised data, which is cheap to create but often noisy. Previous works have shown that significant improvements can be reached by injecting information about the confusion between clean and noisy labels in this additional training data into the classifier training. However, for noise estimation, these approaches either do not take the input features (in our case word embeddings) into account, or they need to learn the noise modeling from scratch which can be difficult in a low-resource setting. We propose to cluster the training data using the input features and then compute different confusion matrices for each cluster. To the best of our knowledge, our approach is the first to leverage feature-dependent noise modeling with pre-initialized confusion matrices. We evaluate on low-resource named entity recognition settings in several languages, showing that our methods improve upon other confusion-matrix based methods by up to 9%. D19-1362 10.18653/v1/D19-1362 @@ -4230,7 +4230,7 @@ HagaiTaitelbaum GalChechik JacobGoldberger - 3558–3563 + 3560–3565 In this paper we present a novel approach to simultaneously representing multiple languages in a common space. Procrustes Analysis (PA) is commonly used to find the optimal orthogonal word mapping in the bilingual case. The proposed Multi Pairwise Procrustes Analysis (MPPA) is a natural extension of the PA algorithm to multilingual word mapping. Unlike previous PA extensions that require a k-way dictionary, this approach requires only pairwise bilingual dictionaries that are much easier to construct. D19-1363 10.18653/v1/D19-1363 @@ -4244,7 +4244,7 @@ SaloniPotdar ShiyuChang MoYu - 3564–3570 + 3566–3572 Out-of-domain (OOD) detection for low-resource text classification is a realistic but understudied task. The goal is to detect the OOD cases with limited in-domain (ID) training data, since in machine learning applications we observe that training data is often insufficient. In this work, we propose an OOD-resistant Prototypical Network to tackle this zero-shot OOD detection and few-shot ID classification task. Evaluations on real-world datasets show that the proposed solution outperforms state-of-the-art methods in zero-shot OOD detection task, while maintaining a competitive performance on ID classification task. D19-1364 10.18653/v1/D19-1364 @@ -4256,7 +4256,7 @@ LiliMou ZhoujunLi WenhanChao - 3571–3576 + 3573–3578 Formality text style transfer plays an important role in various NLP applications, such as non-native speaker assistants and child education. Early studies normalize informal sentences with rules, before statistical and neural models become a prevailing method in the field. While a rule-based system is still a common preprocessing step for formality style transfer in the neural era, it could introduce noise if we use the rules in a naive way such as data preprocessing. To mitigate this problem, we study how to harness rules into a state-of-the-art neural network that is typically pretrained on massive corpora. We propose three fine-tuning methods in this paper and achieve a new state-of-the-art on benchmark datasets D19-1365 10.18653/v1/D19-1365 @@ -4268,7 +4268,7 @@ Hong-YouChen Chi-JenLu Shou-DeLin - 3577–3582 + 3579–3584 The objective of non-parallel text style transfer, or controllable text generation, is to alter specific attributes (e.g. sentiment, mood, tense, politeness, etc) of a given text while preserving its remaining attributes and content. Generative adversarial network (GAN) is a popular model to ensure the transferred sentences are realistic and have the desired target styles. However, training GAN often suffers from mode collapse problem, which causes that the transferred text is little related to the original text. In this paper, we propose a new GAN model with a word-level conditional architecture and a two-phase training procedure. By using a style-related condition architecture before generating a word, our model is able to maintain style-unrelated words while changing the others. By separating the training procedure into reconstruction and transfer phases, our model is able to learn a proper text generation process, which further improves the content preservation. We test our model on polarity sentiment transfer and multiple-attribute transfer tasks. The empirical results show that our model achieves comparable evaluation scores in both transfer accuracy and fluency but significantly outperforms other state-of-the-art models in content compatibility on three real-world datasets. D19-1366 D19-1366.Attachment.zip @@ -4281,7 +4281,7 @@ TongXiao ChunliangZhang JingboZhu - 3583–3588 + 3585–3590 In this paper, we study differentiable neural architecture search (NAS) methods for natural language processing. In particular, we improve differentiable architecture search by removing the softmax-local constraint. Also, we apply differentiable NAS to named entity recognition (NER). It is the first time that differentiable NAS methods are adopted in NLP tasks other than language modeling. On both the PTB language modeling and CoNLL-2003 English NER data, our method outperforms strong baselines. It achieves a new state-of-the-art on the NER task. D19-1367 10.18653/v1/D19-1367 @@ -4292,7 +4292,7 @@ MashaNaslidnyk DavePalfrey ArpitMittal - 3589–3594 + 3591–3596 Bilinear models such as DistMult and ComplEx are effective methods for knowledge graph (KG) completion. However, they require large batch sizes, which becomes a performance bottleneck when training on large scale datasets due to memory constraints. In this paper we use occurrences of entity-relation pairs in the dataset to construct a joint learning model and to increase the quality of sampled negatives during training. We show on three standard datasets that when these two techniques are combined, they give a significant improvement in performance, especially when the batch size and the number of generated negative examples are low relative to the size of the dataset. We then apply our techniques to a dataset containing 2 million entities and demonstrate that our model outperforms the baseline by 2.8% absolute on hits@1. D19-1368 D19-1368.Attachment.zip @@ -4301,7 +4301,7 @@ Single Training Dimension Selection for Word Embedding with <fixed-case>PCA</fixed-case> YuWang - 3595–3600 + 3597–3602 In this paper, we present a fast and reliable method based on PCA to select the number of dimensions for word embeddings. First, we train one embedding with a generous upper bound (e.g. 1,000) of dimensions. Then we transform the embeddings using PCA and incrementally remove the lesser dimensions one at a time while recording the embeddings’ performance on language tasks. Lastly, we select the number of dimensions, balancing model size and accuracy. Experiments using various datasets and language tasks demonstrate that we are able to train about 10 times fewer sets of embeddings while retaining optimal performance. Researchers interested in training the best-performing embeddings for downstream tasks, such as sentiment analysis, question answering and hypernym extraction, as well as those interested in embedding compression should find the method helpful. D19-1369 10.18653/v1/D19-1369 @@ -4313,7 +4313,7 @@ GrahamNeubig TaylorBerg-Kirkpatrick YimingYang - 3601–3612 + 3603–3614 When trained effectively, the Variational Autoencoder (VAE) is both a powerful language model and an effective representation learning framework. In practice, however, VAEs are trained with the evidence lower bound (ELBO) as a surrogate objective to the intractable marginal data likelihood. This approach to training yields unstable results, frequently leading to a disastrous local optimum known as posterior collapse. In this paper, we investigate a simple fix for posterior collapse which yields surprisingly effective results. The combination of two known heuristics, previously considered only in isolation, substantially improves held-out likelihood, reconstruction, and latent representation learning when compared with previous state-of-the-art methods. More interestingly, while our experiments demonstrate superiority on these principle evaluations, our method obtains a worse ELBO. We use these results to argue that the typical surrogate objective for VAEs may not be sufficient or necessarily appropriate for balancing the goals of representation learning and data distribution modeling. D19-1370 10.18653/v1/D19-1370 @@ -4323,7 +4323,7 @@ IzBeltagy KyleLo ArmanCohan - 3613–3618 + 3615–3620 Obtaining large-scale annotated data for NLP tasks in the scientific domain is challenging and expensive. We release SciBERT, a pretrained language model based on BERT (Devlin et. al., 2018) to address the lack of high-quality, large-scale labeled scientific data. SciBERT leverages unsupervised pretraining on a large multi-domain corpus of scientific publications to improve performance on downstream scientific NLP tasks. We evaluate on a suite of tasks including sequence tagging, sentence classification and dependency parsing, with datasets from a variety of scientific domains. We demonstrate statistically significant improvements over BERT and achieve new state-of-the-art results on several of these tasks. The code and pretrained models are available at https://github.com/allenai/scibert/. D19-1371 10.18653/v1/D19-1371 @@ -4332,7 +4332,7 @@ Humor Detection: A Transformer Gets the Last Laugh OrionWeller KevinSeppi - 3619–3623 + 3621–3625 Much previous work has been done in attempting to identify humor in text. In this paper we extend that capability by proposing a new task: assessing whether or not a joke is humorous. We present a novel way of approaching this problem by building a model that learns to identify humorous jokes based on ratings gleaned from Reddit pages, consisting of almost 16,000 labeled instances. Using these ratings to determine the level of humor, we then employ a Transformer architecture for its advantages in learning from sentence context. We demonstrate the effectiveness of this approach and show results that are comparable to human performance. We further demonstrate our model’s increased capabilities on humor identification problems, such as the previously created datasets for short jokes and puns. These experiments show that this method outperforms all previous work done on these tasks, with an F-measure of 93.1% for the Puns dataset and 98.6% on the Short Jokes dataset. D19-1372 10.18653/v1/D19-1372 @@ -4342,7 +4342,7 @@ Alham FikriAji KennethHeafield NikolayBogoychev - 3624–3629 + 3626–3631 One way to reduce network traffic in multi-node data-parallel stochastic gradient descent is to only exchange the largest gradients. However, doing so damages the gradient and degrades the model’s performance. Transformer models degrade dramatically while the impact on RNNs is smaller. We restore gradient quality by combining the compressed global gradient with the node’s locally computed uncompressed gradient. Neural machine translation experiments show that Transformer convergence is restored while RNNs converge faster. With our method, training on 4 nodes converges up to 1.5x as fast as with uncompressed gradients and scales 3.5x relative to single-node training. D19-1373 10.18653/v1/D19-1373 @@ -4355,7 +4355,7 @@ NaveenArivazhagan XinLi AmeliaArcher - 3630–3634 + 3632–3636 We propose a practical scheme to train a single multilingual sequence labeling model that yields state of the art results and is small and fast enough to run on a single CPU. Starting from a public multilingual BERT checkpoint, our final model is 6x smaller and 27x faster, and has higher accuracy than a state-of-the-art multilingual baseline. We show that our model especially outperforms on low-resource languages, and works on codemixed input text without being explicitly trained on codemixed examples. We showcase the effectiveness of our method by reporting on part-of-speech tagging and morphological prediction on 70 treebanks and 48 languages. D19-1374 D19-1374.Attachment.zip @@ -4366,7 +4366,7 @@ ZijianZhao SuZhu KaiYu - 3635–3641 + 3637–3643 Spoken Language Understanding (SLU) converts user utterances into structured semantic representations. Data sparsity is one of the main obstacles of SLU due to the high cost of human annotation, especially when domain changes or a new domain comes. In this work, we propose a data augmentation method with atomic templates for SLU, which involves minimum human efforts. The atomic templates produce exemplars for fine-grained constituents of semantic representations. We propose an encoder-decoder model to generate the whole utterance from atomic exemplars. Moreover, the generator could be transferred from source domains to help a new domain which has little data. Experimental results show that our method achieves significant improvements on DSTC 2&3 dataset which is a domain adaptation setting of SLU. D19-1375 10.18653/v1/D19-1375 @@ -4376,7 +4376,7 @@ HaoPeng RoySchwartz Noah A.Smith - 3642–3649 + 3644–3651 We present PaLM, a hybrid parser and neural language model. Building on an RNN language model, PaLM adds an attention layer over text spans in the left context. An unsupervised constituency parser can be derived from its attention weights, using a greedy decoding algorithm. We evaluate PaLM on language modeling, and empirically show that it outperforms strong baselines. If syntactic annotations are available, the attention component can be trained in a supervised manner, providing syntactically-informed representations of the context, and further improving language modeling performance. D19-1376 D19-1376.Attachment.pdf @@ -4387,7 +4387,7 @@ QingkaiMin YuefengShi YueZhang - 3650–3656 + 3652–3658 The task of semantic parsing is highly useful for dialogue and question answering systems. Many datasets have been proposed to map natural language text into SQL, among which the recent Spider dataset provides cross-domain samples with multiple tables and complex queries. We build a Spider dataset for Chinese, which is currently a low-resource language in this task area. Interesting research questions arise from the uniqueness of the language, which requires word segmentation, and also from the fact that SQL keywords and columns of DB tables are typically written in English. We compare character- and word-based encoders for a semantic parser, and different embedding schemes. Results show that word-based semantic parser is subject to segmentation errors and cross-lingual word embeddings are useful for text-to-SQL. D19-1377 10.18653/v1/D19-1377 @@ -4397,7 +4397,7 @@ BenBogin MattGardner JonathanBerant - 3657–3662 + 3659–3664 State-of-the-art semantic parsers rely on auto-regressive decoding, emitting one symbol at a time. When tested against complex databases that are unobserved at training time (zero-shot), the parser often struggles to select the correct set of database constants in the new database, due to the local nature of decoding. %since their decisions are based on weak, local information only. In this work, we propose a semantic parser that globally reasons about the structure of the output query to make a more contextually-informed selection of database constants. We use message-passing through a graph neural network to softly select a subset of database constants for the output query, conditioned on the question. Moreover, we train a model to rank queries based on the global alignment of database constants to question words. We apply our techniques to the current state-of-the-art model for Spider, a zero-shot semantic parsing dataset with complex databases, increasing accuracy from 39.4% to 47.4%. D19-1378 D19-1378.Attachment.pdf @@ -4408,7 +4408,7 @@ HirokiOuchi JunSuzuki KentaroInui - 3663–3669 + 3665–3671 In transductive learning, an unlabeled test set is used for model training. Although this setting deviates from the common assumption of a completely unseen test set, it is applicable in many real-world scenarios, wherein the texts to be processed are known in advance. However, despite its practical advantages, transductive learning is underexplored in natural language processing. Here we conduct an empirical study of transductive learning for neural models and demonstrate its utility in syntactic and semantic tasks. Specifically, we fine-tune language models (LMs) on an unlabeled test set to obtain test-set-specific word representations. Through extensive experiments, we demonstrate that despite its simplicity, transductive LM fine-tuning consistently improves state-of-the-art neural models in in-domain and out-of-domain settings. D19-1379 D19-1379.Attachment.pdf @@ -4419,7 +4419,7 @@ NadaAlmarwani HananAldarmaki MonaDiab - 3670–3676 + 3672–3678 Vector averaging remains one of the most popular sentence embedding methods in spite of its obvious disregard for syntactic structure. While more complex sequential or convolutional networks potentially yield superior classification performance, the improvements in classification accuracy are typically mediocre compared to the simple vector averaging. As an efficient alternative, we propose the use of discrete cosine transform (DCT) to compress word sequences in an order-preserving manner. The lower order DCT coefficients represent the overall feature patterns in sentences, which results in suitable embeddings for tasks that could benefit from syntactic features. Our results in semantic probing tasks demonstrate that DCT embeddings indeed preserve more syntactic information compared with vector averaging. With practically equivalent complexity, the model yields better overall performance in downstream classification tasks that correlate with syntactic features, which illustrates the capacity of DCT to preserve word order information. D19-1380 10.18653/v1/D19-1380 @@ -4429,7 +4429,7 @@ Kurt JunsheanEspinosa MakotoMiwa SophiaAnaniadou - 3677–3684 + 3679–3686 We tackle the nested and overlapping event detection task and propose a novel search-based neural network (SBNN) structured prediction model that treats the task as a search problem on a relation graph of trigger-argument structures. Unlike existing structured prediction tasks such as dependency parsing, the task targets to detect DAG structures, which constitute events, from the relation graph. We define actions to construct events and use all the beams in a beam search to detect all event structures that may be overlapping and nested. The search process constructs events in a bottom-up manner while modelling the global properties for nested and overlapping structures simultaneously using neural networks. We show that the model achieves performance comparable to the state-of-the-art model Turku Event Extraction System (TEES) on the BioNLP Cancer Genetics (CG) Shared Task 2013 without the use of any syntactic and hand-engineered features. Further analyses on the development set show that our model is more computationally efficient while yielding higher F1-score performance. D19-1381 10.18653/v1/D19-1381 @@ -4440,7 +4440,7 @@ YuanZhang ChrisTar JasonBaldridge - 3685–3690 + 3687–3692 Most existing work on adversarial data generation focuses on English. For example, PAWS (Paraphrase Adversaries from Word Scrambling) consists of challenging English paraphrase identification pairs from Wikipedia and Quora. We remedy this gap with PAWS-X, a new dataset of 23,659 human translated PAWS evaluation pairs in six typologically distinct languages: French, Spanish, German, Chinese, Japanese, and Korean. We provide baseline numbers for three models with different capacity to capture non-local context and sentence structure, and using different multilingual training and evaluation regimes. Multilingual BERT fine-tuned on PAWS English plus machine-translated data performs the best, with a range of 83.1-90.8 accuracy across the non-English languages and an average accuracy gain of 23% over the next best model. PAWS-X shows the effectiveness of deep, multilingual pre-training while also leaving considerable headroom as a new challenge to drive multilingual research that better captures structure and contextual information. D19-1382 10.18653/v1/D19-1382 @@ -4452,7 +4452,7 @@ DanielKing BhavanaDalvi DanWeld - 3691–3697 + 3693–3699 As a step toward better document-level understanding, we explore classification of a sequence of sentences into their corresponding categories, a task that requires understanding sentences in context of the document. Recent successful models for this task have used hierarchical models to contextualize sentence representations, and Conditional Random Fields (CRFs) to incorporate dependencies between subsequent labels. In this work, we show that pretrained language models, BERT (Devlin et al., 2018) in particular, can be used for this task to capture contextual dependencies without the need for hierarchical encoding nor a CRF. Specifically, we construct a joint sentence representation that allows BERT Transformer layers to directly utilize contextual information from all words in all sentences. Our approach achieves state-of-the-art results on four datasets, including a new dataset of structured scientific abstracts. D19-1383 10.18653/v1/D19-1383 @@ -4462,7 +4462,7 @@ LauraHarding Graesser KyunghyunCho DouweKiela - 3698–3708 + 3700–3710 We describe a multi-agent communication framework for examining high-level linguistic phenomena at the community-level. We demonstrate that complex linguistic behavior observed in natural language can be reproduced in this simple setting: i) the outcome of contact between communities is a function of inter- and intra-group connectivity; ii) linguistic contact either converges to the majority protocol, or in balanced cases leads to novel creole languages of lower complexity; and iii) a linguistic continuum emerges where neighboring languages are more mutually intelligible than farther removed languages. We conclude that at least some of the intricate properties of language evolution need not depend on complex evolved linguistic capabilities, but can emerge from simple social exchanges between perceptually-enabled agents playing communication games. D19-1384 10.18653/v1/D19-1384 @@ -4471,7 +4471,7 @@ <fixed-case>T</fixed-case>alk<fixed-case>D</fixed-case>own: A Corpus for Condescension Detection in Context ZijianWang ChristopherPotts - 3709–3717 + 3711–3719 Condescending language use is caustic; it can bring dialogues to an end and bifurcate communities. Thus, systems for condescension detection could have a large positive impact. A challenge here is that condescension is often impossible to detect from isolated utterances, as it depends on the discourse and social context. To address this, we present TalkDown, a new labeled dataset of condescending linguistic acts in context. We show that extending a language-only model with representations of the discourse improves performance, and we motivate techniques for dealing with the low rates of condescension overall. We also use our model to estimate condescension rates in various online communities and relate these differences to differing community norms. D19-1385 10.18653/v1/D19-1385 @@ -4480,7 +4480,7 @@ Summary Cloze: A New Task for Content Selection in Topic-Focused Summarization DanielDeutsch DanRoth - 3718–3727 + 3720–3729 A key challenge in topic-focused summarization is determining what information should be included in the summary, a problem known as content selection. In this work, we propose a new method for studying content selection in topic-focused summarization called the summary cloze task. The goal of the summary cloze task is to generate the next sentence of a summary conditioned on the beginning of the summary, a topic, and a reference document(s). The main challenge is deciding what information in the references is relevant to the topic and partial summary and should be included in the summary. Although the cloze task does not address all aspects of the traditional summarization problem, the more narrow scope of the task allows us to collect a large-scale datset of nearly 500k summary cloze instances from Wikipedia. We report experimental results on this new dataset using various extractive models and a two-step abstractive model that first extractively selects a small number of sentences and then abstractively summarizes them. Our results show that the topic and partial summary help the models identify relevant content, but the task remains a significant challenge. D19-1386 D19-1386.Attachment.zip @@ -4490,7 +4490,7 @@ Text Summarization with Pretrained Encoders YangLiu MirellaLapata - 3728–3738 + 3730–3740 Bidirectional Encoder Representations from Transformers (BERT) represents the latest incarnation of pretrained language models which have recently advanced a wide range of natural language processing tasks. In this paper, we showcase how BERT can be usefully applied in text summarization and propose a general framework for both extractive and abstractive models. We introduce a novel document-level encoder based on BERT which is able to express the semantics of a document and obtain representations for its sentences. Our extractive model is built on top of this encoder by stacking several inter-sentence Transformer layers. For abstractive summarization, we propose a new fine-tuning schedule which adopts different optimizers for the encoder and the decoder as a means of alleviating the mismatch between the two (the former is pretrained while the latter is not). We also demonstrate that a two-staged fine-tuning approach can further boost the quality of the generated summaries. Experiments on three datasets show that our model achieves state-of-the-art results across the board in both extractive and abstractive settings. D19-1387 D19-1387.Attachment.pdf @@ -4504,7 +4504,7 @@ ZhangmingChan DongyanZhao RuiYan - 3739–3749 + 3741–3751 Under special circumstances, summaries should conform to a particular style with patterns, such as court judgments and abstracts in academic papers. To this end, the prototype document-summary pairs can be utilized to generate better summaries. There are two main challenges in this task: (1) the model needs to incorporate learned patterns from the prototype, but (2) should avoid copying contents other than the patternized words—such as irrelevant facts—into the generated summaries. To tackle these challenges, we design a model named Prototype Editing based Summary Generator (PESG). PESG first learns summary patterns and prototype facts by analyzing the correlation between a prototype document and its summary. Prototype facts are then utilized to help extract facts from the input document. Next, an editing generator generates new summary based on the summary pattern or extracted facts. Finally, to address the second challenge, a fact checker is used to estimate mutual information between the input document and generated summary, providing an additional signal for the generator. Extensive experiments conducted on a large-scale real-world text summarization dataset show that PESG achieves the state-of-the-art performance in terms of both automatic metrics and human evaluations. D19-1388 10.18653/v1/D19-1388 @@ -4515,7 +4515,7 @@ AriHoltzman JanBuys YejinChoi - 3750–3759 + 3752–3761 The principle of the Information Bottleneck (Tishby et al., 1999) produces a summary of information X optimized to predict some other relevant information Y. In this paper, we propose a novel approach to unsupervised sentence summarization by mapping the Information Bottleneck principle to a conditional language modelling objective: given a sentence, our approach seeks a compressed sentence that can best predict the next sentence. Our iterative algorithm under the Information Bottleneck objective searches gradually shorter subsequences of the given sentence while maximizing the probability of the next sentence conditioned on the summary. Using only pretrained language models with no direct supervision, our approach can efficiently perform extractive sentence summarization over a large corpus. Building on our unsupervised extractive summarization, we also present a new approach to self-supervised abstractive summarization, where a transformer-based language model is trained on the output summaries of our unsupervised method. Empirical results demonstrate that our extractive method outperforms other unsupervised models on multiple automatic metrics. In addition, we find that our self-supervised abstractive model outperforms unsupervised baselines (including our own) by human evaluation along multiple attributes. D19-1389 D19-1389.Attachment.zip @@ -4527,7 +4527,7 @@ YangZhao HuiSu DietrichKlakow - 3760–3771 + 3762–3773 Pointer Generators have been the de facto standard for modern summarization systems. However, this architecture faces two major drawbacks: Firstly, the pointer is limited to copying the exact words while ignoring possible inflections or abstractions, which restricts its power of capturing richer latent alignment. Secondly, the copy mechanism results in a strong bias towards extractive generations, where most sentences are produced by simply copying from the source text. In this paper, we address these problems by allowing the model to “edit” pointed tokens instead of always hard copying them. The editing is performed by transforming the pointed word vector into a target space with a learned relation embedding. On three large-scale summarization dataset, we show the model is able to (1) capture more latent alignment relations than exact word matches, (2) improve word alignment accuracy, allowing for better model interpretation and controlling, (3) generate higher-quality summaries validated by both qualitative and quantitative evaluations and (4) bring more abstraction to the generated summaries. D19-1390 D19-1390.Attachment.pdf @@ -4538,7 +4538,7 @@ BailinWang IvanTitov MirellaLapata - 3772–3783 + 3774–3785 Semantic parsing aims to map natural language utterances onto machine interpretable meaning representations, aka programs whose execution against a real-world environment produces a denotation. Weakly-supervised semantic parsers are trained on utterance-denotation pairs treating programs as latent. The task is challenging due to the large search space and spuriousness of programs which may execute to the correct answer but do not generalize to unseen examples. Our goal is to instill an inductive bias in the parser to help it distinguish between spurious and correct programs. We capitalize on the intuition that correct programs would likely respect certain structural constraints were they to be aligned to the question (e.g., program fragments are unlikely to align to overlapping text spans) and propose to model alignments as structured latent variables. In order to make the latent-alignment framework tractable, we decompose the parsing task into (1) predicting a partial “abstract program” and (2) refining it while modeling structured alignments with differential dynamic programming. We obtain state-of-the-art performance on the WikiTableQuestions and WikiSQL datasets. When compared to a standard attention baseline, we observe that the proposed structured-alignment mechanism is highly beneficial. D19-1391 10.18653/v1/D19-1391 @@ -4549,7 +4549,7 @@ XutaiMa KevinDuh BenjaminVan Durme - 3784–3796 + 3786–3798 We unify different broad-coverage semantic parsing tasks into a transduction parsing paradigm, and propose an attention-based neural transducer that incrementally builds meaning representation via a sequence of semantic relations. By leveraging multiple attention mechanisms, the neural transducer can be effectively trained without relying on a pre-trained aligner. Experiments separately conducted on three broad-coverage semantic parsing tasks – AMR, SDP and UCCA – demonstrate that our attention-based neural transducer improves the state of the art on both AMR and UCCA, and is competitive with the state of the art on SDP. D19-1392 10.18653/v1/D19-1392 @@ -4558,7 +4558,7 @@ Core Semantic First: A Top-down Approach for <fixed-case>AMR</fixed-case> Parsing DengCai WaiLam - 3797–3807 + 3799–3809 We introduce a novel scheme for parsing a piece of text into its Abstract Meaning Representation (AMR): Graph Spanning based Parsing (GSP). One novel characteristic of GSP is that it constructs a parse graph incrementally in a top-down fashion. Starting from the root, at each step, a new node and its connections to existing nodes will be jointly predicted. The output graph spans the nodes by the distance to the root, following the intuition of first grasping the main ideas then digging into more details. The core semantic first principle emphasizes capturing the main ideas of a sentence, which is of great interest. We evaluate our model on the latest AMR sembank and achieve the state-of-the-art performance in the sense that no heuristic graph re-categorization is adopted. More importantly, the experiments show that our parser is especially good at obtaining the core semantics. D19-1393 D19-1393.Attachment.pdf @@ -4568,7 +4568,7 @@ Don’t paraphrase, detect! Rapid and Effective Data Collection for Semantic Parsing JonathanHerzig JonathanBerant - 3808–3818 + 3810–3820 A major hurdle on the road to conversational interfaces is the difficulty in collecting data that maps language utterances to logical forms. One prominent approach for data collection has been to automatically generate pseudo-language paired with logical forms, and paraphrase the pseudo-language to natural language through crowdsourcing (Wang et al., 2015). However, this data collection procedure often leads to low performance on real data, due to a mismatch between the true distribution of examples and the distribution induced by the data collection procedure. In this paper, we thoroughly analyze two sources of mismatch in this process: the mismatch in logical form distribution and the mismatch in language distribution between the true and induced distributions. We quantify the effects of these mismatches, and propose a new data collection approach that mitigates them. Assuming access to unlabeled utterances from the true distribution, we combine crowdsourcing with a paraphrase model to detect correct logical forms for the unlabeled utterances. On two datasets, our method leads to 70.6 accuracy on average on the true distribution, compared to 51.3 in paraphrasing-based data collection. D19-1394 10.18653/v1/D19-1394 @@ -4581,7 +4581,7 @@ LiqiangNie WeiliGuan ChengYang - 3819–3827 + 3821–3829 Distantly-supervised relation extraction has proven to be effective to find relational facts from texts. However, the existing approaches treat labels as independent and meaningless one-hot vectors, which cause a loss of potential label information for selecting valid instances. In this paper, we propose a novel multi-layer attention-based model to improve relation extraction with joint label embedding. The model makes full use of both structural information from Knowledge Graphs and textual information from entity descriptions to learn label embeddings through gating integration while avoiding the imposed noise with an attention mechanism. Then the learned label embeddings are used as another atten- tion over the instances (whose embeddings are also enhanced with the entity descriptions) for improving relation extraction. Extensive experiments demonstrate that our model significantly outperforms state-of-the-art methods. D19-1395 10.18653/v1/D19-1395 @@ -4593,7 +4593,7 @@ KangLiu JunZhao ShengpingLiu - 3828–3838 + 3830–3840 The lack of word boundaries information has been seen as one of the main obstacles to develop a high performance Chinese named entity recognition (NER) system. Fortunately, the automatically constructed lexicon contains rich word boundaries information and word semantic information. However, integrating lexical knowledge in Chinese NER tasks still faces challenges when it comes to self-matched lexical words as well as the nearest contextual lexical words. We present a Collaborative Graph Network to solve these challenges. Experiments on various datasets show that our model not only outperforms the state-of-the-art (SOTA) results, but also achieves a speed that is six to fifteen times faster than that of the SOTA model. D19-1396 10.18653/v1/D19-1396 @@ -4604,7 +4604,7 @@ LiyuanLiu MaosenZhang XiangRen - 3839–3848 + 3841–3850 In recent years there is a surge of interest in applying distant supervision (DS) to automatically generate training data for relation extraction (RE). In this paper, we study the problem what limits the performance of DS-trained neural models, conduct thorough analyses, and identify a factor that can influence the performance greatly, shifted label distribution. Specifically, we found this problem commonly exists in real-world DS datasets, and without special handing, typical DS-RE models cannot automatically adapt to this shift, thus achieving deteriorated performance. To further validate our intuition, we develop a simple yet effective adaptation method for DS-trained models, bias adjustment, which updates models learned over the source domain (i.e., DS training set) with a label distribution estimated on the target domain (i.e., test set). Experiments demonstrate that bias adjustment achieves consistent performance gains on DS-trained models, especially on neural models, with an up to 23% relative F1 improvement, which verifies our assumptions. Our code and data can be found at https://github.com/INK-USC/shifted-label-distribution. D19-1397 D19-1397.Attachment.pdf @@ -4616,7 +4616,7 @@ GangWang YansongFeng JinpengHuai - 3849–3859 + 3851–3861 Many existing relation extraction (RE) models make decisions globally using integer linear programming (ILP). However, it is nontrivial to make use of integer linear programming as a blackbox solver for RE. Its cost of time and memory may become unacceptable with the increase of data scale, and redundant information needs to be encoded cautiously for ILP. In this paper, we propose an easy first approach for relation extraction with information redundancies, embedded in the results produced by local sentence level extractors, during which conflict decisions are resolved with domain and uniqueness constraints. Information redundancies are leveraged to support both easy first collective inference for easy decisions in the first stage and ILP for hard decisions in a subsequent stage. Experimental study shows that our approach improves the efficiency and accuracy of RE, and outperforms both ILP and neural network-based methods. D19-1398 10.18653/v1/D19-1398 @@ -4625,7 +4625,7 @@ Dependency-Guided <fixed-case>LSTM</fixed-case>-<fixed-case>CRF</fixed-case> for Named Entity Recognition ZhanmingJie WeiLu - 3860–3870 + 3862–3872 Dependency tree structures capture long-distance and syntactic relationships between words in a sentence. The syntactic relations (e.g., nominal subject, object) can potentially infer the existence of certain named entities. In addition, the performance of a named entity recognizer could benefit from the long-distance dependencies between the words in dependency trees. In this work, we propose a simple yet effective dependency-guided LSTM-CRF model to encode the complete dependency trees and capture the above properties for the task of named entity recognition (NER). The data statistics show strong correlations between the entity types and dependency relations. We conduct extensive experiments on several standard datasets and demonstrate the effectiveness of the proposed model in improving NER and achieving state-of-the-art performance. Our analysis reveals that the significant improvements mainly result from the dependency relations and long-distance interactions provided by dependency trees. D19-1399 D19-1399.Attachment.zip @@ -4637,7 +4637,7 @@ GalLavee IdoGuy KiraRadinsky - 3871–3881 + 3873–3883 Large training datasets are required to achieve competitive performance in most natural language tasks. The acquisition process for these datasets is labor intensive, expensive, and time consuming. This process is also prone to human errors. In this work, we show that cross-cultural differences can be harnessed for natural language text classification. We present a transfer-learning framework that leverages widely-available unaligned bilingual corpora for classification tasks, using no task-specific data. Our empirical evaluation on two tasks – formality classification and sarcasm detection – shows that the cross-cultural difference between German and American English, as manifested in product review text, can be applied to achieve good performance for formality classification, while the difference between Japanese and American English can be applied to achieve good performance for sarcasm detection – both without any task-specific labeled data. D19-1400 10.18653/v1/D19-1400 @@ -4647,7 +4647,7 @@ OrenMelamud MihaelaBornea KenBarker - 3882–3891 + 3884–3893 Supervised learning models often perform poorly at low-shot tasks, i.e. tasks for which little labeled data is available for training. One prominent approach for improving low-shot learning is to use unsupervised pre-trained neural models. Another approach is to obtain richer supervision by collecting annotator rationales (explanations supporting label annotations). In this work, we combine these two approaches to improve low-shot text classification with two novel methods: a simple bag-of-words embedding approach; and a more complex context-aware method, based on the BERT model. In experiments with two English text classification datasets, we demonstrate substantial performance gains from combining pre-training with rationales. Furthermore, our investigation of a range of train-set sizes reveals that the simple bag-of-words approach is the clear top performer when there are only a few dozen training instances or less, while more complex models, such as BERT or CNN, require more training data to shine. D19-1401 10.18653/v1/D19-1401 @@ -4656,7 +4656,7 @@ <fixed-case>P</fixed-case>ro<fixed-case>S</fixed-case>eqo: Projection Sequence Networks for On-Device Text Classification ZornitsaKozareva SujithRavi - 3892–3901 + 3894–3903 We propose a novel on-device sequence model for text classification using recurrent projections. Our model ProSeqo uses dynamic recurrent projections without the need to store or look up any pre-trained embeddings. This results in fast and compact neural networks that can perform on-device inference for complex short and long text classification tasks. We conducted exhaustive evaluation on multiple text classification tasks. Results show that ProSeqo outperformed state-of-the-art neural and on-device approaches for short text classification tasks such as dialog act and intent prediction. To the best of our knowledge, ProSeqo is the first on-device long text classification neural model. It achieved comparable results to previous neural approaches for news article, answers and product categorization, while preserving small memory footprint and maintaining high accuracy. D19-1402 10.18653/v1/D19-1402 @@ -4669,7 +4669,7 @@ XiaodanZhu PingJian JianSun - 3902–3911 + 3904–3913 Text classification tends to struggle when data is deficient or when it needs to adapt to unseen classes. In such challenging scenarios, recent studies have used meta-learning to simulate the few-shot task, in which new queries are compared to a small support set at the sample-wise level. However, this sample-wise comparison may be severely disturbed by the various expressions in the same class. Therefore, we should be able to learn a general representation of each class in the support set and then compare it to new queries. In this paper, we propose a novel Induction Network to learn such a generalized class-wise representation, by innovatively leveraging the dynamic routing algorithm in meta-learning. In this way, we find the model is able to induce and generalize better. We evaluate the proposed model on a well-studied sentiment classification dataset (English) and a real-world dialogue intent classification dataset (Chinese). Experiment results show that on both datasets, the proposed model significantly outperforms the existing state-of-the-art approaches, proving the effectiveness of class-wise generalization in few-shot text classification. D19-1403 10.18653/v1/D19-1403 @@ -4679,7 +4679,7 @@ WenpengYin JamaalHay DanRoth - 3912–3921 + 3914–3923 Zero-shot text classification (0Shot-TC) is a challenging NLU problem to which little attention has been paid by the research community. 0Shot-TC aims to associate an appropriate label with a piece of text, irrespective of the text domain and the aspect (e.g., topic, emotion, event, etc.) described by the label. And there are only a few articles studying 0Shot-TC, all focusing only on topical categorization which, we argue, is just the tip of the iceberg in 0Shot-TC. In addition, the chaotic experiments in literature make no uniform comparison, which blurs the progress. This work benchmarks the 0Shot-TC problem by providing unified datasets, standardized evaluations, and state-of-the-art baselines. Our contributions include: i) The datasets we provide facilitate studying 0Shot-TC relative to conceptually different and diverse aspects: the “topic” aspect includes “sports” and “politics” as labels; the “emotion” aspect includes “joy” and “anger”; the “situation” aspect includes “medical assistance” and “water shortage”. ii) We extend the existing evaluation setup (label-partially-unseen) – given a dataset, train on some labels, test on all labels – to include a more challenging yet realistic evaluation label-fully-unseen 0Shot-TC (Chang et al., 2008), aiming at classifying text snippets without seeing task specific training data at all. iii) We unify the 0Shot-TC of diverse aspects within a textual entailment formulation and study it this way. D19-1404 10.18653/v1/D19-1404 @@ -4690,7 +4690,7 @@ VivekGupta MaitreyMehta VivekSrikumar - 3922–3933 + 3924–3935 While neural models show remarkable accuracy on individual predictions, their internal beliefs can be inconsistent across examples. In this paper, we formalize such inconsistency as a generalization of prediction error. We propose a learning framework for constraining models using logic rules to regularize them away from inconsistency. Our framework can leverage both labeled and unlabeled examples and is directly compatible with off-the-shelf learning schemes without model redesign. We instantiate our framework on natural language inference, where experiments show that enforcing invariants stated in logic can help make the predictions of neural models both accurate and consistent. D19-1405 10.18653/v1/D19-1405 @@ -4702,7 +4702,7 @@ AleksanderNagaev AigulNugmanova Ivan P.Yamshchikov - 3934–3943 + 3936–3945 This paper shows that standard assessment methodology for style transfer has several significant problems. First, the standard metrics for style accuracy and semantics preservation vary significantly on different re-runs. Therefore one has to report error margins for the obtained results. Second, starting with certain values of bilingual evaluation understudy (BLEU) between input and output and accuracy of the sentiment transfer the optimization of these two standard metrics diverge from the intuitive goal of the style transfer task. Finally, due to the nature of the task itself, there is a specific dependence between these two metrics that could be easily manipulated. Under these circumstances, we suggest taking BLEU between input and human-written reformulations into consideration for benchmarks. We also propose three new architectures that outperform state of the art in terms of this metric. D19-1406 10.18653/v1/D19-1406 @@ -4714,7 +4714,7 @@ JianfengGao WenDong ChangyouChen - 3944–3954 + 3946–3956 Deep latent variable models (LVM) such as variational auto-encoder (VAE) have recently played an important role in text generation. One key factor is the exploitation of smooth latent structures to guide the generation. However, the representation power of VAEs is limited due to two reasons: (1) the Gaussian assumption is often made on the variational posteriors; and meanwhile (2) a notorious “posterior collapse” issue occurs. In this paper, we advocate sample-based representations of variational distributions for natural language, leading to implicit latent features, which can provide flexible representation power compared with Gaussian-based posteriors. We further develop an LVM to directly match the aggregated posterior to the prior. It can be viewed as a natural extension of VAEs with a regularization of maximizing mutual information, mitigating the “posterior collapse” issue. We demonstrate the effectiveness and versatility of our models in various text generation scenarios, including language modeling, unaligned style transfer, and dialog response generation. The source code to reproduce our experimental results is available on GitHub. D19-1407 D19-1407.Attachment.zip @@ -4724,7 +4724,7 @@ Text Emotion Distribution Learning from Small Sample: A Meta-Learning Approach ZhenjieZhao XiaojuanMa - 3955–3965 + 3957–3967 Text emotion distribution learning (EDL) aims to develop models that can predict the intensity values of a sentence across a set of emotion categories. Existing methods based on supervised learning require a large amount of well-labelled training data, which is difficult to obtain due to inconsistent perception of fine-grained emotion intensity. In this paper, we propose a meta-learning approach to learn text emotion distributions from a small sample. Specifically, we propose to learn low-rank sentence embeddings by tensor decomposition to capture their contextual semantic similarity, and use K-nearest neighbors (KNNs) of each sentence in the embedding space to generate sample clusters. We then train a meta-learner that can adapt to new data with only a few training samples on the clusters, and further fit the meta-learner on KNNs of a testing sample for EDL. In this way, we effectively augment the learning ability of a model on the small sample. To demonstrate the performance, we compare the proposed approach with state-of-the-art EDL methods on a widely used EDL dataset: SemEval 2007 Task 14 (Strapparava and Mihalcea, 2007). Results show the superiority of our method on small-sample emotion distribution learning. D19-1408 10.18653/v1/D19-1408 @@ -4735,7 +4735,7 @@ SamuelCarton ShiyanYan QiaozhuMei - 3966–3979 + 3968–3981 We conduct a large-scale, systematic study to evaluate the existing evaluation methods for natural language generation in the context of generating online product reviews. We compare human-based evaluators with a variety of automated evaluation procedures, including discriminative evaluators that measure how well machine-generated text can be distinguished from human-written text, as well as word overlap metrics that assess how similar the generated text compares to human-written references. We determine to what extent these different evaluators agree on the ranking of a dozen of state-of-the-art generators for online product reviews. We find that human evaluators do not correlate well with discriminative evaluators, leaving a bigger question of whether adversarial accuracy is the correct objective for natural language generation. In general, distinguishing machine-generated text is challenging even for human evaluators, and human decisions correlate better with lexical overlaps. We find lexical diversity an intriguing metric that is indicative of the assessments of different evaluators. A post-experiment survey of participants provides insights into how to evaluate and improve the quality of natural language generation systems. D19-1409 D19-1409.Attachment.pdf @@ -4745,7 +4745,7 @@ Sentence-<fixed-case>BERT</fixed-case>: Sentence Embeddings using <fixed-case>S</fixed-case>iamese <fixed-case>BERT</fixed-case>-Networks NilsReimers IrynaGurevych - 3980–3990 + 3982–3992 BERT (Devlin et al., 2018) and RoBERTa (Liu et al., 2019) has set a new state-of-the-art performance on sentence-pair regression tasks like semantic textual similarity (STS). However, it requires that both sentences are fed into the network, which causes a massive computational overhead: Finding the most similar pair in a collection of 10,000 sentences requires about 50 million inference computations (~65 hours) with BERT. The construction of BERT makes it unsuitable for semantic similarity search as well as for unsupervised tasks like clustering. In this publication, we present Sentence-BERT (SBERT), a modification of the pretrained BERT network that use siamese and triplet network structures to derive semantically meaningful sentence embeddings that can be compared using cosine-similarity. This reduces the effort for finding the most similar pair from 65 hours with BERT / RoBERTa to about 5 seconds with SBERT, while maintaining the accuracy from BERT. We evaluate SBERT and SRoBERTa on common STS tasks and transfer learning tasks, where it outperforms other state-of-the-art sentence embeddings methods. D19-1410 10.18653/v1/D19-1410 @@ -4757,7 +4757,7 @@ YipingJin DittayaWanvarie MasashiSugiyama - 3991–4000 + 3993–4002 We consider a document classification problem where document labels are absent but only relevant keywords of a target class and unlabeled documents are given. Although heuristic methods based on pseudo-labeling have been considered, theoretical understanding of this problem has still been limited. Moreover, previous methods cannot easily incorporate well-developed techniques in supervised text classification. In this paper, we propose a theoretically guaranteed learning framework that is simple to implement and has flexible choices of models, e.g., linear models or neural networks. We demonstrate how to optimize the area under the receiver operating characteristic curve (AUC) effectively and also discuss how to adjust it to optimize other well-known evaluation metrics such as the accuracy and F1-measure. Finally, we show the effectiveness of our framework using benchmark datasets. D19-1411 D19-1411.Attachment.rar @@ -4770,7 +4770,7 @@ RuoyuJia SujianLi JingmingLiu - 4001–4013 + 4003–4015 This paper presents a new sequence-to-sequence (seq2seq) pre-training method PoDA (Pre-training of Denoising Autoencoders), which learns representations suitable for text generation tasks. Unlike encoder-only (e.g., BERT) or decoder-only (e.g., OpenAI GPT) pre-training approaches, PoDA jointly pre-trains both the encoder and decoder by denoising the noise-corrupted text, and it also has the advantage of keeping the network architecture unchanged in the subsequent fine-tuning stage. Meanwhile, we design a hybrid model of Transformer and pointer-generator networks as the backbone architecture for PoDA. We conduct experiments on two text generation tasks: abstractive summarization, and grammatical error correction. Results on four datasets show that PoDA can improve model performance over strong baselines without using any task-specific techniques and significantly speed up convergence. D19-1412 10.18653/v1/D19-1412 @@ -4779,7 +4779,7 @@ Dialog Intent Induction with Deep Multi-View Clustering HughPerkins YiYang - 4014–4023 + 4016–4025 We introduce the dialog intent induction task and present a novel deep multi-view clustering approach to tackle the problem. Dialog intent induction aims at discovering user intents from user query utterances in human-human conversations such as dialogs between customer support agents and customers. Motivated by the intuition that a dialog intent is not only expressed in the user query utterance but also captured in the rest of the dialog, we split a conversation into two independent views and exploit multi-view clustering techniques for inducing the dialog intent. In par- ticular, we propose alternating-view k-means (AV-KMEANS) for joint multi-view represen- tation learning and clustering analysis. The key innovation is that the instance-view representations are updated iteratively by predicting the cluster assignment obtained from the alternative view, so that the multi-view representations of the instances lead to similar cluster assignments. Experiments on two public datasets show that AV-KMEANS can induce better dialog intent clusters than state-of-the-art unsupervised representation learning methods and standard multi-view clustering approaches. D19-1413 10.18653/v1/D19-1413 @@ -4790,7 +4790,7 @@ AramGalstyan GregVer Steeg GuillermoCecchi - 4024–4034 + 4026–4036 Recently, kernelized locality sensitive hashcodes have been successfully employed as representations of natural language text, especially showing high relevance to biomedical relation extraction tasks. In this paper, we propose to optimize the hashcode representations in a nearly unsupervised manner, in which we only use data points, but not their class labels, for learning. The optimized hashcode representations are then fed to a supervised classifi er following the prior work. This nearly unsupervised approach allows fine-grained optimization of each hash function, which is particularly suitable for building hashcode representations generalizing from a training set to a test set. We empirically evaluate the proposed approach for biomedical relation extraction tasks, obtaining significant accuracy improvements w.r.t. state-of-the-art supervised and semi-supervised approaches. D19-1414 10.18653/v1/D19-1414 @@ -4800,7 +4800,7 @@ DaniloCroce DanieleRossini RobertoBasili - 4035–4044 + 4037–4046 While NLP systems become more pervasive, their accountability gains value as a focal point of effort. Epistemological opaqueness of nonlinear learning methods, such as deep learning models, can be a major drawback for their adoptions. In this paper, we discuss the application of Layerwise Relevance Propagation over a linguistically motivated neural architecture, the Kernel-based Deep Architecture, in order to trace back connections between linguistic properties of input instances and system decisions. Such connections then guide the construction of argumentations on network’s inferences, i.e., explanations based on real examples, semantically related to the input. We propose here a methodology to evaluate the transparency and coherence of analogy-based explanations modeling an audit stage for the system. Quantitative analysis on two semantic tasks, i.e., question classification and semantic role labeling, show that the explanatory capabilities (native in KDAs) are effective and they pave the way to more complex argumentation methods. D19-1415 10.18653/v1/D19-1415 @@ -4809,7 +4809,7 @@ Enhancing Variational Autoencoders with Mutual Information Neural Estimation for Text Generation DongQian William K.Cheung - 4045–4055 + 4047–4057 While broadly applicable to many natural language processing (NLP) tasks, variational autoencoders (VAEs) are hard to train due to the posterior collapse issue where the latent variable fails to encode the input data effectively. Various approaches have been proposed to alleviate this problem to improve the capability of the VAE. In this paper, we propose to introduce a mutual information (MI) term between the input and its latent variable to regularize the objective of the VAE. Since estimating the MI in the high-dimensional space is intractable, we employ neural networks for the estimation of the MI and provide a training algorithm based on the convex duality approach. Our experimental results on three benchmark datasets demonstrate that the proposed model, compared to the state-of-the-art baselines, exhibits less posterior collapse and has comparable or better performance in language modeling and text generation. We also qualitatively evaluate the inferred latent space and show that the proposed model can generate more reasonable and diverse sentences via linear interpolation in the latent space. D19-1416 D19-1416.Attachment.pdf @@ -4820,7 +4820,7 @@ AmeyaPrabhu CharlesDognin ManeeshSingh - 4056–4066 + 4058–4068 The exploding cost and time needed for data labeling and model training are bottlenecks for training DNN models on large datasets. Identifying smaller representative data samples with strategies like active learning can help mitigate such bottlenecks. Previous works on active learning in NLP identify the problem of sampling bias in the samples acquired by uncertainty-based querying and develop costly approaches to address it. Using a large empirical study, we demonstrate that active set selection using the posterior entropy of deep models like FastText.zip (FTZ) is robust to sampling biases and to various algorithmic choices (query size and strategies) unlike that suggested by traditional literature. We also show that FTZ based query strategy produces sample sets similar to those from more sophisticated approaches (e.g ensemble networks). Finally, we show the effectiveness of the selected samples by creating tiny high-quality datasets, and utilizing them for fast and cheap training of large models. Based on the above, we propose a simple baseline for deep active text classification that outperforms the state of the art. We expect the presented work to be useful and informative for dataset compression and for problems involving active, semi-supervised or online learning scenarios. Code and models are available at: https://github.com/drimpossible/Sampling-Bias-Active-Learning. D19-1417 D19-1417.Attachment.pdf @@ -4831,7 +4831,7 @@ ChristopherClark MarkYatskar LukeZettlemoyer - 4067–4080 + 4069–4082 State-of-the-art models often make use of superficial patterns in the data that do not generalize well to out-of-domain or adversarial settings. For example, textual entailment models often learn that particular key words imply entailment, irrespective of context, and visual question answering models learn to predict prototypical answers, without considering evidence in the image. In this paper, we show that if we have prior knowledge of such biases, we can train a model to be more robust to domain shift. Our method has two stages: we (1) train a naive model that makes predictions exclusively based on dataset biases, and (2) train a robust model as part of an ensemble with the naive one in order to encourage it to focus on other patterns in the data that are more likely to generalize. Experiments on five datasets with out-of-domain test sets show significantly improved robustness in all settings, including a 12 point gain on a changing priors visual question answering dataset and a 9 point gain on an adversarial question answering test set. D19-1418 10.18653/v1/D19-1418 @@ -4846,7 +4846,7 @@ SvenGowal KrishnamurthyDvijotham PushmeetKohli - 4081–4091 + 4083–4093 Neural networks are part of many contemporary NLP systems, yet their empirical successes come at the price of vulnerability to adversarial attacks. Previous work has used adversarial training and data augmentation to partially mitigate such brittleness, but these are unlikely to find worst-case adversaries due to the complexity of the search space arising from discrete text perturbations. In this work, we approach the problem from the opposite direction: to formally verify a system’s robustness against a predefined class of adversarial attacks. We study text classification under synonym replacements or character flip perturbations. We propose modeling these input perturbations as a simplex and then using Interval Bound Propagation – a formal model verification method. We modify the conventional log-likelihood training objective to train models that can be efficiently verified, which would otherwise come with exponential search complexity. The resulting models show only little difference in terms of nominal accuracy, but have much improved verified accuracy under perturbations and come with an efficiently computable formal guarantee on worst case adversaries. D19-1419 D19-1419.Attachment.zip @@ -4858,7 +4858,7 @@ ShiyuChang YangZhang TommiJaakkola - 4092–4101 + 4094–4103 Selective rationalization has become a common mechanism to ensure that predictive models reveal how they use any available features. The selection may be soft or hard, and identifies a subset of input features relevant for prediction. The setup can be viewed as a co-operate game between the selector (aka rationale generator) and the predictor making use of only the selected features. The co-operative setting may, however, be compromised for two reasons. First, the generator typically has no direct access to the outcome it aims to justify, resulting in poor performance. Second, there’s typically no control exerted on the information left outside the selection. We revise the overall co-operative framework to address these challenges. We introduce an introspective model which explicitly predicts and incorporates the outcome into the selection process. Moreover, we explicitly control the rationale complement via an adversary so as not to leave any useful information out of the selection. We show that the two complementary mechanisms maintain both high predictive accuracy and lead to comprehensive rationales. D19-1420 D19-1420.Attachment.zip @@ -4868,7 +4868,7 @@ Experimenting with Power Divergences for Language Modeling MatthieuLabeau Shay B.Cohen - 4102–4112 + 4104–4114 Neural language models are usually trained using Maximum-Likelihood Estimation (MLE). The corresponding objective function for MLE is derived from the Kullback-Leibler (KL) divergence between the empirical probability distribution representing the data and the parametric probability distribution output by the model. However, the word frequency discrepancies in natural language make performance extremely uneven: while the perplexity is usually very low for frequent words, it is especially difficult to predict rare words. In this paper, we experiment with several families (alpha, beta and gamma) of power divergences, generalized from the KL divergence, for learning language models with an objective different than standard MLE. Intuitively, these divergences should affect the way the probability mass is spread during learning, notably by prioritizing performances on high or low-frequency words. In addition, we implement and experiment with various sampling-based objectives, where the computation of the output layer is only done on a small subset of the vocabulary. They are derived as power generalizations of a softmax approximated via Importance Sampling, and Noise Contrastive Estimation, for accelerated learning. Our experiments on the Penn Treebank and Wikitext-2 show that these power divergences can indeed be used to prioritize learning on the frequent or rare words, and lead to general performance improvements in the case of sampling-based learning. D19-1421 D19-1421.Attachment.pdf @@ -4878,7 +4878,7 @@ Hierarchically-Refined Label Attention Network for Sequence Labeling LeyangCui YueZhang - 4113–4126 + 4115–4128 CRF has been used as a powerful model for statistical sequence labeling. For neural sequence labeling, however, BiLSTM-CRF does not always lead to better results compared with BiLSTM-softmax local classification. This can be because the simple Markov label transition model of CRF does not give much information gain over strong neural encoding. For better representing label sequences, we investigate a hierarchically-refined label attention network, which explicitly leverages label embeddings and captures potential long-term label dependency by giving each word incrementally refined label distributions with hierarchical attention. Results on POS tagging, NER and CCG supertagging show that the proposed model not only improves the overall tagging accuracy with similar number of parameters, but also significantly speeds up the training and testing compared to BiLSTM-CRF. D19-1422 D19-1422.Attachment.rar @@ -4890,7 +4890,7 @@ AditiRaghunathan KeremGöksel PercyLiang - 4127–4140 + 4129–4142 State-of-the-art NLP models can often be fooled by adversaries that apply seemingly innocuous label-preserving transformations (e.g., paraphrasing) to input text. The number of possible transformations scales exponentially with text length, so data augmentation cannot cover all transformations of an input. This paper considers one exponentially large family of label-preserving transformations, in which every word in the input can be replaced with a similar word. We train the first models that are provably robust to all word substitutions in this family. Our training procedure uses Interval Bound Propagation (IBP) to minimize an upper bound on the worst-case loss that any combination of word substitutions can induce. To evaluate models’ robustness to these transformations, we measure accuracy on adversarially chosen word substitutions applied to test examples. Our IBP-trained models attain 75% adversarial accuracy on both sentiment analysis on IMDB and natural language inference on SNLI; in comparison, on IMDB, models trained normally and ones trained with data augmentation achieve adversarial accuracy of only 12% and 41%, respectively. D19-1423 D19-1423.Attachment.zip @@ -4902,7 +4902,7 @@ LiDong FuruWei KeXu - 4141–4150 + 4143–4152 Language model pre-training, such as BERT, has achieved remarkable results in many NLP tasks. However, it is unclear why the pre-training-then-fine-tuning paradigm can improve performance and generalization capability across different tasks. In this paper, we propose to visualize loss landscapes and optimization trajectories of fine-tuning BERT on specific datasets. First, we find that pre-training reaches a good initial point across downstream tasks, which leads to wider optima and easier optimization compared with training from scratch. We also demonstrate that the fine-tuning procedure is robust to overfitting, even though BERT is highly over-parameterized for downstream tasks. Second, the visualization results indicate that fine-tuning BERT tends to generalize better because of the flat and wide optima, and the consistency between the training loss surface and the generalization error surface. Third, the lower layers of BERT are more invariant during fine-tuning, which suggests that the layers that are close to input learn more transferable representations of language. D19-1424 10.18653/v1/D19-1424 @@ -4913,9 +4913,8 @@ ShulyWintner Noah A.Smith YuliaTsvetkov - 4151–4161 - Despite impressive performance on many text classification tasks, deep neural networks tend to learn frequent superficial patterns that are specific to the training data and do not always generalize well. In this work, we observe this limitation with respect to the task of native language identification. We find that standard text classifiers which perform well on the test set end up learning topical features which are confounds of the prediction task (e.g., if the input text mentions Sweden, the classifier predicts that the author’s native language is Swedish). We propose a method that represents the latent topical confounds and a model which “unlearns” confounding features by predicting both the label of the input text and the confound; but we train the two predictors adversarially in an alternating fashion to learn a text representation that predicts the correct label but is less prone to using information about the confound. We show that this model generalizes better and learns features that are indicative of the writing style rather than the content.The code is available at: https://github.com/Sachin19/adversarial-classify - + 4153–4163 + Despite impressive performance on many text classification tasks, deep neural networks tend to learn frequent superficial patterns that are specific to the training data and do not always generalize well. In this work, we observe this limitation with respect to the task of native language identification. We find that standard text classifiers which perform well on the test set end up learning topical features which are confounds of the prediction task (e.g., if the input text mentions Sweden, the classifier predicts that the author’s native language is Swedish). We propose a method that represents the latent topical confounds and a model which “unlearns” confounding features by predicting both the label of the input text and the confound; but we train the two predictors adversarially in an alternating fashion to learn a text representation that predicts the correct label but is less prone to using information about the confound. We show that this model generalizes better and learns features that are indicative of the writing style rather than the content. D19-1425 10.18653/v1/D19-1425 @@ -4924,7 +4923,7 @@ ShashankSrivastava IgorLabutov TomMitchell - 4162–4172 + 4164–4174 Natural language has recently been explored as a new medium of supervision for training machine learning models. Here, we explore learning classification tasks using language in a conversational setting – where the automated learner does not simply receive language input from a teacher, but can proactively engage the teacher by asking questions. We present a reinforcement learning framework, where the learner’s actions correspond to question types and the reward for asking a question is based on how the teacher’s response changes performance of the resulting machine learning model on the learning task. In this framework, learning good question-asking strategies corresponds to asking sequences of questions that maximize the cumulative (discounted) reward, and hence quickly lead to effective classifiers. Empirical analysis across three domains shows that learned question-asking strategies expedite classifier training by asking appropriate questions at different points in the learning process. The approach allows learning classifiers from a blend of strategies, including learning from observations, explanations and clarifications. D19-1426 10.18653/v1/D19-1426 @@ -4933,7 +4932,7 @@ Language Modeling for Code-Switching: Evaluation, Integration of Monolingual Data, and Discriminative Training HilaGonen YoavGoldberg - 4173–4183 + 4175–4185 We focus on the problem of language modeling for code-switched language, in the context of automatic speech recognition (ASR). Language modeling for code-switched language is challenging for (at least) three reasons: (1) lack of available large-scale code-switched data for training; (2) lack of a replicable evaluation setup that is ASR directed yet isolates language modeling performance from the other intricacies of the ASR system; and (3) the reliance on generative modeling. We tackle these three issues: we propose an ASR-motivated evaluation setup which is decoupled from an ASR system and the choice of vocabulary, and provide an evaluation dataset for English-Spanish code-switching. This setup lends itself to a discriminative training approach, which we demonstrate to work better than generative language modeling. Finally, we explore a variety of training protocols and verify the effectiveness of training with large amounts of monolingual data followed by fine-tuning with small amounts of code-switched data, for both the generative and discriminative cases. D19-1427 D19-1427.Attachment.zip @@ -4945,7 +4944,7 @@ ClaireGardent ChloéBraud AntoineBordes - 4184–4194 + 4186–4196 Query-based open-domain NLP tasks require information synthesis from long and diverse web results. Current approaches extractively select portions of web text as input to Sequence-to-Sequence models using methods such as TF-IDF ranking. We propose constructing a local graph structured knowledge base for each query, which compresses the web search information and reduces redundancy. We show that by linearizing the graph into a structured input sequence, models can encode the graph representations within a standard Sequence-to-Sequence setting. For two generative tasks with very long text input, long-form question answering and multi-document summarization, feeding graph representations as input can achieve better performance than using retrieved text portions. D19-1428 D19-1428.Attachment.zip @@ -4957,7 +4956,7 @@ ShujianHuang Xin-YuDai JiajunChen - 4195–4204 + 4197–4206 In sequence labeling, previous domain adaptation methods focus on the adaptation from the source domain to the entire target domain without considering the diversity of individual target domain samples, which may lead to negative transfer results for certain samples. Besides, an important characteristic of sequence labeling tasks is that different elements within a given sample may also have diverse domain relevance, which requires further consideration. To take the multi-level domain relevance discrepancy into account, in this paper, we propose a fine-grained knowledge fusion model with the domain relevance modeling scheme to control the balance between learning from the target domain data and learning from the source domain model. Experiments on three sequence labeling tasks show that our fine-grained knowledge fusion model outperforms strong baselines and other state-of-the-art sequence labeling domain adaptation methods. D19-1429 10.18653/v1/D19-1429 @@ -4970,7 +4969,7 @@ TaoQin JianhuangLai Tie-YanLiu - 4205–4215 + 4207–4216 While target-side monolingual data has been proven to be very useful to improve neural machine translation (briefly, NMT) through back translation, source-side monolingual data is not well investigated. In this work, we study how to use both the source-side and target-side monolingual data for NMT, and propose an effective strategy leveraging both of them. First, we generate synthetic bitext by translating monolingual data from the two domains into the other domain using the models pretrained on genuine bitext. Next, a model is trained on a noised version of the concatenated synthetic bitext where each source sequence is randomly corrupted. Finally, the model is fine-tuned on the genuine bitext and a clean version of a subset of the synthetic bitext without adding any noise. Our approach achieves state-of-the-art results on WMT16, WMT17, WMT18 English\leftrightarrowGerman translations and WMT19 German\toFrench translations, which demonstrate the effectiveness of our method. We also conduct a comprehensive study on how each part in the pipeline works. D19-1430 D19-1430.Attachment.zip @@ -4983,7 +4982,7 @@ WeiZhang QiangChen HuajunChen - 4216–4225 + 4217–4226 Link prediction is an important way to complete knowledge graphs (KGs), while embedding-based methods, effective for link prediction in KGs, perform poorly on relations that only have a few associative triples. In this work, we propose a Meta Relational Learning (MetaR) framework to do the common but challenging few-shot link prediction in KGs, namely predicting new triples about a relation by only observing a few associative triples. We solve few-shot link prediction by focusing on transferring relation-specific meta information to make model learn the most important knowledge and learn faster, corresponding to relation meta and gradient meta respectively in MetaR. Empirically, our model achieves state-of-the-art results on few-shot link prediction KG benchmarks. D19-1431 10.18653/v1/D19-1431 @@ -4994,7 +4993,7 @@ ShioriSagawa TatsunoriHashimoto PercyLiang - 4226–4236 + 4227–4237 Language models are generally trained on data spanning a wide range of topics (e.g., news, reviews, fiction), but they might be applied to an a priori unknown target distribution (e.g., restaurant reviews). In this paper, we first show that training on text outside the test distribution can degrade test performance when using standard maximum likelihood (MLE) training. To remedy this without the knowledge of the test distribution, we propose an approach which trains a model that performs well over a wide range of potential test distributions. In particular, we derive a new distributionally robust optimization (DRO) procedure which minimizes the loss of the model over the worst-case mixture of topics with sufficient overlap with the training distribution. Our approach, called topic conditional value at risk (topic CVaR), obtains a 5.5 point perplexity reduction over MLE when the language models are trained on a mixture of Yelp reviews and news and tested only on reviews. D19-1432 10.18653/v1/D19-1432 @@ -5003,7 +5002,7 @@ Unsupervised Domain Adaptation of Contextualized Embeddings for Sequence Labeling XiaochuangHan JacobEisenstein - 4237–4247 + 4238–4248 Contextualized word embeddings such as ELMo and BERT provide a foundation for strong performance across a wide range of natural language processing tasks by pretraining on large corpora of unlabeled text. However, the applicability of this approach is unknown when the target domain varies substantially from the pretraining corpus. We are specifically interested in the scenario in which labeled data is available in only a canonical source domain such as newstext, and the target domain is distinct from both the labeled and pretraining texts. To address this scenario, we propose domain-adaptive fine-tuning, in which the contextualized embeddings are adapted by masked language modeling on text from the target domain. We test this approach on sequence labeling in two challenging domains: Early Modern English and Twitter. Both domains differ substantially from existing pretraining corpora, and domain-adaptive fine-tuning yields substantial improvements over strong BERT baselines, with particularly impressive results on out-of-vocabulary words. We conclude that domain-adaptive fine-tuning offers a simple and effective approach for the unsupervised adaptation of sequence labeling to difficult new domains. D19-1433 D19-1433.Attachment.zip @@ -5014,7 +5013,7 @@ John P.Lalor HaoWu HongYu - 4248–4258 + 4249–4259 Incorporating Item Response Theory (IRT) into NLP tasks can provide valuable information about model performance and behavior. Traditionally, IRT models are learned using human response pattern (RP) data, presenting a significant bottleneck for large data sets like those required for training deep neural networks (DNNs). In this work we propose learning IRT models using RPs generated from artificial crowds of DNN models. We demonstrate the effectiveness of learning IRT models using DNN-generated data through quantitative and qualitative analyses for two NLP tasks. Parameters learned from human and machine RPs for natural language inference and sentiment analysis exhibit medium to large positive correlations. We demonstrate a use-case for latent difficulty item parameters, namely training set filtering, and show that using difficulty to sample training data outperforms baseline methods. Finally, we highlight cases where human expectation about item difficulty does not match difficulty as estimated from the machine RPs. D19-1434 D19-1434.Attachment.zip @@ -5027,7 +5026,7 @@ RasnaGoyal SabyasachiGhosh VihariPiratla - 4259–4269 + 4260–4270 We present a Parallel Iterative Edit (PIE) model for the problem of local sequence transduction arising in tasks like Grammatical error correction (GEC). Recent approaches are based on the popular encoder-decoder (ED) model for sequence to sequence learning. The ED model auto-regressively captures full dependency among output tokens but is slow due to sequential decoding. The PIE model does parallel decoding, giving up the advantage of modeling full dependency in the output, yet it achieves accuracy competitive with the ED model for four reasons: 1. predicting edits instead of tokens, 2. labeling sequences instead of generating sequences, 3. iteratively refining predictions to capture dependencies, and 4. factorizing logits over edits and their token argument to harness pre-trained language models like BERT. Experiments on tasks spanning GEC, OCR correction and spell correction demonstrate that the PIE model is an accurate and significantly faster alternative for local sequence transduction. D19-1435 D19-1435.Attachment.pdf @@ -5039,7 +5038,7 @@ FeiHuang MinlieHuang XiaoyanZhu - 4270–4280 + 4271–4281 Most of the existing generative adversarial networks (GAN) for text generation suffer from the instability of reinforcement learning training algorithms such as policy gradient, leading to unstable performance. To tackle this problem, we propose a novel framework called Adversarial Reward Augmented Maximum Likelihood (ARAML). During adversarial training, the discriminator assigns rewards to samples which are acquired from a stationary distribution near the data rather than the generator’s distribution. The generator is optimized with maximum likelihood estimation augmented by the discriminator’s rewards instead of policy gradient. Experiments show that our model can outperform state-of-the-art text GANs with a more stable training process. D19-1436 10.18653/v1/D19-1436 @@ -5051,7 +5050,7 @@ XianLi GrahamNeubig EduardHovy - 4281–4291 + 4282–4292 Most sequence-to-sequence (seq2seq) models are autoregressive; they generate each token by conditioning on previously generated tokens. In contrast, non-autoregressive seq2seq models generate all tokens in one pass, which leads to increased efficiency through parallel processing on hardware such as GPUs. However, directly modeling the joint distribution of all tokens simultaneously is challenging, and even with increasingly complex model structures accuracy lags significantly behind autoregressive models. In this paper, we propose a simple, efficient, and effective model for non-autoregressive sequence generation using latent variable models. Specifically, we turn to generative flow, an elegant technique to model complex distributions using neural networks, and design several layers of flow tailored for modeling the conditional density of sequential latent variables. We evaluate this model on three neural machine translation (NMT) benchmark datasets, achieving comparable performance with state-of-the-art non-autoregressive NMT models and almost constant decoding time w.r.t the sequence length. D19-1437 D19-1437.Attachment.zip @@ -5063,7 +5062,7 @@ LiangZhao JianyuWang JoelHestness - 4292–4301 + 4293–4302 Compositional generalization is a basic mechanism in human language learning, but current neural networks lack such ability. In this paper, we conduct fundamental research for encoding compositionality in neural networks. Conventional methods use a single representation for the input sentence, making it hard to apply prior knowledge of compositionality. In contrast, our approach leverages such knowledge with two representations, one generating attention maps, and the other mapping attended input words to output symbols. We reduce the entropy in each representation to improve generalization. Our experiments demonstrate significant improvements over the conventional methods in five NLP tasks including instruction learning and machine translation. In the SCAN domain, it boosts accuracies from 14.0% to 98.8% in Jump task, and from 92.0% to 99.7% in TurnLeft task. It also beats human performance on a few-shot learning task. We hope the proposed approach can help ease future research towards human-level compositional language learning. D19-1438 D19-1438.Attachment.zip @@ -5077,7 +5076,7 @@ YordanYordanov PhilBlunsom ThomasLukasiewicz - 4302–4311 + 4303–4312 Pronoun resolution is a major area of natural language understanding. However, large-scale training sets are still scarce, since manually labelling data is costly. In this work, we introduce WikiCREM (Wikipedia CoREferences Masked) a large-scale, yet accurate dataset of pronoun disambiguation instances. We use a language-model-based approach for pronoun resolution in combination with our WikiCREM dataset. We compare a series of models on a collection of diverse and challenging coreference resolution problems, where we match or outperform previous state-of-the-art approaches on 6 out of 7 datasets, such as GAP, DPR, WNLI, PDP, WinoBias, and WinoGender. We release our model to be used off-the-shelf for solving pronoun disambiguation. D19-1439 D19-1439.Attachment.zip @@ -5087,7 +5086,7 @@ Identifying and Explaining Discriminative Attributes ArminsStepanjans AndréFreitas - 4312–4321 + 4313–4322 Identifying what is at the center of the meaning of a word and what discriminates it from other words is a fundamental natural language inference task. This paper describes an explicit word vector representation model (WVM) to support the identification of discriminative attributes. A core contribution of the paper is a quantitative and qualitative comparative analysis of different types of data sources and Knowledge Bases in the construction of explainable and explicit WVMs: (i) knowledge graphs built from dictionary definitions, (ii) entity-attribute-relationships graphs derived from images and (iii) commonsense knowledge graphs. Using a detailed quantitative and qualitative analysis, we demonstrate that these data sources have complementary semantic aspects, supporting the creation of explicit semantic vector spaces. The explicit vector spaces are evaluated using the task of discriminative attribute identification, showing comparable performance to the state-of-the-art systems in the task (F1-score = 0.69), while delivering full model transparency and explainability. D19-1440 10.18653/v1/D19-1440 @@ -5098,7 +5097,7 @@ YuCheng ZheGan JingjingLiu - 4322–4331 + 4323–4332 Pre-trained language models such as BERT have proven to be highly effective for natural language processing (NLP) tasks. However, the high demand for computing resources in training such models hinders their application in practice. In order to alleviate this resource hunger in large-scale model training, we propose a Patient Knowledge Distillation approach to compress an original large model (teacher) into an equally-effective lightweight shallow network (student). Different from previous knowledge distillation methods, which only use the output from the last layer of the teacher network for distillation, our student model patiently learns from multiple intermediate layers of the teacher model for incremental knowledge extraction, following two strategies: (i) PKD-Last: learning from the last k layers; and (ii) PKD-Skip: learning from every k layers. These two patient distillation schemes enable the exploitation of rich information in the teacher’s hidden layers, and encourage the student model to patiently learn from and imitate the teacher through a multi-layer distillation process. Empirically, this translates into improved results on multiple NLP tasks with a significant gain in training efficiency, without sacrificing model accuracy. D19-1441 10.18653/v1/D19-1441 @@ -5107,7 +5106,7 @@ Neural <fixed-case>G</fixed-case>aussian Copula for Variational Autoencoder Prince ZizhuangWang William YangWang - 4332–4342 + 4333–4343 Variational language models seek to estimate the posterior of latent variables with an approximated variational posterior. The model often assumes the variational posterior to be factorized even when the true posterior is not. The learned variational posterior under this assumption does not capture the dependency relationships over latent variables. We argue that this would cause a typical training problem called posterior collapse observed in all other variational language models. We propose Gaussian Copula Variational Autoencoder (VAE) to avert this problem. Copula is widely used to model correlation and dependencies of high-dimensional random variables, and therefore it is helpful to maintain the dependency relationships that are lost in VAE. The empirical results show that by modeling the correlation of latent variables explicitly using a neural parametric copula, we can avert this training difficulty while getting competitive results among all other VAE approaches. D19-1442 10.18653/v1/D19-1442 @@ -5119,7 +5118,7 @@ MakotoYamada Louis-PhilippeMorency RuslanSalakhutdinov - 4343–4352 + 4344–4353 Transformer is a powerful architecture that achieves superior performance on various sequence learning tasks, including neural machine translation, language understanding, and sequence prediction. At the core of the Transformer is the attention mechanism, which concurrently processes all inputs in the streams. In this paper, we present a new formulation of attention via the lens of the kernel. To be more precise, we realize that the attention can be seen as applying kernel smoother over the inputs with the kernel scores being the similarities between inputs. This new formulation gives us a better way to understand individual components of the Transformer’s attention, such as the better way to integrate the positional embedding. Another important advantage of our kernel-based formulation is that it paves the way to a larger space of composing Transformer’s attention. As an example, we propose a new variant of Transformer’s attention which models the input as a product of symmetric kernels. This approach achieves competitive performance to the current state of the art model with less computation. In our experiments, we empirically study different kernel construction strategies on two widely used tasks: neural machine translation and sequence prediction. D19-1443 10.18653/v1/D19-1443 @@ -5129,7 +5128,7 @@ JiaweiWu WenhanXiong William YangWang - 4353–4363 + 4354–4364 Many tasks in natural language processing can be viewed as multi-label classification problems. However, most of the existing models are trained with the standard cross-entropy loss function and use a fixed prediction policy (e.g., a threshold of 0.5) for all the labels, which completely ignores the complexity and dependencies among different labels. In this paper, we propose a meta-learning method to capture these complex label dependencies. More specifically, our method utilizes a meta-learner to jointly learn the training policies and prediction policies for different labels. The training policies are then used to train the classifier with the cross-entropy loss function, and the prediction policies are further implemented for prediction. Experimental results on fine-grained entity typing and text classification demonstrate that our proposed method can obtain more accurate multi-label classification results. D19-1444 10.18653/v1/D19-1444 @@ -5140,7 +5139,7 @@ AlexeyRomanov AnnaRogers AnnaRumshisky - 4364–4373 + 4365–4374 BERT-based architectures currently give state-of-the-art performance on many NLP tasks, but little is known about the exact mechanisms that contribute to its success. In the current work, we focus on the interpretation of self-attention, which is one of the fundamental underlying components of BERT. Using a subset of GLUE tasks and a set of handcrafted features-of-interest, we propose the methodology and carry out a qualitative and quantitative analysis of the information encoded by the individual BERT’s heads. Our findings suggest that there is a limited set of attention patterns that are repeated across different heads, indicating the overall model overparametrization. While different heads consistently use the same attention patterns, they have varying impact on performance across different tasks. We show that manually disabling attention in certain heads leads to a performance improvement over the regular fine-tuned BERT models. D19-1445 D19-1445.Attachment.zip @@ -5155,7 +5154,7 @@ TaoQin JianhuangLai Tie-YanLiu - 4374–4383 + 4375–4384 Neural machine translation, which achieves near human-level performance in some languages, strongly relies on the large amounts of parallel sentences, which hinders its applicability to low-resource language pairs. Recent works explore the possibility of unsupervised machine translation with monolingual data only, leading to much lower accuracy compared with the supervised one. Observing that weakly paired bilingual documents are much easier to collect than bilingual sentences, e.g., from Wikipedia, news websites or books, in this paper, we investigate training translation models with weakly paired bilingual documents. Our approach contains two components. 1) We provide a simple approach to mine implicitly bilingual sentence pairs from document pairs which can then be used as supervised training signals. 2) We leverage the topic consistency of two weakly paired documents and learn the sentence translation model by constraining the word distribution-level alignments. We evaluate our method on weakly paired documents from Wikipedia on six tasks, the widely used WMT16 German\leftrightarrowEnglish, WMT13 Spanish\leftrightarrowEnglish and WMT16 Romanian\leftrightarrowEnglish translation tasks. We obtain 24.1/30.3, 28.1/27.6 and 30.1/27.6 BLEU points separately, outperforming previous results by more than 5 BLEU points in each direction and reducing the gap between unsupervised translation and supervised translation up to 50%. D19-1446 10.18653/v1/D19-1446 @@ -5165,7 +5164,7 @@ JasonLee KyunghyunCho DouweKiela - 4384–4394 + 4385–4395 Emergent multi-agent communication protocols are very different from natural language and not easily interpretable by humans. We find that agents that were initially pretrained to produce natural language can also experience detrimental language drift: when a non-linguistic reward is used in a goal-based task, e.g. some scalar success metric, the communication protocol may easily and radically diverge from natural language. We recast translation as a multi-agent communication game and examine auxiliary training constraints for their effectiveness in mitigating language drift. We show that a combination of syntactic (language model likelihood) and semantic (visual grounding) constraints gives the best communication performance, allowing pre-trained agents to retain English syntax while learning to accurately convey the intended meaning. D19-1447 10.18653/v1/D19-1447 @@ -5175,7 +5174,7 @@ ElenaVoita RicoSennrich IvanTitov - 4395–4405 + 4396–4406 We seek to understand how the representations of individual tokens and the structure of the learned feature space evolve between layers in deep neural networks under different learning objectives. We chose the Transformers for our analysis as they have been shown effective with various tasks, including machine translation (MT), standard left-to-right language models (LM) and masked language modeling (MLM). Previous work used black-box probing tasks to show that the representations learned by the Transformer differ significantly depending on the objective. In this work, we use canonical correlation analysis and mutual information estimators to study how information flows across Transformer layers and observe that the choice of the objective determines this process. For example, as you go from bottom to top layers, information about the past in left-to-right language models gets vanished and predictions about the future get formed. In contrast, for MLM, representations initially acquire information about the context around the token, partially forgetting the token identity and producing a more generalized token representation. The token identity then gets recreated at the top MLM layers. D19-1448 D19-1448.Attachment.zip @@ -5187,7 +5186,7 @@ GoranGlavaš RoiReichart AnnaKorhonen - 4406–4417 + 4407–4418 Recent efforts in cross-lingual word embedding (CLWE) learning have predominantly focused on fully unsupervised approaches that project monolingual embeddings into a shared cross-lingual space without any cross-lingual signal. The lack of any supervision makes such approaches conceptually attractive. Yet, their only core difference from (weakly) supervised projection-based CLWE methods is in the way they obtain a seed dictionary used to initialize an iterative self-learning procedure. The fully unsupervised methods have arguably become more robust, and their primary use case is CLWE induction for pairs of resource-poor and distant languages. In this paper, we question the ability of even the most robust unsupervised CLWE approaches to induce meaningful CLWEs in these more challenging settings. A series of bilingual lexicon induction (BLI) experiments with 15 diverse languages (210 language pairs) show that fully unsupervised CLWE methods still fail for a large number of language pairs (e.g., they yield zero BLI performance for 87/210 pairs). Even when they succeed, they never surpass the performance of weakly supervised methods (seeded with 500-1,000 translation pairs) using the same self-learning procedure in any BLI setup, and the gaps are often substantial. These findings call for revisiting the main motivations behind fully unsupervised CLWE methods. D19-1449 D19-1449.Attachment.zip @@ -5198,7 +5197,7 @@ HaozhouWang JamesHenderson PaolaMerlo - 4418–4429 + 4419–4430 Distributed representations of words which map each word to a continuous vector have proven useful in capturing important linguistic information not only in a single language but also across different languages. Current unsupervised adversarial approaches show that it is possible to build a mapping matrix that aligns two sets of monolingual word embeddings without high quality parallel data, such as a dictionary or a sentence-aligned corpus. However, without an additional step of refinement, the preliminary mapping learnt by these methods is unsatisfactory, leading to poor performance for typologically distant languages. In this paper, we propose a weakly-supervised adversarial training method to overcome this limitation, based on the intuition that mapping across languages is better done at the concept level than at the word level. We propose a concept-based adversarial training method which improves the performance of previous unsupervised adversarial methods for most languages, and especially for typologically distant language pairs. D19-1450 10.18653/v1/D19-1450 @@ -5211,7 +5210,7 @@ WeiLu JimmyLin XuSun - 4430–4440 + 4431–4441 Multilingual knowledge graphs (KGs), such as YAGO and DBpedia, represent entities in different languages. The task of cross-lingual entity alignment is to match entities in a source language with their counterparts in target languages. In this work, we investigate embedding-based approaches to encode entities from multilingual KGs into the same vector space, where equivalent entities are close to each other. Specifically, we apply graph convolutional networks (GCNs) to combine multi-aspect information of entities, including topological connections, relations, and attributes of entities, to learn entity embeddings. To exploit the literal descriptions of entities expressed in different languages, we propose two uses of a pretrained multilingual BERT model to bridge cross-lingual gaps. We further propose two strategies to integrate GCN-based and BERT-based modules to boost performance. Extensive experiments on two benchmark datasets demonstrate that our method significantly outperforms existing systems. D19-1451 10.18653/v1/D19-1451 @@ -5221,7 +5220,7 @@ MitraMohtarami JamesGlass PreslavNakov - 4441–4451 + 4442–4452 We study cross-lingual stance detection, which aims to leverage labeled data in one language to identify the relative perspective (or stance) of a given document with respect to a claim in a different target language. In particular, we introduce a novel contrastive language adaptation approach applied to memory networks, which ensures accurate alignment of stances in the source and target languages, and can effectively deal with the challenge of limited labeled data in the target language. The evaluation results on public benchmark datasets and comparison against current state-of-the-art approaches demonstrate the effectiveness of our approach. D19-1452 10.18653/v1/D19-1452 @@ -5232,7 +5231,7 @@ StephanPeitz UdhyakumarNallasamy MatthiasPaulik - 4452–4461 + 4453–4462 The state of the art in machine translation (MT) is governed by neural approaches, which typically provide superior translation accuracy over statistical approaches. However, on the closely related task of word alignment, traditional statistical word alignment models often remain the go-to solution. In this paper, we present an approach to train a Transformer model to produce both accurate translations and alignments. We extract discrete alignments from the attention probabilities learnt during regular neural machine translation model training and leverage them in a multi-task framework to optimize towards translation and alignment objectives. We demonstrate that our approach produces competitive results compared to GIZA++ trained IBM alignment models without sacrificing translation accuracy and outperforms previous attempts on Transformer model based word alignment. Finally, by incorporating IBM model alignments into our multi-task training, we report significantly better alignment accuracies compared to GIZA++ on three publicly available data sets. D19-1453 D19-1453.Attachment.zip @@ -5245,7 +5244,7 @@ DerekChen RonanLe Bras YejinChoi - 4462–4472 + 4463–4473 We introduce Social IQa, the first large-scale benchmark for commonsense reasoning about social situations. Social IQa contains 38,000 multiple choice questions for probing emotional and social intelligence in a variety of everyday situations (e.g., Q: “Jordan wanted to tell Tracy a secret, so Jordan leaned towards Tracy. Why did Jordan do this?” A: “Make sure no one else could hear”). Through crowdsourcing, we collect commonsense questions along with correct and incorrect answers about social interactions, using a new framework that mitigates stylistic artifacts in incorrect answers by asking workers to provide the right answer to a different but related question. Empirical results show that our benchmark is challenging for existing question-answering models based on pretrained language models, compared to human performance (>20% gap). Notably, we further establish Social IQa as a resource for transfer learning of commonsense knowledge, achieving state-of-the-art performance on multiple commonsense reasoning tasks (Winograd Schemas, COPA). D19-1454 10.18653/v1/D19-1454 @@ -5254,7 +5253,7 @@ Self-Assembling Modular Networks for Interpretable Multi-Hop Reasoning YichenJiang MohitBansal - 4473–4483 + 4474–4484 Multi-hop QA requires a model to connect multiple pieces of evidence scattered in a long context to answer the question. The recently proposed HotpotQA (Yang et al., 2018) dataset is comprised of questions embodying four different multi-hop reasoning paradigms (two bridge entity setups, checking multiple properties, and comparing two entities), making it challenging for a single neural network to handle all four. In this work, we present an interpretable, controller-based Self-Assembling Neural Modular Network (Hu et al., 2017, 2018) for multi-hop reasoning, where we design four novel modules (Find, Relocate, Compare, NoOp) to perform unique types of language reasoning. Based on a question, our layout controller RNN dynamically infers a series of reasoning modules to construct the entire network. Empirically, we show that our dynamic, multi-hop modular network achieves significant improvements over the static, single-hop baseline (on both regular and adversarial evaluation). We further demonstrate the interpretability of our model via three analyses. First, the controller can softly decompose the multi-hop question into multiple single-hop sub-questions to promote compositional reasoning behavior of the main network. Second, the controller can predict layouts that conform to the layouts designed by human experts. Finally, the intermediate module can infer the entity that connects two distantly-located supporting facts by addressing the sub-question from the controller. D19-1455 10.18653/v1/D19-1455 @@ -5265,7 +5264,7 @@ IgnacioCases LauriKarttunen ChristopherPotts - 4484–4494 + 4485–4495 Deep learning models for semantics are generally evaluated using naturalistic corpora. Adversarial testing methods, in which models are evaluated on new examples with known semantic properties, have begun to reveal that good performance at these naturalistic tasks can hide serious shortcomings. However, we should insist that these evaluations be fair – that the models are given data sufficient to support the requisite kinds of generalization. In this paper, we define and motivate a formal notion of fairness in this sense. We then apply these ideas to natural language inference by constructing very challenging but provably fair artificial datasets and showing that standard neural models fail to generalize in the required ways; only task-specific models that jointly compose the premise and hypothesis are able to achieve high performance, and even these models do not solve the task perfectly. D19-1456 D19-1456.Attachment.zip @@ -5278,7 +5277,7 @@ AntoineBosselut Wen-tauYih PeterClark - 4495–4504 + 4496–4505 Our goal is to better comprehend procedural text, e.g., a paragraph about photosynthesis, by not only predicting what happens, but *why* some actions need to happen before others. Our approach builds on a prior process comprehension framework for predicting actions’ effects, to also identify subsequent steps that those effects enable. We present our new model (XPAD) that biases effect predictions towards those that (1) explain more of the actions in the paragraph and (2) are more plausible with respect to background knowledge. We also extend an existing benchmark dataset for procedural text comprehension, ProPara, by adding the new task of explaining actions by predicting their dependencies. We find that XPAD significantly outperforms prior systems on this task, while maintaining the performance on the original task in ProPara. The dataset is available at http://data.allenai.org/propara D19-1457 10.18653/v1/D19-1457 @@ -5290,7 +5289,7 @@ JinDong JoellePineau William L.Hamilton - 4505–4514 + 4506–4515 The recent success of natural language understanding (NLU) systems has been troubled by results highlighting the failure of these models to generalize in a systematic and robust way. In this work, we introduce a diagnostic benchmark suite, named CLUTRR, to clarify some key issues related to the robustness and systematicity of NLU systems. Motivated by the classic work on inductive logic programming, CLUTRR requires that an NLU system infer kinship relations between characters in short stories. Successful performance on this task requires both extracting relationships between entities, as well as inferring the logical rules governing these relationships. CLUTRR allows us to precisely measure a model’s ability for systematic generalization by evaluating on held-out combinations of logical rules, and allows us to evaluate a model’s robustness by adding curated noise facts. Our empirical results highlight a substantial performance gap between state-of-the-art NLU models (e.g., BERT and MAC) and a graph neural network model that works directly with symbolic inputs—with the graph-based model exhibiting both stronger generalization and greater robustness. D19-1458 D19-1458.Attachment.zip @@ -5308,7 +5307,7 @@ AmitDubey Kyu-YoungKim AndyCedilnik - 4515–4524 + 4516–4525 A significant barrier to progress in data-driven approaches to building dialog systems is the lack of high quality, goal-oriented conversational data. To help satisfy this elementary requirement, we introduce the initial release of the Taskmaster-1 dataset which includes 13,215 task-based dialogs comprising six domains. Two procedures were used to create this collection, each with unique advantages. The first involves a two-person, spoken “Wizard of Oz” (WOz) approach in which trained agents and crowdsourced workers interact to complete the task while the second is “self-dialog” in which crowdsourced workers write the entire dialog themselves. We do not restrict the workers to detailed scripts or to a small knowledge base and hence we observe that our dataset contains more realistic and diverse conversations in comparison to existing datasets. We offer several baseline models including state of the art neural seq2seq architectures with benchmark performance as well as qualitative human evaluations. Dialogs are labeled with API calls and arguments, a simple and cost effective approach which avoids the requirement of complex annotation schema. The layer of abstraction between the dialog model and the service provider API allows for a given model to interact with multiple services that provide similar functionally. Finally, the dataset will evoke interest in written vs. spoken language, discourse patterns, error handling and other linguistic phenomena related to dialog system research, development and design. D19-1459 10.18653/v1/D19-1459 @@ -5322,7 +5321,7 @@ YiZhang AdelYoussef MonaDiab - 4525–4535 + 4526–4536 The need for high-quality, large-scale, goal-oriented dialogue datasets continues to grow as virtual assistants become increasingly wide-spread. However, publicly available datasets useful for this area are limited either in their size, linguistic diversity, domain coverage, or annotation granularity. In this paper, we present strategies toward curating and annotating large scale goal oriented dialogue data. We introduce the MultiDoGO dataset to overcome these limitations. With a total of over 81K dialogues harvested across six domains, MultiDoGO is over 8 times the size of MultiWOZ, the other largest comparable dialogue dataset currently available to the public. Over 54K of these harvested conversations are annotated for intent classes and slot labels. We adopt a Wizard-of-Oz approach wherein a crowd-sourced worker (the “customer”) is paired with a trained annotator (the “agent”). The data curation process was controlled via biases to ensure a diversity in dialogue flows following variable dialogue policies. We provide distinct class label tags for agents vs. customer utterances, along with applicable slot labels. We also compare and contrast our strategies on annotation granularity, i.e. turn vs. sentence level. Furthermore, we compare and contrast annotations curated by leveraging professional annotators vs the crowd. We believe our strategies for eliciting and annotating such a dialogue dataset scales across modalities and domains and potentially languages in the future. To demonstrate the efficacy of our devised strategies we establish neural baselines for classification on the agent and customer utterances as well as slot labeling for each domain. D19-1460 D19-1460.Attachment.pdf @@ -5334,7 +5333,7 @@ SamuelHumeau BharathChintagunta JasonWeston - 4536–4545 + 4537–4546 The detection of offensive language in the context of a dialogue has become an increasingly important application of natural language processing. The detection of trolls in public forums (Galán-García et al., 2016), and the deployment of chatbots in the public domain (Wolf et al., 2017) are two examples that show the necessity of guarding against adversarially offensive behavior on the part of humans. In this work, we develop a training scheme for a model to become robust to such human attacks by an iterative build it, break it, fix it scheme with humans and models in the loop. In detailed experiments we show this approach is considerably more robust than previous systems. Further, we show that offensive language used within a conversation critically depends on the dialogue context, and cannot be viewed as a single sentence offensive detection task as in most previous work. Our newly collected tasks and methods are all made open source and publicly available. D19-1461 D19-1461.Attachment.zip @@ -5346,7 +5345,7 @@ DeyiXiong BonnieWebber ChangjianHu - 4546–4556 + 4547–4557 Ellipsis and co-reference are common and ubiquitous especially in multi-turn dialogues. In this paper, we treat the resolution of ellipsis and co-reference in dialogue as a problem of generating omitted or referred expressions from the dialogue context. We therefore propose a unified end-to-end Generative Ellipsis and CO-reference Resolution model (GECOR) in the context of dialogue. The model can generate a new pragmatically complete user utterance by alternating the generation and copy mode for each user utterance. A multi-task learning framework is further proposed to integrate the GECOR into an end-to-end task-oriented dialogue. In order to train both the GECOR and the multi-task learning framework, we manually construct a new dataset on the basis of the public dataset CamRest676 with both ellipsis and co-reference annotation. On this dataset, intrinsic evaluations on the resolution of ellipsis and co-reference show that the GECOR model significantly outperforms the sequence-to-sequence (seq2seq) baseline model in terms of EM, BLEU and F1 while extrinsic evaluations on the downstream dialogue task demonstrate that our multi-task learning framework with GECOR achieves a higher success rate of task completion than TSCP, a state-of-the-art end-to-end task-oriented dialogue model. D19-1462 10.18653/v1/D19-1462 @@ -5358,7 +5357,7 @@ FengJi HaiqingChen YinZhang - 4557–4566 + 4558–4567 How to incorporate external knowledge into a neural dialogue model is critically important for dialogue systems to behave like real humans. To handle this problem, memory networks are usually a great choice and a promising way. However, existing memory networks do not perform well when leveraging heterogeneous information from different sources. In this paper, we propose a novel and versatile external memory networks called Heterogeneous Memory Networks (HMNs), to simultaneously utilize user utterances, dialogue history and background knowledge tuples. In our method, historical sequential dialogues are encoded and stored into the context-aware memory enhanced by gating mechanism while grounding knowledge tuples are encoded and stored into the context-free memory. During decoding, the decoder augmented with HMNs recurrently selects each word in one response utterance from these two memories and a general vocabulary. Experimental results on multiple real-world datasets show that HMNs significantly outperform the state-of-the-art data-driven task-oriented dialogue models in most domains. D19-1463 10.18653/v1/D19-1463 @@ -5368,8 +5367,8 @@ ChenZhang QiuchiLi DaweiSong - 4567–4577 - Due to their inherent capability in semantic alignment of aspects and their context words, attention mechanism and Convolutional Neural Networks (CNNs) are widely applied for aspect-based sentiment classification. However, these models lack a mechanism to account for relevant syntactical constraints and long-range word dependencies, and hence may mistakenly recognize syntactically irrelevant contextual words as clues for judging aspect sentiment. To tackle this problem, we propose to build a Graph Convolutional Network (GCN) over the dependency tree of a sentence to exploit syntactical information and word dependencies. Based on it, a novel aspect-specific sentiment classification framework is raised. Experiments on three benchmarking collections illustrate that our proposed model has comparable effectiveness to a range of state-of-the-art modelsCode and preprocessed datasets are available at https://github.com/GeneZC/ASGCNhttps://github.com/GeneZC/ASGCN., and further demonstrate that both syntactical information and long-range word dependencies are properly captured by the graph convolution structure. + 4568–4578 + Due to their inherent capability in semantic alignment of aspects and their context words, attention mechanism and Convolutional Neural Networks (CNNs) are widely applied for aspect-based sentiment classification. However, these models lack a mechanism to account for relevant syntactical constraints and long-range word dependencies, and hence may mistakenly recognize syntactically irrelevant contextual words as clues for judging aspect sentiment. To tackle this problem, we propose to build a Graph Convolutional Network (GCN) over the dependency tree of a sentence to exploit syntactical information and word dependencies. Based on it, a novel aspect-specific sentiment classification framework is raised. Experiments on three benchmarking collections illustrate that our proposed model has comparable effectiveness to a range of state-of-the-art models, and further demonstrate that both syntactical information and long-range word dependencies are properly captured by the graph convolution structure. D19-1464 10.18653/v1/D19-1464 @@ -5381,7 +5380,7 @@ LingzhiWang XixinWu Kam-FaiWong - 4578–4588 + 4579–4589 Aspect words, indicating opinion targets, are essential in expressing and understanding human opinions. To identify aspects, most previous efforts focus on using sequence tagging models trained on human-annotated data. This work studies unsupervised aspect extraction and explores how words appear in global context (on sentence level) and local context (conveyed by neighboring words). We propose a novel neural model, capable of coupling global and local representation to discover aspect words. Experimental results on two benchmarks, laptop and restaurant reviews, show that our model significantly outperforms the state-of-the-art models from previous studies evaluated with varying metrics. Analysis on model output show our ability to learn meaningful and coherent aspect representations. We further investigate how words distribute in global and local context, and find that aspect and non-aspect words do exhibit different context, interpreting our superiority in unsupervised aspect extraction. D19-1465 10.18653/v1/D19-1465 @@ -5394,7 +5393,7 @@ LidongBing YuZhang QiangYang - 4589–4599 + 4590–4600 Joint extraction of aspects and sentiments can be effectively formulated as a sequence labeling problem. However, such formulation hinders the effectiveness of supervised methods due to the lack of annotated sequence data in many domains. To address this issue, we firstly explore an unsupervised domain adaptation setting for this task. Prior work can only use common syntactic relations between aspect and opinion words to bridge the domain gaps, which highly relies on external linguistic resources. To resolve it, we propose a novel Selective Adversarial Learning (SAL) method to align the inferred correlation vectors that automatically capture their latent relations. The SAL method can dynamically learn an alignment weight for each word such that more important words can possess higher alignment weights to achieve fine-grained (word-level) adaptation. Empirically, extensive experiments demonstrate the effectiveness of the proposed SAL method. D19-1466 D19-1466.Attachment.pdf @@ -5409,7 +5408,7 @@ ZhongSu RenhongCheng XiaoweiShen - 4600–4609 + 4601–4610 Aspect level sentiment classification is a fine-grained sentiment analysis task. To detect the sentiment towards a particular aspect in a sentence, previous studies have developed various attention-based methods for generating aspect-specific sentence representations. However, the attention may inherently introduce noise and downgrade the performance. In this paper, we propose constrained attention networks (CAN), a simple yet effective solution, to regularize the attention for multi-aspect sentiment analysis, which alleviates the drawback of the attention mechanism. Specifically, we introduce orthogonal regularization on multiple aspects and sparse regularization on each single aspect. Experimental results on two public datasets demonstrate the effectiveness of our approach. We further extend our approach to multi-task settings and outperform the state-of-the-art methods. D19-1467 D19-1467.Attachment.zip @@ -5420,7 +5419,7 @@ GiannisKaramanolakis DanielHsu LuisGravano - 4610–4620 + 4611–4621 User-generated reviews can be decomposed into fine-grained segments (e.g., sentences, clauses), each evaluating a different aspect of the principal entity (e.g., price, quality, appearance). Automatically detecting these aspects can be useful for both users and downstream opinion mining applications. Current supervised approaches for learning aspect classifiers require many fine-grained aspect labels, which are labor-intensive to obtain. And, unfortunately, unsupervised topic models often fail to capture the aspects of interest. In this work, we consider weakly supervised approaches for training aspect classifiers that only require the user to provide a small set of seed words (i.e., weakly positive indicators) for the aspects of interest. First, we show that current weakly supervised approaches fail to leverage the predictive power of seed words for aspect detection. Next, we propose a student-teacher approach that effectively leverages seed words in a bag-of-words classifier (teacher); in turn, we use the teacher to train a second model (student) that is potentially more powerful (e.g., a neural network that uses pre-trained word embeddings). Finally, we show that iterative co-training can be used to cope with noisy seed words, leading to both improved teacher and student models. Our proposed approach consistently outperforms previous weakly supervised approaches (by 14.1 absolute F1 points on average) in six different domains of product reviews and six multilingual datasets of restaurant reviews. D19-1468 D19-1468.Attachment.zip @@ -5434,7 +5433,7 @@ XiaoLin DanJurafsky AjayDivakaran - 4621–4631 + 4622–4632 Computing author intent from multimodal data like Instagram posts requires modeling a complex relationship between text and image. For example, a caption might evoke an ironic contrast with the image, so neither caption nor image is a mere transcript of the other. Instead they combine—via what has been called meaning multiplication (Bateman et al.)- to create a new meaning that has a more complex relation to the literal meanings of text and image. Here we introduce a multimodal dataset of 1299 Instagram posts labeled for three orthogonal taxonomies: the authorial intent behind the image-caption pair, the contextual relationship between the literal meanings of the image and caption, and the semiotic relationship between the signified meanings of the image and caption. We build a baseline deep multimodal classifier to validate the taxonomy, showing that employing both text and image improves intent detection by 9.6 compared to using only the image modality, demonstrating the commonality of non-intersective meaning multiplication. The gain with multimodality is greatest when the image and caption diverge semiotically. Our dataset offers a new resource for the study of the rich meanings that result from pairing text and image. D19-1469 10.18653/v1/D19-1469 @@ -5445,7 +5444,7 @@ JingLi LuWang Kam-FaiWong - 4632–4642 + 4633–4643 The prevalent use of social media leads to a vast amount of online conversations being produced on a daily basis. It presents a concrete challenge for individuals to better discover and engage in social media discussions. In this paper, we present a novel framework to automatically recommend conversations to users based on their prior conversation behaviors. Built on neural collaborative filtering, our model explores deep semantic features that measure how a user’s preferences match an ongoing conversation’s context. Furthermore, to identify salient characteristics from interleaving user interactions, our model incorporates graph-structured networks, where both replying relations and temporal features are encoded as conversation context. Experimental results on two large-scale datasets collected from Twitter and Reddit show that our model yields better performance than previous state-of-the-art models, which only utilize lexical features and ignore past user interactions in the conversations. D19-1470 10.18653/v1/D19-1470 @@ -5457,7 +5456,7 @@ HaolinJin AmbreenNazir LingSun - 4643–4652 + 4644–4653 Recently, neural networks based on multi-task learning have achieved promising performance on fake news detection, which focuses on learning shared features among tasks as complementarity features to serve different tasks. However, in most of the existing approaches, the shared features are completely assigned to different tasks without selection, which may lead to some useless and even adverse features integrated into specific tasks. In this paper, we design a sifted multi-task learning method with a selected sharing layer for fake news detection. The selected sharing layer adopts gate mechanism and attention mechanism to filter and select shared feature flows between tasks. Experiments on two public and widely used competition datasets, i.e. RumourEval and PHEME, demonstrate that our proposed method achieves the state-of-the-art performance and boosts the F1-score by more than 0.87%, 1.31%, respectively. D19-1471 10.18653/v1/D19-1471 @@ -5468,7 +5467,7 @@ RenatoFerreira Pinto Junior GraemeHirst YangXu - 4653–4662 + 4654–4663 We present a text-based framework for investigating moral sentiment change of the public via longitudinal corpora. Our framework is based on the premise that language use can inform people’s moral perception toward right or wrong, and we build our methodology by exploring moral biases learned from diachronic word embeddings. We demonstrate how a parameter-free model supports inference of historical shifts in moral sentiment toward concepts such as slavery and democracy over centuries at three incremental levels: moral relevance, moral polarity, and fine-grained moral dimensions. We apply this methodology to visualizing moral time courses of individual concepts and analyzing the relations between psycholinguistic variables and rates of moral sentiment change at scale. Our work offers opportunities for applying natural language processing toward characterizing moral sentiment change in society. D19-1472 D19-1472.Attachment.zip @@ -5479,7 +5478,7 @@ BeiYu YingyaLi JunWang - 4663–4673 + 4664–4674 Causal interpretation of correlational findings from observational studies has been a major type of misinformation in science communication. Prior studies on identifying inappropriate use of causal language relied on manual content analysis, which is not scalable for examining a large volume of science publications. In this study, we first annotated a corpus of over 3,000 PubMed research conclusion sentences, then developed a BERT-based prediction model that classifies conclusion sentences into “no relationship”, “correlational”, “conditional causal”, and “direct causal” categories, achieving an accuracy of 0.90 and a macro-F1 of 0.88. We then applied the prediction model to measure the causal language use in the research conclusions of about 38,000 observational studies in PubMed. The prediction result shows that 21.7% studies used direct causal language exclusively in their conclusions, and 32.4% used some direct causal language. We also found that the ratio of causal language use differs among authors from different countries, challenging the notion of a shared consensus on causal language use in the global science community. Our prediction model could also be used to help identify the inappropriate use of causal language in science publications. D19-1473 10.18653/v1/D19-1473 @@ -5491,7 +5490,7 @@ HongmingZhang YangqiuSong Dit-YanYeung - 4674–4683 + 4675–4684 Current research on hate speech analysis is typically oriented towards monolingual and single classification tasks. In this paper, we present a new multilingual multi-aspect hate speech analysis dataset and use it to test the current state-of-the-art multilingual multitask learning approaches. We evaluate our dataset in various classification settings, then we discuss how to leverage our annotations in order to improve hate speech detection and classification in general. D19-1474 10.18653/v1/D19-1474 @@ -5505,7 +5504,7 @@ CasperHansen ChristianHansen Jakob GrueSimonsen - 4684–4696 + 4685–4697 We contribute the largest publicly available dataset of naturally occurring factual claims for the purpose of automatic claim verification. It is collected from 26 fact checking websites in English, paired with textual sources and rich metadata, and labelled for veracity by human expert journalists. We present an in-depth analysis of the dataset, highlighting characteristics and challenges. Further, we present results for automatic veracity prediction, both with established baselines and with a novel method for joint ranking of evidence pages and predicting veracity that outperforms all baselines. Significant performance increases are achieved by encoding evidence, and by modelling metadata. Our best-performing model achieves a Macro F1 of 49.2%, showing that this is a challenging testbed for claim veracity prediction. D19-1475 10.18653/v1/D19-1475 @@ -5516,7 +5515,7 @@ QinliangSu XiaojunQuan WeijiaZhang - 4697–4705 + 4698–4706 Textual network embeddings aim to learn a low-dimensional representation for every node in the network so that both the structural and textual information from the networks can be well preserved in the representations. Traditionally, the structural and textual embeddings were learned by models that rarely take the mutual influences between them into account. In this paper, a deep neural architecture is proposed to effectively fuse the two kinds of informations into one representation. The novelties of the proposed architecture are manifested in the aspects of a newly defined objective function, the complementary information fusion method for structural and textual features, and the mutual gate mechanism for textual feature extraction. Experimental results show that the proposed model outperforms the comparing methods on all three datasets. D19-1476 10.18653/v1/D19-1476 @@ -5527,7 +5526,7 @@ DiegoMarcheggiani SabineSchulte im Walde RaquelFernández - 4706–4716 + 4707–4717 Information about individuals can help to better understand what they say, particularly in social media where texts are short. Current approaches to modelling social media users pay attention to their social connections, but exploit this information in a static way, treating all connections uniformly. This ignores the fact, well known in sociolinguistics, that an individual may be part of several communities which are not equally relevant in all communicative situations. We present a model based on Graph Attention Networks that captures this observation. It dynamically explores the social graph of a user, computes a user representation given the most relevant connections for a target task, and combines it with linguistic information to make a prediction. We apply our model to three different tasks, evaluate it against alternative models, and analyse the results extensively, showing that it significantly outperforms other current methods. D19-1477 D19-1477.Attachment.pdf @@ -5539,7 +5538,7 @@ BareaSinno AlexRosenfeld Junyi JessyLi - 4717–4729 + 4718–4730 Insightful findings in political science often require researchers to analyze documents of a certain subject or type, yet these documents are usually contained in large corpora that do not distinguish between pertinent and non-pertinent documents. In contrast, we can find corpora that label relevant documents but have limitations (e.g., from a single source or era), preventing their use for political science research. To bridge this gap, we present adaptive ensembling, an unsupervised domain adaptation framework, equipped with a novel text classification model and time-aware training to ensure our methods work well with diachronic corpora. Experiments on an expert-annotated dataset show that our framework outperforms strong benchmarks. Further analysis indicates that our methods are more stable, learn better representations, and extract cleaner corpora for fine-grained analysis. D19-1478 D19-1478.Attachment.pdf @@ -5549,7 +5548,7 @@ A Hierarchical Location Prediction Neural Network for Twitter User Geolocation BinxuanHuang KathleenCarley - 4731–4741 + 4732–4742 Accurate estimation of user location is important for many online services. Previous neural network based methods largely ignore the hierarchical structure among locations. In this paper, we propose a hierarchical location prediction neural network for Twitter user geolocation. Our model first predicts the home country for a user, then uses the country result to guide the city-level prediction. In addition, we employ a character-aware word embedding layer to overcome the noisy information in tweets. With the feature fusion layer, our model can accommodate various feature combinations and achieves state-of-the-art results over three commonly used benchmarks under different feature settings. It not only improves the prediction accuracy but also greatly reduces the mean error distance. D19-1480 10.18653/v1/D19-1480 @@ -5558,7 +5557,7 @@ Trouble on the Horizon: Forecasting the Derailment of Online Conversations as they Develop Jonathan P.Chang CristianDanescu-Niculescu-Mizil - 4742–4753 + 4743–4754 Online discussions often derail into toxic exchanges between participants. Recent efforts mostly focused on detecting antisocial behavior after the fact, by analyzing single comments in isolation. To provide more timely notice to human moderators, a system needs to preemptively detect that a conversation is heading towards derailment before it actually turns toxic. This means modeling derailment as an emerging property of a conversation rather than as an isolated utterance-level event. Forecasting emerging conversational properties, however, poses several inherent modeling challenges. First, since conversations are dynamic, a forecasting model needs to capture the flow of the discussion, rather than properties of individual comments. Second, real conversations have an unknown horizon: they can end or derail at any time; thus a practical forecasting model needs to assess the risk in an online fashion, as the conversation develops. In this work we introduce a conversational forecasting model that learns an unsupervised representation of conversational dynamics and exploits it to predict future derailment as the conversation develops. By applying this model to two new diverse datasets of online conversations with labels for antisocial events, we show that it outperforms state-of-the-art systems at forecasting derailment. D19-1481 10.18653/v1/D19-1481 @@ -5570,7 +5569,7 @@ YinyinLiu ElizabethBelding William YangWang - 4754–4763 + 4755–4764 Countering online hate speech is a critical yet challenging task, but one which can be aided by the use of Natural Language Processing (NLP) techniques. Previous research has primarily focused on the development of NLP methods to automatically and effectively detect online hate speech while disregarding further action needed to calm and discourage individuals from using hate speech in the future. In addition, most existing hate speech datasets treat each post as an isolated instance, ignoring the conversational context. In this paper, we propose a novel task of generative hate speech intervention, where the goal is to automatically generate responses to intervene during online conversations that contain hate speech. As a part of this work, we introduce two fully-labeled large-scale hate speech intervention datasets collected from Gab and Reddit. These datasets provide conversation segments, hate speech labels, as well as intervention responses written by Mechanical Turk Workers. In this paper, we also analyze the datasets to understand the common intervention strategies and explore the performance of common automatic response generation methods on these new datasets to provide a benchmark for future research. D19-1482 10.18653/v1/D19-1482 @@ -5582,7 +5581,7 @@ DesmondPatton CharlotteSelous KathyMcKeown - 4764–4774 + 4765–4775 Gang-involved youth in cities such as Chicago sometimes post on social media to express their aggression towards rival gangs and previous research has demonstrated that a deep learning approach can predict aggression and loss in posts. To address the possibility of bias in this sensitive application, we developed an approach to systematically interpret the state of the art model. We found, surprisingly, that it frequently bases its predictions on stop words such as “a” or “on”, an approach that could harm social media users who have no aggressive intentions. To tackle this bias, domain experts annotated the rationales, highlighting words that explain why a tweet is labeled as “aggression”. These new annotations enable us to quantitatively measure how justified the model predictions are, and build models that drastically reduce bias. Our study shows that in high stake scenarios, accuracy alone cannot guarantee a good system and we need new evaluation methods. D19-1483 D19-1483.Attachment.pdf @@ -5593,7 +5592,7 @@ EllaRabinovich MasihSultani SuzanneStevenson - 4775–4785 + 4776–4786 In contrast to many decades of research on oral code-switching, the study of written multilingual productions has only recently enjoyed a surge of interest. Many open questions remain regarding the sociolinguistic underpinnings of written code-switching, and progress has been limited by a lack of suitable resources. We introduce a novel, large, and diverse dataset of written code-switched productions, curated from topical threads of multiple bilingual communities on the Reddit discussion platform, and explore questions that were mainly addressed in the context of spoken language thus far. We investigate whether findings in oral code-switching concerning content and style, as well as speaker proficiency, are carried over into written code-switching in discussion forums. The released dataset can further facilitate a range of research and practical activities. D19-1484 D19-1484.Attachment.zip @@ -5604,7 +5603,7 @@ PenghuiWei NanXu WenjiMao - 4786–4797 + 4787–4798 Automatically verifying rumorous information has become an important and challenging task in natural language processing and social media analytics. Previous studies reveal that people’s stances towards rumorous messages can provide indicative clues for identifying the veracity of rumors, and thus determining the stances of public reactions is a crucial preceding step for rumor veracity prediction. In this paper, we propose a hierarchical multi-task learning framework for jointly predicting rumor stance and veracity on Twitter, which consists of two components. The bottom component of our framework classifies the stances of tweets in a conversation discussing a rumor via modeling the structural property based on a novel graph convolutional network. The top component predicts the rumor veracity by exploiting the temporal dynamics of stance evolution. Experimental results on two benchmark datasets show that our method outperforms previous methods in both rumor stance classification and veracity prediction. D19-1485 10.18653/v1/D19-1485 @@ -5618,7 +5617,7 @@ QimaiLi Xiao-MingWu Albert Y.S.Lam - 4798–4808 + 4799–4809 Intent classification is an important building block of dialogue systems. With the burgeoning of conversational AI, existing systems are not capable of handling numerous fast-emerging intents, which motivates zero-shot intent classification. Nevertheless, research on this problem is still in the incipient stage and few methods are available. A recently proposed zero-shot intent classification method, IntentCapsNet, has been shown to achieve state-of-the-art performance. However, it has two unaddressed limitations: (1) it cannot deal with polysemy when extracting semantic capsules; (2) it hardly recognizes the utterances of unseen intents in the generalized zero-shot intent classification setting. To overcome these limitations, we propose to reconstruct capsule networks for zero-shot intent classification. First, we introduce a dimensional attention mechanism to fight against polysemy. Second, we reconstruct the transformation matrices for unseen intents by utilizing abundant latent information of the labeled utterances, which significantly improves the model generalization ability. Experimental results on two task-oriented dialogue datasets in different languages show that our proposed method outperforms IntentCapsNet and other strong baselines. D19-1486 10.18653/v1/D19-1486 @@ -5630,7 +5629,7 @@ YangSong TaoZhang Ji-RongWen - 4809–4819 + 4810–4820 Person-job fit has been an important task which aims to automatically match job positions with suitable candidates. Previous methods mainly focus on solving the match task in single-domain setting, which may not work well when labeled data is limited. We study the domain adaptation problem for person-job fit. We first propose a deep global match network for capturing the global semantic interactions between two sentences from a job posting and a candidate resume respectively. Furthermore, we extend the match network and implement domain adaptation in three levels, sentence-level representation, sentence-level match, and global match. Extensive experiment results on a large real-world dataset consisting of six domains have demonstrated the effectiveness of the proposed model, especially when there is not sufficient labeled data. D19-1487 10.18653/v1/D19-1487 @@ -5642,7 +5641,7 @@ ChuanShi HouyeJi XiaoliLi - 4820–4829 + 4821–4830 Short text classification has found rich and critical applications in news and tweet tagging to help users find relevant information. Due to lack of labeled training data in many practical use cases, there is a pressing need for studying semi-supervised short text classification. Most existing studies focus on long texts and achieve unsatisfactory performance on short texts due to the sparsity and limited labeled data. In this paper, we propose a novel heterogeneous graph neural network based method for semi-supervised short text classification, leveraging full advantage of few labeled data and large unlabeled data through information propagation along the graph. In particular, we first present a flexible HIN (heterogeneous information network) framework for modeling the short texts, which can integrate any type of additional information as well as capture their relations to address the semantic sparsity. Then, we propose Heterogeneous Graph ATtention networks (HGAT) to embed the HIN for short text classification based on a dual-level attention mechanism, including node-level and type-level attentions. The attention mechanism can learn the importance of different neighboring nodes as well as the importance of different node (information) types to a current node. Extensive experimental results have demonstrated that our proposed model outperforms state-of-the-art methods across six benchmark datasets significantly. D19-1488 10.18653/v1/D19-1488 @@ -5662,7 +5661,7 @@ SeanKross MichelleMazurek HalDaumé III - 4830–4841 + 4831–4842 The readability of a digital text can influence people’s ability to learn new things about a range topics from digital resources (e.g., Wikipedia, WebMD). Readability also impacts search rankings, and is used to evaluate the performance of NLP systems. Despite this, we lack a thorough understanding of how to validly measure readability at scale, especially for domain-specific texts. In this work, we present a comparison of the validity of well-known readability measures and introduce a novel approach, Smart Cloze, which is designed to address shortcomings of existing measures. We compare these approaches across four different corpora: crowdworker-generated stories, Wikipedia articles, security and privacy advice, and health information. On these corpora, we evaluate the convergent and content validity of each measure, and detail tradeoffs in score precision, domain-specificity, and participant burden. These results provide a foundation for more accurate readability measurements and better evaluation of new natural-language-processing systems and tools. D19-1489 10.18653/v1/D19-1489 @@ -5673,7 +5672,7 @@ LuZong YikangYang JionglongSu - 4842–4851 + 4843–4852 With the development of NLP technologies, news can be automatically categorized and labeled according to a variety of characteristics, at the same time be represented as low dimensional embeddings. However, it lacks a systematic approach that effectively integrates the inherited features and inter-textual knowledge of news to represent the collective information with a dense vector. With the aim of filling this gap, the News2vec model is proposed to allow the distributed representation of news taking into account its associated features. To describe the cross-document linkages between news, a network consisting of news and its attributes is constructed. Moreover, the News2vec model treats the news node as a bag of features by developing the Subnode model. Based on the biased random walk and the skip-gram model, each news feature is mapped to a vector, and the news is thus represented as the sum of its features. This approach offers an easy solution to create embeddings for unseen news nodes based on its attributes. To evaluate our model, dimension reduction plots and correlation heat-maps are created to visualize the news vectors, together with the application of two downstream tasks, the stock movement prediction and news recommendation. By comparing with other established text/sentence embedding models, we show that News2vec achieves state-of-the-art performance on these news-related tasks. D19-1490 10.18653/v1/D19-1490 @@ -5682,7 +5681,7 @@ Recursive Context-Aware Lexical Simplification SianGooding EkaterinaKochmar - 4852–4862 + 4853–4863 This paper presents a novel architecture for recursive context-aware lexical simplification, REC-LS, that is capable of (1) making use of the wider context when detecting the words in need of simplification and suggesting alternatives, and (2) taking previous simplification steps into account. We show that our system outputs lexical simplifications that are grammatically correct and semantically appropriate, and outperforms the current state-of-the-art systems in lexical simplification. D19-1491 D19-1491.Attachment.pdf @@ -5693,7 +5692,7 @@ SaraRosenthal KenBarker ZhichengLiang - 4863–4872 + 4864–4873 Electronic Health Records (EHRs) contain both structured content and unstructured (text) content about a patient’s medical history. In the unstructured text parts, there are common sections such as Assessment and Plan, Social History, and Medications. These sections help physicians find information easily and can be used by an information retrieval system to return specific information sought by a user. However, it is common that the exact format of sections in a particular EHR does not adhere to known patterns. Therefore, being able to predict sections and headers in EHRs automatically is beneficial to physicians. Prior approaches in EHR section prediction have only used text data from EHRs and have required significant manual annotation. We propose using sections from medical literature (e.g., textbooks, journals, web content) that contain content similar to that found in EHR sections. Our approach uses data from a different kind of source where labels are provided without the need of a time-consuming annotation effort. We use this data to train two models: an RNN and a BERT-based model. We apply the learned models along with source data via transfer learning to predict sections in EHRs. Our results show that medical literature can provide helpful supervision signal for this classification task. D19-1492 10.18653/v1/D19-1492 @@ -5707,7 +5706,7 @@ JianqiangHuang YongfengHuang XingXie - 4873–4882 + 4874–4883 News recommendation is important for online news platforms to help users find interested news and alleviate information overload. Existing news recommendation methods usually rely on the news click history to model user interest. However, these methods may suffer from the data sparsity problem, since the news click behaviors of many users in online news platforms are usually very limited. Fortunately, some other kinds of user behaviors such as webpage browsing and search queries can also provide useful clues of users’ news reading interest. In this paper, we propose a neural news recommendation approach which can exploit heterogeneous user behaviors. Our approach contains two major modules, i.e., news representation and user representation. In the news representation module, we learn representations of news from their titles via CNN networks, and apply attention networks to select important words. In the user representation module, we propose an attentive multi-view learning framework to learn unified representations of users from their heterogeneous behaviors such as search queries, clicked news and browsed webpages. In addition, we use word- and record-level attentions to select informative words and behavior records. Experiments on a real-world dataset validate the effectiveness of our approach. D19-1493 10.18653/v1/D19-1493 @@ -5720,7 +5719,7 @@ SuyuGe YongfengHuang XingXie - 4883–4892 + 4884–4893 User and item representation learning is critical for recommendation. Many of existing recommendation methods learn representations of users and items based on their ratings and reviews. However, the user-user and item-item relatedness are usually not considered in these methods, which may be insufficient. In this paper, we propose a neural recommendation approach which can utilize useful information from both review content and user-item graphs. Since reviews and graphs have different characteristics, we propose to use a multi-view learning framework to incorporate them as different views. In the review content-view, we propose to use a hierarchical model to first learn sentence representations from words, then learn review representations from sentences, and finally learn user/item representations from reviews. In addition, we propose to incorporate a three-level attention network into this view to select important words, sentences and reviews for learning informative user and item representations. In the graph-view, we propose a hierarchical graph neural network to jointly model the user-item, user-user and item-item relatedness by capturing the first- and second-order interactions between users and items in the user-item graph. In addition, we apply attention mechanism to model the importance of these interactions to learn informative user and item representations. Extensive experiments on four benchmark datasets validate the effectiveness of our approach. D19-1494 10.18653/v1/D19-1494 @@ -5732,7 +5731,7 @@ TingLiu ZhongyangLi JunwenDuan - 4893–4902 + 4894–4903 Prior work has proposed effective methods to learn event representations that can capture syntactic and semantic information over text corpus, demonstrating their effectiveness for downstream tasks such as script event prediction. On the other hand, events extracted from raw texts lacks of commonsense knowledge, such as the intents and emotions of the event participants, which are useful for distinguishing event pairs when there are only subtle differences in their surface realizations. To address this issue, this paper proposes to leverage external commonsense knowledge about the intent and sentiment of the event. Experiments on three event-related tasks, i.e., event similarity, script event prediction and stock market prediction, show that our model obtains much better event embeddings for the tasks, achieving 78% improvements on hard similarity task, yielding more precise inferences on subsequent events under given contexts, and better accuracies in predicting the volatilities of the stock market. D19-1495 10.18653/v1/D19-1495 @@ -5743,7 +5742,7 @@ Jyun-YuJiang Kai-WeiChang WeiWang - 4903–4912 + 4904–4913 Adversarial attacks against machine learning models have threatened various real-world applications such as spam filtering and sentiment analysis. In this paper, we propose a novel framework, learning to discriminate perturbations (DISP), to identify and adjust malicious perturbations, thereby blocking adversarial attacks for text classification models. To identify adversarial attacks, a perturbation discriminator validates how likely a token in the text is perturbed and provides a set of potential perturbations. For each potential perturbation, an embedding estimator learns to restore the embedding of the original word based on the context and a replacement token is chosen based on approximate kNN search. DISP can block adversarial attacks for any NLP model without modifying the model structure or training procedure. Extensive experiments on two benchmark datasets demonstrate that DISP significantly outperforms baseline methods in blocking adversarial attacks for text classification. In addition, in-depth analysis shows the robustness of DISP across different situations. D19-1496 10.18653/v1/D19-1496 @@ -5754,7 +5753,7 @@ Wayne XinZhao Eddy JingYin Ji-RongWen - 4913–4923 + 4914–4924 Citation count prediction (CCP) has been an important research task for automatically estimating the future impact of a scholarly paper. Previous studies mainly focus on extracting or mining useful features from the paper itself or the associated authors. An important kind of data signals, peer review text, has not been utilized for the CCP task. In this paper, we take the initiative to utilize peer review data for the CCP task with a neural prediction model. Our focus is to learn a comprehensive semantic representation for peer review text for improving the prediction performance. To achieve this goal, we incorporate the abstract-review match mechanism and the cross-review match mechanism to learn deep features from peer review text. We also consider integrating hand-crafted features via a wide component. The deep and wide components jointly make the prediction. Extensive experiments have demonstrated the usefulness of the peer review data and the effectiveness of the proposed model. Our dataset has been released online. D19-1497 10.18653/v1/D19-1497 @@ -5764,7 +5763,7 @@ FeniaChristopoulou MakotoMiwa SophiaAnaniadou - 4924–4935 + 4925–4936 Document-level relation extraction is a complex human process that requires logical inference to extract relationships between named entities in text. Existing approaches use graph-based neural models with words as nodes and edges as relations between them, to encode relations across sentences. These models are node-based, i.e., they form pair representations based solely on the two target node representations. However, entity relations can be better expressed through unique edge representations formed as paths between nodes. We thus propose an edge-oriented graph neural model for document-level relation extraction. The model utilises different types of nodes and edges to create a document-level graph. An inference mechanism on the graph edges enables to learn intra- and inter-sentence relations using multi-instance learning internally. Experiments on two document-level biomedical datasets for chemical-disease and gene-disease associations show the usefulness of the proposed edge-oriented approach. D19-1498 10.18653/v1/D19-1498 @@ -5778,7 +5777,7 @@ DongyanZhao ShumingShi RuiYan - 4936–4945 + 4937–4946 Text style transfer task requires the model to transfer a sentence of one style to another style while retaining its original content meaning, which is a challenging problem that has long suffered from the shortage of parallel data. In this paper, we first propose a semi-supervised text style transfer model that combines the small-scale parallel data with the large-scale nonparallel data. With these two types of training data, we introduce a projection function between the latent space of different styles and design two constraints to train it. We also introduce two other simple but effective semi-supervised methods to compare with. To evaluate the performance of the proposed methods, we build and release a novel style transfer dataset that alters sentences between the style of ancient Chinese poem and the modern Chinese. D19-1499 10.18653/v1/D19-1499 @@ -5790,7 +5789,7 @@ ShomirWilson ThomasNorton NormanSadeh - 4946–4957 + 4947–4958 Privacy policies are long and complex documents that are difficult for users to read and understand. Yet, they have legal effects on how user data can be collected, managed and used. Ideally, we would like to empower users to inform themselves about the issues that matter to them, and enable them to selectively explore these issues. We present PrivacyQA, a corpus consisting of 1750 questions about the privacy policies of mobile applications, and over 3500 expert annotations of relevant answers. We observe that a strong neural baseline underperforms human performance by almost 0.3 F1 on PrivacyQA, suggesting considerable room for improvement for future systems. Further, we use this dataset to categorically identify challenges to question answerability, with domain-general implications for any question answering system. The PrivacyQA corpus offers a challenging corpus for question answering, with genuine real world utility. D19-1500 10.18653/v1/D19-1500 @@ -5805,7 +5804,7 @@ KunGai DongyanZhao RuiYan - 4958–4967 + 4959–4968 Different from other text generation tasks, in product description generation, it is of vital importance to generate faithful descriptions that stick to the product attribute information. However, little attention has been paid to this problem. To bridge this gap we propose a model named Fidelity-oriented Product Description Generator (FPDG). FPDG takes the entity label of each word into account, since the product attribute information is always conveyed by entity words. Specifically, we first propose a Recurrent Neural Network (RNN) decoder based on the Entity-label-guided Long Short-Term Memory (ELSTM) cell, taking both the embedding and the entity label of each word as input. Second, we establish a keyword memory that stores the entity labels as keys and keywords as values, and FPDG will attend to keywords through attending to their entity labels. Experiments conducted a large-scale real-world product description dataset show that our model achieves the state-of-the-art performance in terms of both traditional generation metrics as well as human evaluations. Specifically, FPDG increases the fidelity of the generated descriptions by 25%. D19-1501 10.18653/v1/D19-1501 @@ -5816,7 +5815,7 @@ LeiHou JuanziLi TiansiDong - 4968–4977 + 4969–4978 This paper addresses the problem of inferring the fine-grained type of an entity from a knowledge base. We convert this problem into the task of graph-based semi-supervised classification, and propose Hierarchical Multi Graph Convolutional Network (HMGCN), a novel Deep Learning architecture to tackle this problem. We construct three kinds of connectivity matrices to capture different kinds of semantic correlations between entities. A recursive regularization is proposed to model the subClassOf relations between types in given type hierarchy. Extensive experiments with two large-scale public datasets show that our proposed method significantly outperforms four state-of-the-art methods. D19-1502 10.18653/v1/D19-1502 @@ -5828,7 +5827,7 @@ LinhTran GangLee IzhakShafran - 4978–4989 + 4979–4990 Recently we proposed the Span Attribute Tagging (SAT) Model to infer clinical entities (e.g., symptoms) and their properties (e.g., duration). It tackles the challenge of large label space and limited training data using a hierarchical two-stage approach that identifies the span of interest in a tagging step and assigns labels to the span in a classification step. We extend the SAT model to jointly infer not only entities and their properties but also relations between them. Most relation extraction models restrict inferring relations between tokens within a few neighboring sentences, mainly to avoid high computational complexity. In contrast, our proposed Relation-SAT (R-SAT) model is computationally efficient and can infer relations over the entire conversation, spanning an average duration of 10 minutes. We evaluate our model on a corpus of clinical conversations. When the entities are given, the R-SAT outperforms baselines in identifying relations between symptoms and their properties by about 32% (0.82 vs 0.62 F-score) and by about 50% (0.60 vs 0.41 F-score) on medications and their properties. On the more difficult task of jointly inferring entities and relations, the R-SAT model achieves a performance of 0.34 and 0.45 for symptoms and medications respectively, which is significantly better than 0.18 and 0.35 for the baseline model. The contributions of different components of the model are quantified using ablation analysis. D19-1503 10.18653/v1/D19-1503 @@ -5839,7 +5838,7 @@ SungjunCho DavidBindel DavidMimno - 4990–5000 + 4991–5001 Despite great scalability on large data and their ability to understand correlations between topics, spectral topic models have not been widely used due to the absence of reliability in real data and lack of practical implementations. This paper aims to solidify the foundations of spectral topic inference and provide a practical implementation for anchor-based topic modeling. Beginning with vocabulary curation, we scrutinize every single inference step with other viable options. We also evaluate our matrix-based approach against popular alternatives including a tensor-based spectral method as well as probabilistic algorithms. Our quantitative and qualitative experiments demonstrate the power of Rectified Anchor Word algorithm in various real datasets, providing a complete guide to practical correlated topic modeling. D19-1504 D19-1504.Attachment.zip @@ -5852,7 +5851,7 @@ MichaelGamon Sujay KumarJauhar ChangTienLu - 5001–5010 + 5002–5011 Management of collaborative documents can be difficult, given the profusion of edits and comments that multiple authors make during a document’s evolution. Reliably modeling the relationship between edits and comments is a crucial step towards helping the user keep track of a document in flux. A number of authoring tasks, such as categorizing and summarizing edits, detecting completed to-dos, and visually rearranging comments could benefit from such a contribution. Thus, in this paper we explore the relationship between comments and edits by defining two novel, related tasks: Comment Ranking and Edit Anchoring. We begin by collecting a dataset with more than half a million comment-edit pairs based on Wikipedia revision histories. We then propose a hierarchical multi-layer deep neural-network to model the relationship between edits and comments. Our architecture tackles both Comment Ranking and Edit Anchoring tasks by encoding specific edit actions such as additions and deletions, while also accounting for document context. In a number of evaluation settings, our experimental results show that our approach outperforms several strong baselines significantly. We are able to achieve a precision@1 of 71.0% and a precision@3 of 94.4% for Comment Ranking, while we achieve 74.4% accuracy on Edit Anchoring. D19-1505 10.18653/v1/D19-1505 @@ -5862,7 +5861,7 @@ PrabhuKaliamoorthi SujithRavi ZornitsaKozareva - 5011–5020 + 5012–5021 Recently, there has been a great interest in the development of small and accurate neural networks that run entirely on devices such as mobile phones, smart watches and IoT. This enables user privacy, consistent user experience and low latency. Although a wide range of applications have been targeted from wake word detection to short text classification, yet there are no on-device networks for long text classification. We propose a novel projection attention neural network PRADO that combines trainable projections with attention and convolutions. We evaluate our approach on multiple large document text classification tasks. Our results show the effectiveness of the trainable projection model in finding semantically similar phrases and reaching high performance while maintaining compact size. Using this approach, we train tiny neural networks just 200 Kilobytes in size that improve over prior CNN and LSTM models and achieve near state of the art performance on multiple long document classification tasks. We also apply our model for transfer learning, show its robustness and ability to further improve the performance in limited data scenarios. D19-1506 10.18653/v1/D19-1506 @@ -5870,7 +5869,7 @@ Subword Language Model for Query Auto-Completion GyuwanKim - 5021–5031 + 5022–5032 Current neural query auto-completion (QAC) systems rely on character-level language models, but they slow down when queries are long. We present how to utilize subword language models for the fast and accurate generation of query completion candidates. Representing queries with subwords shorten a decoding length significantly. To deal with issues coming from introducing subword language model, we develop a retrace algorithm and a reranking method by approximate marginalization. As a result, our model achieves up to 2.5 times faster while maintaining a similar quality of generated results compared to the character-level baseline. Also, we propose a new evaluation metric, mean recoverable length (MRL), measuring how many upcoming characters the model could complete correctly. It provides more explicit meaning and eliminates the need for prefix length sampling for existing rank-based metrics. Moreover, we performed a comprehensive analysis with ablation study to figure out the importance of each component. D19-1507 10.18653/v1/D19-1507 @@ -5883,7 +5882,7 @@ HuaixiaoTou ZhongyuWei TingChen - 5032–5041 + 5033–5042 Symptom diagnosis is a challenging yet profound problem in natural language processing. Most previous research focus on investigating the standard electronic medical records for symptom diagnosis, while the dialogues between doctors and patients that contain more rich information are not well studied. In this paper, we first construct a dialogue symptom diagnosis dataset based on an online medical forum with a large amount of dialogues between patients and doctors. Then, we provide some benchmark models on this dataset to boost the research of dialogue symptom diagnosis. In order to further enhance the performance of symptom diagnosis over dialogues, we propose a global attention mechanism to capture more symptom related information, and build a symptom graph to model the associations between symptoms rather than treating each symptom independently. Experimental results show that both the global attention and symptom graph are effective to boost dialogue symptom diagnosis. In particular, our proposed model achieves the state-of-the-art performance on the constructed dataset. D19-1508 10.18653/v1/D19-1508 @@ -5896,7 +5895,7 @@ ChandraBhagavatula ElizabethClark YejinChoi - 5042–5052 + 5043–5053 Counterfactual reasoning requires predicting how alternative events, contrary to what actually happened, might have resulted in different outcomes. Despite being considered a necessary component of AI-complete systems, few resources have been developed for evaluating counterfactual reasoning in narratives. In this paper, we propose Counterfactual Story Rewriting: given an original story and an intervening counterfactual event, the task is to minimally revise the story to make it compatible with the given counterfactual event. Solving this task will require deep understanding of causal narrative chains and counterfactual invariance, and integration of such story reasoning capabilities into conditional language generation models. We present TIMETRAVEL, a new dataset of 29,849 counterfactual rewritings, each with the original story, a counterfactual event, and human-generated revision of the original story compatible with the counterfactual event. Additionally, we include 81,407 counterfactual “branches” without a rewritten storyline to support future work on semi- or un-supervised approaches to counterfactual story rewriting. Finally, we evaluate the counterfactual rewriting capacities of several competitive baselines based on pretrained language models, and assess whether common overlap and model-based automatic metrics for text generation correlate well with human scores for counterfactual rewriting. D19-1509 D19-1509.Attachment.pdf @@ -5909,7 +5908,7 @@ SaschaRothe DaniilMirylenka AliakseiSeveryn - 5053–5064 + 5054–5065 We propose LaserTagger - a sequence tagging approach that casts text generation as a text editing task. Target texts are reconstructed from the inputs using three main edit operations: keeping a token, deleting it, and adding a phrase before the token. To predict the edit operations, we propose a novel model, which combines a BERT encoder with an autoregressive Transformer decoder. This approach is evaluated on English text on four tasks: sentence fusion, sentence splitting, abstractive summarization, and grammar correction. LaserTagger achieves new state-of-the-art results on three of these tasks, performs comparably to a set of strong seq2seq baselines with a large number of training examples, and outperforms them when the number of examples is limited. Furthermore, we show that at inference time tagging can be more than two orders of magnitude faster than comparable seq2seq models, making it more attractive for running in a live environment. D19-1510 D19-1510.Attachment.zip @@ -5921,7 +5920,7 @@ ShiFeng DalingWang YifeiZhang - 5065–5075 + 5066–5076 Generating intriguing question is a key step towards building human-like open-domain chatbots. Although some recent works have focused on this task, compared with questions raised by humans, significant gaps remain in maintaining semantic coherence with post, which may result in generating dull or deviated questions. We observe that the answer has strong semantic coherence to its question and post, which can be used to guide question generation. Thus, we devise two methods to further enhance semantic coherence between post and question under the guidance of answer. First, the coherence score between generated question and answer is used as the reward function in a reinforcement learning framework, to encourage the cases that are consistent with the answer in semantic. Second, we incorporate adversarial training to explicitly control question generation in the direction of question-answer coherence. Extensive experiments show that our two methods outperform state-of-the-art baseline algorithms with large margins in raising semantic coherent questions. D19-1511 10.18653/v1/D19-1511 @@ -5932,7 +5931,7 @@ CanXu WeiWu ZhoujunLi - 5076–5088 + 5077–5089 Automatic news comment generation is beneficial for real applications but has not attracted enough attention from the research community. In this paper, we propose a “read-attend-comment” procedure for news comment generation and formalize the procedure with a reading network and a generation network. The reading network comprehends a news article and distills some important points from it, then the generation network creates a comment by attending to the extracted discrete points and the news title. We optimize the model in an end-to-end manner by maximizing a variational lower bound of the true objective using the back-propagation algorithm. Experimental results on two public datasets indicate that our model can significantly outperform existing methods in terms of both automatic evaluation and human judgment. D19-1512 10.18653/v1/D19-1512 @@ -5942,7 +5941,7 @@ HongyinTang MiaoLi BeihongJin - 5089–5098 + 5090–5099 Text generation is among the most fundamental tasks in natural language processing. In this paper, we propose a text generation model that learns semantics and structural features simultaneously. This model captures structural features by a sequential variational autoencoder component and leverages a topic modeling component based on Gaussian distribution to enhance the recognition of text semantics. To make the reconstructed text more coherent to the topics, the model further adapts the encoder of the topic modeling component for a discriminator. The results of experiments over several datasets demonstrate that our model outperforms several states of the art models in terms of text perplexity and topic coherence. Moreover, the latent representations learned by our model is superior to others in a text classification task. Finally, given the input texts, our model can generate meaningful texts which hold similar structures but under different topics. D19-1513 D19-1513.Attachment.zip @@ -5952,7 +5951,7 @@ <fixed-case>LXMERT</fixed-case>: Learning Cross-Modality Encoder Representations from Transformers HaoTan MohitBansal - 5099–5110 + 5100–5111 Vision-and-language reasoning requires an understanding of visual concepts, language semantics, and, most importantly, the alignment and relationships between these two modalities. We thus propose the LXMERT (Learning Cross-Modality Encoder Representations from Transformers) framework to learn these vision-and-language connections. In LXMERT, we build a large-scale Transformer model that consists of three encoders: an object relationship encoder, a language encoder, and a cross-modality encoder. Next, to endow our model with the capability of connecting vision and language semantics, we pre-train the model with large amounts of image-and-sentence pairs, via five diverse representative pre-training tasks: masked language modeling, masked object prediction (feature regression and label classification), cross-modality matching, and image question answering. These tasks help in learning both intra-modality and cross-modality relationships. After fine-tuning from our pre-trained parameters, our model achieves the state-of-the-art results on two visual question answering datasets (i.e., VQA and GQA). We also show the generalizability of our pre-trained cross-modality model by adapting it to a challenging visual-reasoning task, NLVR2, and improve the previous best result by 22% absolute (54% to 76%). Lastly, we demonstrate detailed ablation studies to prove that both our novel model components and pre-training strategies significantly contribute to our strong results. Code and pre-trained models publicly available at: https://github.com/airsplay/lxmert D19-1514 10.18653/v1/D19-1514 @@ -5961,7 +5960,7 @@ Phrase Grounding by Soft-Label Chain Conditional Random Field JiachengLiu JuliaHockenmaier - 5111–5121 + 5112–5122 The phrase grounding task aims to ground each entity mention in a given caption of an image to a corresponding region in that image. Although there are clear dependencies between how different mentions of the same caption should be grounded, previous structured prediction methods that aim to capture such dependencies need to resort to approximate inference or non-differentiable losses. In this paper, we formulate phrase grounding as a sequence labeling task where we treat candidate regions as potential labels, and use neural chain Conditional Random Fields (CRFs) to model dependencies among regions for adjacent mentions. In contrast to standard sequence labeling tasks, the phrase grounding task is defined such that there may be multiple correct candidate regions. To address this multiplicity of gold labels, we define so-called Soft-Label Chain CRFs, and present an algorithm that enables convenient end-to-end training. Our method establishes a new state-of-the-art on phrase grounding on the Flickr30k Entities dataset. Analysis shows that our model benefits both from the entity dependencies captured by the CRF and from the soft-label training regime. Our code is available at github.com/liujch1998/SoftLabelCCRF D19-1515 @@ -5975,7 +5974,7 @@ YangqiuSong YanSong ChangshuiZhang - 5122–5131 + 5123–5132 Grounding a pronoun to a visual object it refers to requires complex reasoning from various information sources, especially in conversational scenarios. For example, when people in a conversation talk about something all speakers can see, they often directly use pronouns (e.g., it) to refer to it without previous introduction. This fact brings a huge challenge for modern natural language understanding systems, particularly conventional context-based pronoun coreference models. To tackle this challenge, in this paper, we formally define the task of visual-aware pronoun coreference resolution (PCR) and introduce VisPro, a large-scale dialogue PCR dataset, to investigate whether and how the visual information can help resolve pronouns in dialogues. We then propose a novel visual-aware PCR model, VisCoref, for this task and conduct comprehensive experiments and case studies on our dataset. Results demonstrate the importance of the visual information in this PCR case and show the effectiveness of the proposed model. D19-1516 10.18653/v1/D19-1516 @@ -5986,7 +5985,7 @@ YongchengWang ShizheChen QinJin - 5132–5142 + 5133–5143 Multimodal semantic comprehension has attracted increasing research interests recently such as visual question answering and caption generation. However, due to the data limitation, fine-grained semantic comprehension has not been well investigated, which requires to capture semantic details of multimodal contents. In this work, we introduce “YouMakeup”, a large-scale multimodal instructional video dataset to support fine-grained semantic comprehension research in specific domain. YouMakeup contains 2,800 videos from YouTube, spanning more than 420 hours in total. Each video is annotated with a sequence of natural language descriptions for instructional steps, grounded in temporal video range and spatial facial areas. The annotated steps in a video involve subtle difference in actions, products and regions, which requires fine-grained understanding and reasoning both temporally and spatially. In order to evaluate models’ ability for fined-grained comprehension, we further propose two groups of tasks including generation tasks and visual question answering from different aspects. We also establish a baseline of step caption generation for future comparison. The dataset will be publicly available at https://github. com/AIM3-RUC/YouMakeup to support research investigation in fine-grained semantic comprehension. D19-1517 10.18653/v1/D19-1517 @@ -5998,7 +5997,7 @@ ChilieTan XiaolinLi JunXiao - 5143–5152 + 5144–5153 In this paper, we focus on natural language video localization: localizing (ie, grounding) a natural language description in a long and untrimmed video sequence. All currently published models for addressing this problem can be categorized into two types: (i) top-down approach: it does classification and regression for a set of pre-cut video segment candidates; (ii) bottom-up approach: it directly predicts probabilities for each video frame as the temporal boundaries (ie, start and end time point). However, both two approaches suffer several limitations: the former is computation-intensive for densely placed candidates, while the latter has trailed the performance of the top-down counterpart thus far. To this end, we propose a novel dense bottom-up framework: DEnse Bottom-Up Grounding (DEBUG). DEBUG regards all frames falling in the ground truth segment as foreground, and each foreground frame regresses the unique distances from its location to bi-directional ground truth boundaries. Extensive experiments on three challenging benchmarks (TACoS, Charades-STA, and ActivityNet Captions) show that DEBUG is able to match the speed of bottom-up models while surpassing the performance of the state-of-the-art top-down models. D19-1518 10.18653/v1/D19-1518 @@ -6011,7 +6010,7 @@ LihaoLu JiachengLiu JiaweiHan - 5153–5162 + 5154–5163 Everyone makes mistakes. So do human annotators when curating labels for named entity recognition (NER). Such label mistakes might hurt model training and interfere model comparison. In this study, we dive deep into one of the widely-adopted NER benchmark datasets, CoNLL03 NER. We are able to identify label mistakes in about 5.38% test sentences, which is a significant ratio considering that the state-of-the-art test F1 score is already around 93%. Therefore, we manually correct these label mistakes and form a cleaner test set. Our re-evaluation of popular models on this corrected test set leads to more accurate assessments, compared to those on the original test set. More importantly, we propose a simple yet effective framework, CrossWeigh, to handle label mistakes during NER model training. Specifically, it partitions the training data into several folds and train independent NER models to identify potential mistakes in each fold. Then it adjusts the weights of training data accordingly to train the final NER model. Extensive experiments demonstrate significant improvements of plugging various NER models into our proposed framework on three datasets. All implementations and corrected test set are available at our Github repo https://github.com/ZihanWangKi/CrossWeigh. D19-1519 10.18653/v1/D19-1519 @@ -6023,7 +6022,7 @@ ZaidSheikh GrahamNeubig JaimeCarbonell - 5163–5173 + 5164–5174 Most state-of-the-art models for named entity recognition (NER) rely on the availability of large amounts of labeled data, making them challenging to extend to new, lower-resourced languages. However, there are now many proposed solutions to this problem involving either cross-lingual transfer learning, which learns from other highly resourced languages, or active learning, which efficiently selects effective training data based on model predictions. In this paper, we ask the question: given this recent progress, and some amount of human annotation, what is the most effective method for efficiently creating high-quality entity recognizers in under-resourced languages? Based on extensive experimentation using both simulated and real human annotation, we settle on a recipe of starting with a cross-lingual transferred model, then performing targeted annotation of only uncertain entity spans in the target language, minimizing annotator effort. Results demonstrate that cross-lingual transfer is a powerful tool when very little data can be annotated, but an entity-targeted annotation strategy can achieve competitive accuracy quickly, with just one-tenth of training data. D19-1520 D19-1520.Attachment.pdf @@ -6036,7 +6035,7 @@ ChenyanXiong DanielCampos ArnoldOverwijk - 5174–5183 + 5175–5184 This paper studies keyphrase extraction in real-world scenarios where documents are from diverse domains and have variant content quality. We curate and release OpenKP, a large scale open domain keyphrase extraction dataset with near one hundred thousand web documents and expert keyphrase annotations. To handle the variations of domain and content quality, we develop BLING-KPE, a neural keyphrase extraction model that goes beyond language understanding using visual presentations of documents and weak supervision from search queries. Experimental results on OpenKP confirm the effectiveness of BLING-KPE and the contributions of its neural architecture, visual features, and search log weak supervision. Zero-shot evaluations on DUC-2001 demonstrate the improved generalization ability of learning from the open domain data compared to a specific domain. D19-1521 D19-1521.Attachment.zip @@ -6047,7 +6046,7 @@ IvanaBalazevic CarlAllen TimothyHospedales - 5184–5193 + 5185–5194 Knowledge graphs are structured representations of real world facts. However, they typically contain only a small subset of all possible facts. Link prediction is a task of inferring missing facts based on existing ones. We propose TuckER, a relatively straightforward but powerful linear model based on Tucker decomposition of the binary tensor representation of knowledge graph triples. TuckER outperforms previous state-of-the-art models across standard link prediction datasets, acting as a strong baseline for more elaborate models. We show that TuckER is a fully expressive model, derive sufficient bounds on its embedding dimensionalities and demonstrate that several previously introduced linear models can be viewed as special cases of TuckER. D19-1522 D19-1522.Attachment.zip @@ -6057,7 +6056,7 @@ Human-grounded Evaluations of Explanation Methods for Text Classification PiyawatLertvittayakumjorn FrancescaToni - 5194–5204 + 5195–5205 Due to the black-box nature of deep learning models, methods for explaining the models’ results are crucial to gain trust from humans and support collaboration between AIs and humans. In this paper, we consider several model-agnostic and model-specific explanation methods for CNNs for text classification and conduct three human-grounded evaluations, focusing on different purposes of explanations: (1) revealing model behavior, (2) justifying model predictions, and (3) helping humans investigate uncertain predictions. The results highlight dissimilar qualities of the various explanation methods we consider and show the degree to which these methods could serve for each purpose. D19-1523 D19-1523.Attachment.zip @@ -6070,7 +6069,7 @@ ChongFeng AnqingZheng XiaopengLiu - 5205–5214 + 5206–5215 We introduce a new task of modeling the role and function for on-line resource citations in scientific literature. By categorizing the on-line resources and analyzing the purpose of resource citations in scientific texts, it can greatly help resource search and recommendation systems to better understand and manage the scientific resources. For this novel task, we are the first to create an annotation scheme, which models the different granularity of information from a hierarchical perspective. And we construct a dataset SciRes, which includes 3,088 manually annotated resource contexts. In this paper, we propose a possible solution by using a multi-task framework to build the scientific resource classifier (SciResCLF) for jointly recognizing the role and function types. Then we use the classification results to help a scientific resource recommendation (SciResREC) task. Experiments show that our model achieves the best results on both the classification task and the recommendation task. The SciRes dataset is released for future research. D19-1524 D19-1524.Attachment.pdf @@ -6082,7 +6081,7 @@ ShehzeenHussain ShlomoDubnov FarinazKoushanfar - 5215–5224 + 5216–5225 In this work, we develop methods to repurpose text classification neural networks for alternate tasks without modifying the network architecture or parameters. We propose a context based vocabulary remapping method that performs a computationally inexpensive input transformation to reprogram a victim classification model for a new set of sequences. We propose algorithms for training such an input transformation in both white box and black box settings where the adversary may or may not have access to the victim model’s architecture and parameters. We demonstrate the application of our model and the vulnerability of neural networks by adversarially repurposing various text-classification models including LSTM, bi-directional LSTM and CNN for alternate classification tasks. D19-1525 D19-1525.Attachment.pdf @@ -6094,7 +6093,7 @@ QinliangSu DinghanShen ChangyouChen - 5225–5234 + 5226–5235 Hashing is promising for large-scale information retrieval tasks thanks to the efficiency of distance evaluation between binary codes. Generative hashing is often used to generate hashing codes in an unsupervised way. However, existing generative hashing methods only considered the use of simple priors, like Gaussian and Bernoulli priors, which limits these methods to further improve their performance. In this paper, two mixture-prior generative models are proposed, under the objective to produce high-quality hashing codes for documents. Specifically, a Gaussian mixture prior is first imposed onto the variational auto-encoder (VAE), followed by a separate step to cast the continuous latent representation of VAE into binary code. To avoid the performance loss caused by the separate casting, a model using a Bernoulli mixture prior is further developed, in which an end-to-end training is admitted by resorting to the straight-through (ST) discrete gradient estimator. Experimental results on several benchmark datasets demonstrate that the proposed methods, especially the one using Bernoulli mixture priors, consistently outperform existing ones by a substantial margin. D19-1526 10.18653/v1/D19-1526 @@ -6105,7 +6104,7 @@ ZhixinZhou ZhaozhuoXu PingLi - 5235–5245 + 5236–5246 Retrieval of relevant vectors produced by representation learning critically influences the efficiency in natural language processing (NLP) tasks. In this paper, we demonstrate an efficient method for searching vectors via a typical non-metric matching function: inner product. Our method, which constructs an approximate Inner Product Delaunay Graph (IPDG) for top-1 Maximum Inner Product Search (MIPS), transforms retrieving the most suitable latent vectors into a graph search problem with great benefits of efficiency. Experiments on data representations learned for different machine learning tasks verify the outperforming effectiveness and efficiency of the proposed IPDG. D19-1527 10.18653/v1/D19-1527 @@ -6120,7 +6119,7 @@ YangqiuSong WilfredNg DongYu - 5246–5255 + 5247–5256 Conventional word embeddings represent words with fixed vectors, which are usually trained based on co-occurrence patterns among words. In doing so, however, the power of such representations is limited, where the same word might be functionalized separately under different syntactic relations. To address this limitation, one solution is to incorporate relational dependencies of different words into their embeddings. Therefore, in this paper, we propose a multiplex word embedding model, which can be easily extended according to various relations among words. As a result, each word has a center embedding to represent its overall semantics, and several relational embeddings to represent its relational dependencies. Compared to existing models, our model can effectively distinguish words with respect to different relations without introducing unnecessary sparseness. Moreover, to accommodate various relations, we use a small dimension for relational embeddings and our model is able to keep their effectiveness. Experiments on selectional preference acquisition and word similarity demonstrate the effectiveness of the proposed model, and a further study of scalability also proves that our embeddings only need 1/20 of the original embedding size to achieve better performance. D19-1528 10.18653/v1/D19-1528 @@ -6130,7 +6129,7 @@ YukunMa Patrick H.Chen Cho-JuiHsieh - 5256–5265 + 5257–5266 It is challenging to deploy deep neural nets on memory-constrained devices due to the explosion of numbers of parameters. Especially, the input embedding layer and Softmax layer usually dominate the memory usage in an RNN-based language model. For example, input embedding and Softmax matrices in IWSLT-2014 German-to-English data set account for more than 80% of the total model parameters. To compress these embedding layers, we propose MulCode, a novel multi-way multiplicative neural compressor. MulCode learns an adaptively created matrix and its multiplicative compositions. Together with a prior weighted loss, Multicode is more effective than the state-of-the-art compression methods. On the IWSLT-2014 machine translation data set, MulCode achieved 17 times compression rate for the embedding and Softmax matrices, and when combined with quantization technique, our method can achieve 41.38 times compression rate with very little loss in performance. D19-1529 10.18653/v1/D19-1529 @@ -6141,7 +6140,7 @@ HilaGonen RyanCotterell SimoneTeufel - 5266–5274 + 5267–5275 This paper treats gender bias latent in word embeddings. Previous mitigation attempts rely on the operationalisation of gender bias as a projection over a linear subspace. An alternative approach is Counterfactual Data Augmentation (CDA), in which a corpus is duplicated and augmented to remove bias, e.g. by swapping all inherently-gendered words in the copy. We perform an empirical comparison of these approaches on the English Gigaword and Wikipedia, and find that whilst both successfully reduce direct bias and perform well in tasks which quantify embedding quality, CDA variants outperform projection-based methods at the task of drawing non-biased gender analogies by an average of 19% across both corpora. We propose two improvements to CDA: Counterfactual Data Substitution (CDS), a variant of CDA in which potentially biased text is randomly substituted to avoid duplication, and the Names Intervention, a novel name-pairing technique that vastly increases the number of words being treated. CDA/S with the Names Intervention is the only approach which is able to mitigate indirect gender bias: following debiasing, previously biased words are significantly less clustered according to gender (cluster purity is reduced by 49%), thus improving on the state-of-the-art for bias mitigation. D19-1530 D19-1530.Attachment.zip @@ -6156,7 +6155,7 @@ MuhaoChen RyanCotterell Kai-WeiChang - 5275–5283 + 5276–5284 Recent studies have shown that word embeddings exhibit gender bias inherited from the training corpora. However, most studies to date have focused on quantifying and mitigating such bias only in English. These analyses cannot be directly extended to languages that exhibit morphological agreement on gender, such as Spanish and French. In this paper, we propose new metrics for evaluating gender bias in word embeddings of these languages and further demonstrate evidence of gender bias in bilingual embeddings which align these languages with English. Finally, we extend an existing approach to mitigate gender bias in word embedding of these languages under both monolingual and bilingual settings. Experiments on modified Word Embedding Association Test, word similarity, word translation, and word pair translation tasks show that the proposed approaches can effectively reduce the gender bias while preserving the utility of the original embeddings. D19-1531 10.18653/v1/D19-1531 @@ -6165,7 +6164,7 @@ Weakly Supervised Cross-lingual Semantic Relation Classification via Knowledge Distillation YogarshiVyas MarineCarpuat - 5284–5295 + 5285–5296 Words in different languages rarely cover the exact same semantic space. This work characterizes differences in meaning between words across languages using semantic relations that have been used to relate the meaning of English words. However, because of translation ambiguity, semantic relations are not always preserved by translation. We introduce a cross-lingual relation classifier trained only with English examples and a bilingual dictionary. Our classifier relies on a novel attention-based distillation approach to account for translation ambiguity when transferring knowledge from English to cross-lingual settings. On new English-Chinese and English-Hindi test sets, the resulting models largely outperform baselines that more naively rely on bilingual embeddings or dictionaries for cross-lingual transfer, and approach the performance of fully supervised systems on English tasks. D19-1532 D19-1532.Attachment.pdf @@ -6176,7 +6175,7 @@ ChristianHadiwinoto Hwee TouNg Wee ChungGan - 5296–5305 + 5297–5306 Contextualized word representations are able to give different representations for the same word in different contexts, and they have been shown to be effective in downstream natural language processing tasks, such as question answering, named entity recognition, and sentiment analysis. However, evaluation on word sense disambiguation (WSD) in prior work shows that using contextualized word representations does not outperform the state-of-the-art approach that makes use of non-contextualized word embeddings. In this paper, we explore different strategies of integrating pre-trained contextualized word representations and our best strategy achieves accuracies exceeding the best prior published accuracies by significant margins on multiple benchmark WSD datasets. D19-1533 10.18653/v1/D19-1533 @@ -6188,7 +6187,7 @@ SujianLi SameerSingh MattGardner - 5306–5314 + 5307–5315 The ability to understand and work with numbers (numeracy) is critical for many complex reasoning tasks. Currently, most NLP models treat numbers in text in the same way as other tokens—they embed them as distributed vectors. Is this enough to capture numeracy? We begin by investigating the numerical reasoning capabilities of a state-of-the-art question answering model on the DROP dataset. We find this model excels on questions that require numerical reasoning, i.e., it already captures numeracy. To understand how this capability emerges, we probe token embedding methods (e.g., BERT, GloVe) on synthetic list maximum, number decoding, and addition tasks. A surprising degree of numeracy is naturally present in standard embeddings. For example, GloVe and word2vec accurately encode magnitude for numbers up to 1,000. Furthermore, character-level embeddings are even more precise—ELMo captures numeracy the best for all pre-trained methods—but BERT, which uses sub-word units, is less exact. D19-1534 D19-1534.Attachment.zip @@ -6203,7 +6202,7 @@ LeiFang BinZhou DongmeiZhang - 5315–5325 + 5316–5326 Context-dependent semantic parsing has proven to be an important yet challenging task. To leverage the advances in context-independent semantic parsing, we propose to perform follow-up query analysis, aiming to restate context-dependent natural language queries with contextual information. To accomplish the task, we propose STAR, a novel approach with a well-designed two-phase process. It is parser-independent and able to handle multifarious follow-up scenarios in different domains. Experiments on the FollowUp dataset show that STAR outperforms the state-of-the-art baseline by a large margin of nearly 8%. The superiority on parsing results verifies the feasibility of follow-up query analysis. We also explore the extensibility of STAR on the SQA dataset, which is very promising. D19-1535 10.18653/v1/D19-1535 @@ -6212,7 +6211,7 @@ <fixed-case>T</fixed-case>ext2<fixed-case>M</fixed-case>ath: End-to-end Parsing Text into Math Expressions YanyanZou WeiLu - 5326–5336 + 5327–5337 We propose Text2Math, a model for semantically parsing text into math expressions. The model can be used to solve different math related problems including arithmetic word problems and equation parsing problems. Unlike previous approaches, we tackle the problem from an end-to-end structured prediction perspective where our algorithm aims to predict the complete math expression at once as a tree structure, where minimal manual efforts are involved in the process. Empirical results on benchmark datasets demonstrate the efficacy of our approach. D19-1536 D19-1536.Attachment.zip @@ -6230,7 +6229,7 @@ CaimingXiong RichardSocher DragomirRadev - 5337–5348 + 5338–5349 We focus on the cross-domain context-dependent text-to-SQL generation task. Based on the observation that adjacent natural language questions are often linguistically dependent and their corresponding SQL queries tend to overlap, we utilize the interaction history by editing the previous predicted query to improve the generation quality. Our editing mechanism views SQL as sequences and reuses generation results at the token level in a simple manner. It is flexible to change individual tokens and robust to error propagation. Furthermore, to deal with complex table structures in different domains, we employ an utterance-table encoder and a table-aware decoder to incorporate the context of the user utterance and the table schema. We evaluate our approach on the SParC dataset and demonstrate the benefit of editing compared with the state-of-the-art baselines which generate SQL from scratch. Our code is available at https://github.com/ryanzhumich/sparc_atis_pytorch. D19-1537 10.18653/v1/D19-1537 @@ -6240,7 +6239,7 @@ ShexiaHe ZuchaoLi HaiZhao - 5349–5358 + 5350–5359 Recently, semantic role labeling (SRL) has earned a series of success with even higher performance improvements, which can be mainly attributed to syntactic integration and enhanced word representation. However, most of these efforts focus on English, while SRL on multiple languages more than English has received relatively little attention so that is kept underdevelopment. Thus this paper intends to fill the gap on multilingual SRL with special focus on the impact of syntax and contextualized word representation. Unlike existing work, we propose a novel method guided by syntactic rule to prune arguments, which enables us to integrate syntax into multilingual SRL model simply and effectively. We present a unified SRL model designed for multiple languages together with the proposed uniform syntax enhancement. Our model achieves new state-of-the-art results on the CoNLL-2009 benchmarks of all seven languages. Besides, we pose a discussion on the syntactic role among different languages and verify the effectiveness of deep enhanced representation for multilingual SRL. D19-1538 10.18653/v1/D19-1538 @@ -6252,7 +6251,7 @@ YinhanLiu LukeZettlemoyer MichaelAuli - 5359–5368 + 5360–5369 We present a new approach for pretraining a bi-directional transformer model that provides significant performance gains across a variety of language understanding problems. Our model solves a cloze-style word reconstruction task, where each word is ablated and must be predicted given the rest of the text. Experiments demonstrate large performance gains on GLUE and new state of the art results on NER as well as constituency parsing benchmarks, consistent with BERT. We also present a detailed analysis of a number of factors that contribute to effective pretraining, including data domain and size, model capacity, and variations on the cloze objective. D19-1539 10.18653/v1/D19-1539 @@ -6265,7 +6264,7 @@ WeiYang PengShi JimmyLin - 5369–5380 + 5370–5381 A core problem of information retrieval (IR) is relevance matching, which is to rank documents by relevance to a user’s query. On the other hand, many NLP problems, such as question answering and paraphrase identification, can be considered variants of semantic matching, which is to measure the semantic distance between two pieces of short texts. While at a high level both relevance and semantic matching require modeling textual similarity, many existing techniques for one cannot be easily adapted to the other. To bridge this gap, we propose a novel model, HCAN (Hybrid Co-Attention Network), that comprises (1) a hybrid encoder module that includes ConvNet-based and LSTM-based encoders, (2) a relevance matching module that measures soft term matches with importance weighting at multiple granularities, and (3) a semantic matching module with co-attention mechanisms that capture context-aware semantic relatedness. Evaluations on multiple IR and NLP benchmarks demonstrate state-of-the-art effectiveness compared to approaches that do not exploit pretraining on external data. Extensive ablation studies suggest that relevance and semantic matching signals are complementary across many problem settings, regardless of the choice of underlying encoders. D19-1540 10.18653/v1/D19-1540 @@ -6275,7 +6274,7 @@ QingrongXia ZhenghuaLi MinZhang - 5381–5391 + 5382–5392 Semantic role labeling (SRL) aims to identify the predicate-argument structure of a sentence. Inspired by the strong correlation between syntax and semantics, previous works pay much attention to improve SRL performance on exploiting syntactic knowledge, achieving significant results. Pipeline methods based on automatic syntactic trees and multi-task learning (MTL) approaches using standard syntactic trees are two common research orientations. In this paper, we adopt a simple unified span-based model for both span-based and word-based Chinese SRL as a strong baseline. Besides, we present a MTL framework that includes the basic SRL module and a dependency parser module. Different from the commonly used hard parameter sharing strategy in MTL, the main idea is to extract implicit syntactic representations from the dependency parser as external inputs for the basic SRL model. Experiments on the benchmarks of Chinese Proposition Bank 1.0 and CoNLL-2009 Chinese datasets show that our proposed framework can effectively improve the performance over the strong baselines. With the external BERT representations, our framework achieves new state-of-the-art 87.54 and 88.5 F1 scores on the two test data of the two benchmarks, respectively. In-depth analysis are conducted to gain more insights on the proposed framework and the effectiveness of syntax. D19-1541 10.18653/v1/D19-1541 @@ -6284,7 +6283,7 @@ Transfer Fine-Tuning: A <fixed-case>BERT</fixed-case> Case Study YukiArase Jun’ichiTsujii - 5392–5403 + 5393–5404 A semantic equivalence assessment is defined as a task that assesses semantic equivalence in a sentence pair by binary judgment (i.e., paraphrase identification) or grading (i.e., semantic textual similarity measurement). It constitutes a set of tasks crucial for research on natural language understanding. Recently, BERT realized a breakthrough in sentence representation learning (Devlin et al., 2019), which is broadly transferable to various NLP tasks. While BERT’s performance improves by increasing its model size, the required computational power is an obstacle preventing practical applications from adopting the technology. Herein, we propose to inject phrasal paraphrase relations into BERT in order to generate suitable representations for semantic equivalence assessment instead of increasing the model size. Experiments on standard natural language understanding tasks confirm that our method effectively improves a smaller BERT model while maintaining the model size. The generated model exhibits superior performance compared to a larger BERT model on semantic equivalence assessment tasks. Furthermore, it achieves larger performance gains on tasks with limited training datasets for fine-tuning, which is a property desirable for transfer learning. D19-1542 10.18653/v1/D19-1542 @@ -6296,7 +6295,7 @@ HongzhiLiu Jian-GuangLou DongmeiZhang - 5404–5413 + 5405–5414 On text-to-SQL generation, the input utterance usually contains lots of tokens that are related to column names or cells in the table, called table-related tokens. These table-related tokens are troublesome for the downstream neural semantic parser because it brings complex semantics and hinders the sharing across the training examples. However, existing approaches either ignore handling these tokens before the semantic parser or simply use deterministic approaches based on string-match or word embedding similarity. In this work, we propose a more efficient approach to handle table-related tokens before the semantic parser. First, we formulate it as a sequential tagging problem and propose a two-stage anonymization model to learn the semantic relationship between tables and input utterances. Then, we leverage the implicit supervision from SQL queries by policy gradient to guide the training. Experiments demonstrate that our approach consistently improves performances of different neural semantic parsers and significantly outperforms deterministic approaches. D19-1543 10.18653/v1/D19-1543 @@ -6306,7 +6305,7 @@ XinchiChen ChunchuanLyu IvanTitov - 5414–5424 + 5415–5425 Semantic role labeling (SRL) involves extracting propositions (i.e. predicates and their typed arguments) from natural language sentences. State-of-the-art SRL models rely on powerful encoders (e.g., LSTMs) and do not model non-local interaction between arguments. We propose a new approach to modeling these interactions while maintaining efficient inference. Specifically, we use Capsule Networks (Sabour et al., 2017): each proposition is encoded as a tuple of capsules, one capsule per argument type (i.e. role). These tuples serve as embeddings of entire propositions. In every network layer, the capsules interact with each other and with representations of words in the sentence. Each iteration results in updated proposition embeddings and updated predictions about the SRL structure. Our model substantially outperforms the non-refinement baseline model on all 7 CoNLL-2019 languages and achieves state-of-the-art results on 5 languages (including English) for dependency SRL. We analyze the types of mistakes corrected by the refinement procedure. For example, each role is typically (but not always) filled with at most one argument. Whereas enforcing this approximate constraint is not useful with the modern SRL system, iterative procedure corrects the mistakes by capturing this intuition in a flexible and context-sensitive way. D19-1544 10.18653/v1/D19-1544 @@ -6316,7 +6315,7 @@ SrinivasanIyer AlvinCheung LukeZettlemoyer - 5425–5434 + 5426–5435 Programmers typically organize executable source code using high-level coding patterns or idiomatic structures such as nested loops, exception handlers and recursive blocks, rather than as individual code tokens. In contrast, state of the art (SOTA) semantic parsers still map natural language instructions to source code by building the code syntax tree one node at a time. In this paper, we introduce an iterative method to extract code idioms from large source code corpora by repeatedly collapsing most-frequent depth-2 subtrees of their syntax trees, and train semantic parsers to apply these idioms during decoding. Applying idiom-based decoding on a recent context-dependent semantic parsing task improves the SOTA by 2.2% BLEU score while reducing training time by more than 50%. This improved speed enables us to scale up the model by training on an extended training set that is 5\times larger, to further move up the SOTA by an additional 2.3% BLEU and 0.9% exact match. Finally, idioms also significantly improve accuracy of semantic parsing to SQL on the ATIS-SQL dataset, when training data is limited. D19-1545 D19-1545.Attachment.zip @@ -6327,7 +6326,7 @@ RajasAgashe SrinivasanIyer LukeZettlemoyer - 5435–5445 + 5436–5446 Interactive programming with interleaved code snippet cells and natural language markdown is recently gaining popularity in the form of Jupyter notebooks, which accelerate prototyping and collaboration. To study code generation conditioned on a long context history, we present JuICe, a corpus of 1.5 million examples with a curated test set of 3.7K instances based on online programming assignments. Compared with existing contextual code generation datasets, JuICe provides refined human-curated data, open-domain code, and an order of magnitude more training data. Using JuICe, we train models for two tasks: (1) generation of the API call sequence in a code cell, and (2) full code cell generation, both conditioned on the NL-Code history up to a particular code cell. Experiments using current baseline code generation models show that both context and distant supervision aid in generation, and that the dataset is challenging for current systems. D19-1546 10.18653/v1/D19-1546 @@ -6338,7 +6337,7 @@ YuSu HuanSun Wen-tauYih - 5446–5457 + 5447–5458 As a promising paradigm, interactive semantic parsing has shown to improve both semantic parsing accuracy and user confidence in the results. In this paper, we propose a new, unified formulation of the interactive semantic parsing problem, where the goal is to design a model-based intelligent agent. The agent maintains its own state as the current predicted semantic parse, decides whether and where human intervention is needed, and generates a clarification question in natural language. A key part of the agent is a world model: it takes a percept (either an initial question or subsequent feedback from the user) and transitions to a new state. We then propose a simple yet remarkably effective instantiation of our framework, demonstrated on two text-to-SQL datasets (WikiSQL and Spider) with different state-of-the-art base semantic parsers. Compared to an existing interactive semantic parsing approach that treats the base parser as a black box, our approach solicits less user feedback but yields higher run-time accuracy. D19-1547 D19-1547.Attachment.pdf @@ -6352,7 +6351,7 @@ LonghuaQian MinZhang GuodongZhou - 5458–5467 + 5459–5468 Recent studies on AMR-to-text generation often formalize the task as a sequence-to-sequence (seq2seq) learning problem by converting an Abstract Meaning Representation (AMR) graph into a word sequences. Graph structures are further modeled into the seq2seq framework in order to utilize the structural information in the AMR graphs. However, previous approaches only consider the relations between directly connected concepts while ignoring the rich structure in AMR graphs. In this paper we eliminate such a strong limitation and propose a novel structure-aware self-attention approach to better model the relations between indirectly connected concepts in the state-of-the-art seq2seq model, i.e. the Transformer. In particular, a few different methods are explored to learn structural representations between two concepts. Experimental results on English AMR benchmark datasets show that our approach significantly outperforms the state-of-the-art with 29.66 and 31.82 BLEU scores on LDC2015E86 and LDC2017T10, respectively. To the best of our knowledge, these are the best results achieved so far by supervised models on the benchmarks. D19-1548 10.18653/v1/D19-1548 @@ -6361,7 +6360,7 @@ Syntax-Aware Aspect Level Sentiment Classification with Graph Attention Networks BinxuanHuang KathleenCarley - 5468–5476 + 5469–5477 Aspect level sentiment classification aims to identify the sentiment expressed towards an aspect given a context sentence. Previous neural network based methods largely ignore the syntax structure in one sentence. In this paper, we propose a novel target-dependent graph attention network (TD-GAT) for aspect level sentiment classification, which explicitly utilizes the dependency relationship among words. Using the dependency graph, it propagates sentiment features directly from the syntactic context of an aspect target. In our experiments, we show our method outperforms multiple baselines with GloVe embeddings. We also demonstrate that using BERT representations further substantially boosts the performance. D19-1549 10.18653/v1/D19-1549 @@ -6370,7 +6369,7 @@ Learning Explicit and Implicit Structures for Targeted Sentiment Analysis HaoLi WeiLu - 5477–5487 + 5478–5488 Targeted sentiment analysis is the task of jointly predicting target entities and their associated sentiment information. Existing research efforts mostly regard this joint task as a sequence labeling problem, building models that can capture explicit structures in the output space. However, the importance of capturing implicit global structural information that resides in the input space is largely unexplored. In this work, we argue that both types of information (implicit and explicit structural information) are crucial for building a successful targeted sentiment analysis model. Our experimental results show that properly capturing both information is able to lead to better performance than competitive existing approaches. We also conduct extensive experiments to investigate our model’s effectiveness and robustness. D19-1550 D19-1550.Attachment.pdf @@ -6385,7 +6384,7 @@ JianxinLiao TongXu MingLiu - 5488–5497 + 5489–5498 Aspect-level sentiment classification is a crucial task for sentiment analysis, which aims to identify the sentiment polarities of specific targets in their context. The main challenge comes from multi-aspect sentences, which express multiple sentiment polarities towards different targets, resulting in overlapped feature representation. However, most existing neural models tend to utilize static pooling operation or attention mechanism to identify sentimental words, which therefore insufficient for dealing with overlapped features. To solve this problem, we propose to utilize capsule network to construct vector-based feature representation and cluster features by an EM routing algorithm. Furthermore, interactive attention mechanism is introduced in the capsule routing procedure to model the semantic relationship between aspect terms and context. The iterative routing also enables encoding sentence from a global perspective. Experimental results on three datasets show that our proposed model achieves state-of-the-art performance. D19-1551 10.18653/v1/D19-1551 @@ -6397,7 +6396,7 @@ ShoushanLi GuodongZhou MinZhang - 5498–5506 + 5499–5507 There have been a recent line of works to automatically predict the emotions of posts in social media. Existing approaches consider the posts individually and predict their emotions independently. Different from previous researches, we explore the dependence among relevant posts via the authors’ backgrounds, since the authors with similar backgrounds, e.g., gender, location, tend to express similar emotions. However, such personal attributes are not easy to obtain in most social media websites, and it is hard to capture attributes-aware words to connect similar people. Accordingly, we propose a Neural Personal Discrimination (NPD) approach to address above challenges by determining personal attributes from posts, and connecting relevant posts with similar attributes to jointly learn their emotions. In particular, we employ adversarial discriminators to determine the personal attributes, with attention mechanisms to aggregate attributes-aware words. In this way, social correlationship among different posts can be better addressed. Experimental results show the usefulness of personal attributes, and the effectiveness of our proposed NPD approach in capturing such personal attributes with significant gains over the state-of-the-art models. D19-1552 D19-1552.Attachment.zip @@ -6411,7 +6410,7 @@ JunXie QiSu XuSun - 5507–5516 + 5508–5517 The task of unsupervised sentiment modification aims to reverse the sentiment polarity of the input text while preserving its semantic content without any parallel data. Most previous work follows a two-step process. They first separate the content from the original sentiment, and then directly generate text with the target sentiment only based on the content produced by the first step. However, the second step bears both the target sentiment addition and content reconstruction, thus resulting in a lack of specific information like proper nouns in the generated text. To remedy this, we propose a specificity-driven cascading approach in this work, which can effectively increase the specificity of the generated text and further improve content preservation. In addition, we propose a more reasonable metric to evaluate sentiment modification. The experiments show that our approach outperforms competitive baselines by a large margin, which achieves 11% and 38% relative improvements of the overall metric on the Yelp and Amazon datasets, respectively. D19-1553 10.18653/v1/D19-1553 @@ -6424,7 +6423,7 @@ QiZeng YunLiang XuSun - 5517–5526 + 5518–5527 Recent work has shown that current text classification models are fragile and sensitive to simple perturbations. In this work, we propose a novel adversarial training approach, LexicalAT, to improve the robustness of current classification models. The proposed approach consists of a generator and a classifier. The generator learns to generate examples to attack the classifier while the classifier learns to defend these attacks. Considering the diversity of attacks, the generator uses a large-scale lexical knowledge base, WordNet, to generate attacking examples by replacing some words in training examples with their synonyms (e.g., sad and unhappy), neighbor words (e.g., fox and wolf), or super-superior words (e.g., chair and armchair). Due to the discrete generation step in the generator, we use policy gradient, a reinforcement learning approach, to train the two modules. Experiments show LexicalAT outperforms strong baselines and reduces test errors on various neural networks, including CNN, RNN, and BERT. D19-1554 10.18653/v1/D19-1554 @@ -6433,7 +6432,7 @@ Leveraging Structural and Semantic Correspondence for Attribute-Oriented Aspect Sentiment Discovery ZheZhang MunindarSingh - 5527–5537 + 5528–5538 Opinionated text often involves attributes such as authorship and location that influence the sentiments expressed for different aspects. We posit that structural and semantic correspondence is both prevalent in opinionated text, especially when associated with attributes, and crucial in accurately revealing its latent aspect and sentiment structure. However, it is not recognized by existing approaches. We propose Trait, an unsupervised probabilistic model that discovers aspects and sentiments from text and associates them with different attributes. To this end, Trait infers and leverages structural and semantic correspondence using a Markov Random Field. We show empirically that by incorporating attributes explicitly Trait significantly outperforms state-of-the-art baselines both by generating attribute profiles that accord with our intuitions, as shown via visualization, and yielding topics of greater semantic cohesion. D19-1555 10.18653/v1/D19-1555 @@ -6445,7 +6444,7 @@ Florenced’Alché-Buc SlimEssid ChloéClavel - 5538–5547 + 5539–5548 The task of predicting fine grained user opinion based on spontaneous spoken language is a key problem arising in the development of Computational Agents as well as in the development of social network based opinion miners. Unfortunately, gathering reliable data on which a model can be trained is notoriously difficult and existing works rely only on coarsely labeled opinions. In this work we aim at bridging the gap separating fine grained opinion models already developed for written language and coarse grained models developed for spontaneous multimodal opinion mining. We take advantage of the implicit hierarchical structure of opinions to build a joint fine and coarse grained opinion model that exploits different views of the opinion expression. The resulting model shares some properties with attention-based models and is shown to provide competitive results on a recently released multimodal fine grained annotated corpus. D19-1556 D19-1556.Attachment.pdf @@ -6456,7 +6455,7 @@ PrathushaK Sarma YingyuLiang WilliamSethares - 5548–5557 + 5549–5558 This paper proposes a way to improve the performance of existing algorithms for text classification in domains with strong language semantics. A proposed domain adaptation layer learns weights to combine a generic and a domain specific (DS) word embedding into a domain adapted (DA) embedding. The DA word embeddings are then used as inputs to a generic encoder + classifier framework to perform a downstream task such as classification. This adaptation layer is particularly suited to data sets that are modest in size, and which are, therefore, not ideal candidates for (re)training a deep neural network architecture. Results on binary and multi-class classification tasks using popular encoder architectures, including current state-of-the-art methods (with and without the shallow adaptation layer) show the effectiveness of the proposed approach. D19-1557 D19-1557.Attachment.zip @@ -6470,7 +6469,7 @@ HongleiGuo RenhongCheng ZhongSu - 5558–5567 + 5559–5568 Cross-domain sentiment classification has drawn much attention in recent years. Most existing approaches focus on learning domain-invariant representations in both the source and target domains, while few of them pay attention to the domain-specific information. Despite the non-transferability of the domain-specific information, simultaneously learning domain-dependent representations can facilitate the learning of domain-invariant representations. In this paper, we focus on aspect-level cross-domain sentiment classification, and propose to distill the domain-invariant sentiment features with the help of an orthogonal domain-dependent task, i.e. aspect detection, which is built on the aspects varying widely in different domains. We conduct extensive experiments on three public datasets and the experimental results demonstrate the effectiveness of our method. D19-1558 10.18653/v1/D19-1558 @@ -6483,7 +6482,7 @@ JinanXu YufengChen JieZhou - 5568–5579 + 5569–5580 Aspect based sentiment analysis (ABSA) aims to identify the sentiment polarity towards the given aspect in a sentence, while previous models typically exploit an aspect-independent (weakly associative) encoder for sentence representation generation. In this paper, we propose a novel Aspect-Guided Deep Transition model, named AGDT, which utilizes the given aspect to guide the sentence encoding from scratch with the specially-designed deep transition architecture. Furthermore, an aspect-oriented objective is designed to enforce AGDT to reconstruct the given aspect with the generated sentence representation. In doing so, our AGDT can accurately generate aspect-specific sentence representation, and thus conduct more accurate sentiment predictions. Experimental results on multiple SemEval datasets demonstrate the effectiveness of our proposed approach, which significantly outperforms the best reported results with the same setting. D19-1559 10.18653/v1/D19-1559 @@ -6498,7 +6497,7 @@ MinZhang XiaozhongLiu GuodongZhou - 5580–5589 + 5581–5590 Recently, neural networks have shown promising results on Document-level Aspect Sentiment Classification (DASC). However, these approaches often offer little transparency w.r.t. their inner working mechanisms and lack interpretability. In this paper, to simulating the steps of analyzing aspect sentiment in a document by human beings, we propose a new Hierarchical Reinforcement Learning (HRL) approach to DASC. This approach incorporates clause selection and word selection strategies to tackle the data noise problem in the task of DASC. First, a high-level policy is proposed to select aspect-relevant clauses and discard noisy clauses. Then, a low-level policy is proposed to select sentiment-relevant words and discard noisy words inside the selected clauses. Finally, a sentiment rating predictor is designed to provide reward signals to guide both clause and word selection. Experimental results demonstrate the impressive effectiveness of the proposed approach to DASC over the state-of-the-art baselines. D19-1560 10.18653/v1/D19-1560 @@ -6516,7 +6515,7 @@ MichalJacovi RanitAharonov NoamSlonim - 5590–5600 + 5591–5601 In Natural Language Understanding, the task of response generation is usually focused on responses to short texts, such as tweets or a turn in a dialog. Here we present a novel task of producing a critical response to a long argumentative text, and suggest a method based on general rebuttal arguments to address it. We do this in the context of the recently-suggested task of listening comprehension over argumentative content: given a speech on some specified topic, and a list of relevant arguments, the goal is to determine which of the arguments appear in the speech. The general rebuttals we describe here (in English) overcome the need for topic-specific arguments to be provided, by proving to be applicable for a large set of topics. This allows creating responses beyond the scope of topics for which specific arguments are available. All data collected during this work is freely available for research. D19-1561 D19-1561.Attachment.zip @@ -6525,7 +6524,7 @@ Rethinking Attribute Representation and Injection for Sentiment Classification Reinald KimAmplayo - 5601–5612 + 5602–5613 Text attributes, such as user and product information in product reviews, have been used to improve the performance of sentiment classification models. The de facto standard method is to incorporate them as additional biases in the attention mechanism, and more performance gains are achieved by extending the model architecture. In this paper, we show that the above method is the least effective way to represent and inject attributes. To demonstrate this hypothesis, unlike previous models with complicated architectures, we limit our base model to a simple BiLSTM with attention classifier, and instead focus on how and where the attributes should be incorporated in the model. We propose to represent attributes as chunk-wise importance weight matrices and consider four locations in the model (i.e., embedding, encoding, attention, classifier) to inject attributes. Experiments show that our proposed method achieves significant improvements over the standard approach and that attention mechanism is the worst location to inject attributes, contradicting prior work. We also outperform the state-of-the-art despite our use of a simple base model. Finally, we show that these representations transfer well to other tasks. Model implementation and datasets are released here: https://github.com/rktamplayo/CHIM. D19-1562 10.18653/v1/D19-1562 @@ -6540,7 +6539,7 @@ MinYang RuifengXu RuibinMao - 5613–5623 + 5614–5624 Emotion cause analysis, which aims to identify the reasons behind emotions, is a key topic in sentiment analysis. A variety of neural network models have been proposed recently, however, these previous models mostly focus on the learning architecture with local textual information, ignoring the discourse and prior knowledge, which play crucial roles in human text comprehension. In this paper, we propose a new method to extract emotion cause with a hierarchical neural model and knowledge-based regularizations, which aims to incorporate discourse context information and restrain the parameters by sentiment lexicon and common knowledge. The experimental results demonstrate that our proposed method achieves the state-of-the-art performance on two public datasets in different languages (Chinese and English), outperforming a number of competitive baselines by at least 2.08% in F-measure. D19-1563 10.18653/v1/D19-1563 @@ -6556,7 +6555,7 @@ MichalJacovi RanitAharonov NoamSlonim - 5624–5634 + 5625–5635 We explore the task of automatic assessment of argument quality. To that end, we actively collected 6.3k arguments, more than a factor of five compared to previously examined data. Each argument was explicitly and carefully annotated for its quality. In addition, 14k pairs of arguments were annotated independently, identifying the higher quality argument in each pair. In spite of the inherent subjective nature of the task, both annotation schemes led to surprisingly consistent results. We release the labeled datasets to the community. Furthermore, we suggest neural methods based on a recently released language model, for argument ranking as well as for argument-pair classification. In the former task, our results are comparable to state-of-the-art; in the latter task our results significantly outperform earlier methods. D19-1564 D19-1564.Attachment.zip @@ -6569,7 +6568,7 @@ AlbertoBarrón-Cedeño RostislavPetrov PreslavNakov - 5635–5645 + 5636–5646 Propaganda aims at influencing people’s mindset with the purpose of advancing a specific agenda. Previous work has addressed propaganda detection at document level, typically labelling all articles from a propagandistic news outlet as propaganda. Such noisy gold labels inevitably affect the quality of any learning system trained on them. A further issue with most existing systems is the lack of explainability. To overcome these limitations, we propose a novel task: performing fine-grained analysis of texts by detecting all fragments that contain propaganda techniques as well as their type. In particular, we create a corpus of news articles manually annotated at fragment level with eighteen propaganda techniques and propose a suitable evaluation measure. We further design a novel multi-granularity neural network, and we show that it outperforms several strong BERT-based baselines. D19-1565 D19-1565.Attachment.zip @@ -6581,7 +6580,7 @@ Md ShadAkhtar AsifEkbal PushpakBhattacharyya - 5646–5656 + 5647–5657 In recent times, multi-modal analysis has been an emerging and highly sought-after field at the intersection of natural language processing, computer vision, and speech processing. The prime objective of such studies is to leverage the diversified information, (e.g., textual, acoustic and visual), for learning a model. The effective interaction among these modalities often leads to a better system in terms of performance. In this paper, we introduce a recurrent neural network based approach for the multi-modal sentiment and emotion analysis. The proposed model learns the inter-modal interaction among the participating modalities through an auto-encoder mechanism. We employ a context-aware attention module to exploit the correspondence among the neighboring utterances. We evaluate our proposed approach for five standard multi-modal affect analysis datasets. Experimental results suggest the efficacy of the proposed model for both sentiment and emotion analysis over various existing state-of-the-art systems. D19-1566 10.18653/v1/D19-1566 @@ -6590,7 +6589,7 @@ Sequential Learning of Convolutional Features for Effective Text Classification AvinashMadasu VijjiniAnvesh Rao - 5657–5666 + 5658–5667 Text classification has been one of the major problems in natural language processing. With the advent of deep learning, convolutional neural network (CNN) has been a popular solution to this task. However, CNNs which were first proposed for images, face many crucial challenges in the context of text processing, namely in their elementary blocks: convolution filters and max pooling. These challenges have largely been overlooked by the most existing CNN models proposed for text classification. In this paper, we present an experimental study on the fundamental blocks of CNNs in text categorization. Based on this critique, we propose Sequential Convolutional Attentive Recurrent Network (SCARN). The proposed SCARN model utilizes both the advantages of recurrent and convolutional structures efficiently in comparison to previously proposed recurrent convolutional models. We test our model on different text classification datasets across tasks like sentiment analysis and question classification. Extensive experiments establish that SCARN outperforms other recurrent convolutional architectures with significantly less parameters. Furthermore, SCARN achieves better performance compared to equally large various deep CNN and LSTM architectures. D19-1567 10.18653/v1/D19-1567 @@ -6600,7 +6599,7 @@ EsinDurmus FaisalLadhak ClaireCardie - 5667–5677 + 5668–5678 Research in the social sciences and psychology has shown that the persuasiveness of an argument depends not only the language employed, but also on attributes of the source/communicator, the audience, and the appropriateness and strength of the argument’s claims given the pragmatic and discourse context of the argument. Among these characteristics of persuasive arguments, prior work in NLP does not explicitly investigate the effect of the pragmatic and discourse context when determining argument quality. This paper presents a new dataset to initiate the study of this aspect of argumentation: it consists of a diverse collection of arguments covering 741 controversial topics and comprising over 47,000 claims. We further propose predictive models that incorporate the pragmatic and discourse context of argumentative claims and show that they outperform models that rely only on claim-specific linguistic features for predicting the perceived impact of individual claims within a particular line of argument. D19-1568 10.18653/v1/D19-1568 @@ -6612,7 +6611,7 @@ SamuelMensah YongyiMao XudongLiu - 5678–5687 + 5679–5688 We propose a method based on neural networks to identify the sentiment polarity of opinion words expressed on a specific aspect of a sentence. Although a large majority of works typically focus on leveraging the expressive power of neural networks in handling this task, we explore the possibility of integrating dependency trees with neural networks for representation learning. To this end, we present a convolution over a dependency tree (CDT) model which exploits a Bi-directional Long Short Term Memory (Bi-LSTM) to learn representations for features of a sentence, and further enhance the embeddings with a graph convolutional network (GCN) which operates directly on the dependency tree of the sentence. Our approach propagates both contextual and dependency information from opinion words to aspect words, offering discriminative properties for supervision. Experimental results ranks our approach as the new state-of-the-art in aspect-based sentiment classification. D19-1569 10.18653/v1/D19-1569 @@ -6624,7 +6623,7 @@ GuopingHuang ConghuiZhu TiejunZhao - 5688–5694 + 5689–5695 Many Data Augmentation (DA) methods have been proposed for neural machine translation. Existing works measure the superiority of DA methods in terms of their performance on a specific test set, but we find that some DA methods do not exhibit consistent improvements across translation tasks. Based on the observation, this paper makes an initial attempt to answer a fundamental question: what benefits, which are consistent across different methods and tasks, does DA in general obtain? Inspired by recent theoretic advances in deep learning, the paper understands DA from two perspectives towards the generalization ability of a model: input sensitivity and prediction margin, which are defined independent of specific test set thereby may lead to findings with relatively low variance. Extensive experiments show that relatively consistent benefits across five DA methods and four translation tasks are achieved regarding both perspectives. D19-1570 D19-1570.Attachment.pdf @@ -6635,7 +6634,7 @@ KyraYee YannDauphin MichaelAuli - 5695–5700 + 5696–5701 Previous work on neural noisy channel modeling relied on latent variable models that incrementally process the source and target sentence. This makes decoding decisions based on partial source prefixes even though the full source is available. We pursue an alternative approach based on standard sequence to sequence models which utilize the entire source. These models perform remarkably well as channel models, even though they have neither been trained on, nor designed to factor over incomplete target sentences. Experiments with neural language models trained on billions of words show that noisy channel models can outperform a direct model by up to 3.2 BLEU on WMT’17 German-English translation. We evaluate on four language-pairs and our channel models consistently outperform strong alternatives such right-to-left reranking models and ensembles of direct models. D19-1571 D19-1571.Attachment.pdf @@ -6649,7 +6648,7 @@ MarcinKadras SylvainGugger JeremyHoward - 5701–5706 + 5702–5707 Pretrained language models are promising particularly for low-resource languages as they only require unlabelled data. However, training existing models requires huge amounts of compute, while pretrained cross-lingual models often underperform on low-resource languages. We propose Multi-lingual language model Fine-Tuning (MultiFiT) to enable practitioners to train and fine-tune language models efficiently in their own language. In addition, we propose a zero-shot method using an existing pretrained cross-lingual model. We evaluate our methods on two widely used cross-lingual classification datasets where they outperform models pretrained on orders of magnitude more data and compute. We release all models and code. D19-1572 D19-1572.Attachment.zip @@ -6664,7 +6663,7 @@ TaoQin LiweiWang Tie-YanLiu - 5707–5712 + 5708–5713 Due to the unparallelizable nature of the autoregressive factorization, AutoRegressive Translation (ART) models have to generate tokens sequentially during decoding and thus suffer from high inference latency. Non-AutoRegressive Translation (NART) models were proposed to reduce the inference time, but could only achieve inferior translation accuracy. In this paper, we proposed a novel approach to leveraging the hints from hidden states and word alignments to help the training of NART models. The results achieve significant improvement over previous NART models for the WMT14 En-De and De-En datasets and are even comparable to a strong LSTM-based ART baseline but one order of magnitude faster in inference. D19-1573 D19-1573.Attachment.zip @@ -6675,7 +6674,7 @@ AdamFisch JiangGuo ReginaBarzilay - 5713–5719 + 5714–5720 This paper explores the task of leveraging typology in the context of cross-lingual dependency parsing. While this linguistic information has shown great promise in pre-neural parsing, results for neural architectures have been mixed. The aim of our investigation is to better understand this state-of-the-art. Our main findings are as follows: 1) The benefit of typological information is derived from coarsely grouping languages into syntactically-homogeneous clusters rather than from learning to leverage variations along individual typological dimensions in a compositional manner; 2) Typology consistent with the actual corpus statistics yields better transfer performance; 3) Typological similarity is only a rough proxy of cross-lingual transferability with respect to parsing. D19-1574 D19-1574.Attachment.zip @@ -6688,7 +6687,7 @@ JiangGuo YijiaLiu TingLiu - 5720–5726 + 5721–5727 This paper investigates the problem of learning cross-lingual representations in a contextual space. We propose Cross-Lingual BERT Transformation (CLBT), a simple and efficient approach to generate cross-lingual contextualized word embeddings based on publicly available pre-trained BERT models (Devlin et al., 2018). In this approach, a linear transformation is learned from contextual word alignments to align the contextualized embeddings independently trained in different languages. We demonstrate the effectiveness of this approach on zero-shot cross-lingual transfer parsing. Experiments show that our embeddings substantially outperform the previous state-of-the-art that uses static embeddings. We further compare our approach with XLM (Lample and Conneau, 2019), a recently proposed cross-lingual language model trained with massive parallel data, and achieve highly competitive results. D19-1575 D19-1575.Attachment.pdf @@ -6700,7 +6699,7 @@ GeWang YongJiang KeweiTu - 5727–5732 + 5728–5733 The key to multilingual grammar induction is to couple grammar parameters of different languages together by exploiting the similarity between languages. Previous work relies on linguistic phylogenetic knowledge to specify similarity between languages. In this work, we propose a novel universal grammar induction approach that represents language identities with continuous vectors and employs a neural network to predict grammar parameters based on the representation. Without any prior linguistic phylogenetic knowledge, we automatically capture similarity between languages with the vector representations and softly tie the grammar parameters of different languages. In our experiments, we apply our approach to 15 languages across 8 language families and subfamilies in the Universal Dependency Treebank dataset, and we observe substantial performance gain on average over monolingual and multilingual baselines. D19-1576 10.18653/v1/D19-1576 @@ -6712,7 +6711,7 @@ LawrenceWolf-Sonkin HannaWallach RyanCotterell - 5733–5738 + 5734–5739 Many of the world’s languages employ grammatical gender on the lexeme. For instance, in Spanish, house “casa” is feminine, whereas the word for paper “papel” is masculine. To a speaker of a genderless language, this categorization seems to exist with neither rhyme nor reason. But, is the association of nouns to gender classes truly arbitrary? In this work, we present the first large-scale investigation of the arbitrariness of gender assignment that uses canonical correlation analysis as a method for correlating the gender of inanimate nouns with their lexical semantic meaning. We find that the gender systems of 18 languages exhibit a significant correlation with an externally grounded definition of lexical semantics. D19-1577 10.18653/v1/D19-1577 @@ -6722,7 +6721,7 @@ VinodkumarPrabhakaran BenHutchinson MargaretMitchell - 5739–5744 + 5740–5745 Data-driven statistical Natural Language Processing (NLP) techniques leverage large amounts of language data to build models that can understand language. However, most language data reflect the public discourse at the time the data was produced, and hence NLP models are susceptible to learning incidental associations around named referents at a particular point in time, in addition to general linguistic meaning. An NLP system designed to model notions such as sentiment and toxicity should ideally produce scores that are independent of the identity of such entities mentioned in text and their social associations. For example, in a general purpose sentiment analysis system, a phrase such as I hate Katy Perry should be interpreted as having the same sentiment as I hate Taylor Swift. Based on this idea, we propose a generic evaluation framework, Perturbation Sensitivity Analysis, which detects unintended model biases related to named entities, and requires no new annotations or corpora. We demonstrate the utility of this analysis by employing it on two different NLP models — a sentiment model and a toxicity model — applied on online comments in English language from four different genres. D19-1578 10.18653/v1/D19-1578 @@ -6731,7 +6730,7 @@ Automatically Inferring Gender Associations from Language SerinaChang KathyMcKeown - 5745–5751 + 5746–5752 In this paper, we pose the question: do people talk about women and men in different ways? We introduce two datasets and a novel integration of approaches for automatically inferring gender associations from language, discovering coherent word clusters, and labeling the clusters for the semantic concepts they represent. The datasets allow us to compare how people write about women and men in two different settings – one set draws from celebrity news and the other from student reviews of computer science professors. We demonstrate that there are large-scale differences in the ways that people talk about women and men and that these differences vary across domains. Human evaluations show that our methods significantly outperform strong baselines. D19-1579 D19-1579.Attachment.pdf @@ -6750,7 +6749,7 @@ ArinehMirinjian XiangRen MortezaDehghani - 5752–5756 + 5753–5757 Official reports of hate crimes in the US are under-reported relative to the actual number of such incidents. Further, despite statistical approximations, there are no official reports from a large number of US cities regarding incidents of hate. Here, we first demonstrate that event extraction and multi-instance learning, applied to a corpus of local news articles, can be used to predict instances of hate crime. We then use the trained model to detect incidents of hate in cities for which the FBI lacks statistics. Lastly, we train models on predicting homicide and kidnapping, compare the predictions to FBI reports, and establish that incidents of hate are indeed under-reported, compared to other types of crimes, in local press. D19-1580 10.18653/v1/D19-1580 @@ -6760,7 +6759,7 @@ JunSaito YugoMurawaki SadaoKurohashi - 5757–5764 + 5758–5765 Recognizing affective events that trigger positive or negative sentiment has a wide range of natural language processing applications but remains a challenging problem mainly because the polarity of an event is not necessarily predictable from its constituent words. In this paper, we propose to propagate affective polarity using discourse relations. Our method is simple and only requires a very small seed lexicon and a large raw corpus. Our experiments using Japanese data show that our method learns affective events effectively without manually labeled data. It also improves supervised learning results when labeled data are small. D19-1581 10.18653/v1/D19-1581 @@ -6772,7 +6771,7 @@ XiangbinMeng JiafengGuo XueqiCheng - 5765–5769 + 5766–5770 Syntactic relations are broadly used in many NLP tasks. For event detection, syntactic relation representations based on dependency tree can better capture the interrelations between candidate trigger words and related entities than sentence representations. But, existing studies only use first-order syntactic relations (i.e., the arcs) in dependency trees to identify trigger words. For this reason, this paper proposes a new method for event detection, which uses a dependency tree based graph convolution network with aggregative attention to explicitly model and aggregate multi-order syntactic representations in sentences. Experimental comparison with state-of-the-art baselines shows the superiority of the proposed method. D19-1582 10.18653/v1/D19-1582 @@ -6783,7 +6782,7 @@ NitishaJain ParamitaMirza GerhardWeikum - 5770–5775 + 5771–5776 Scalar implicatures are language features that imply the negation of stronger statements, e.g., “She was married twice” typically implicates that she was not married thrice. In this paper we discuss the importance of scalar implicatures in the context of textual information extraction. We investigate how textual features can be used to predict whether a given text segment mentions all objects standing in a certain relationship with a certain subject. Preliminary results on Wikipedia indicate that this prediction is feasible, and yields informative assessments. D19-1583 10.18653/v1/D19-1583 @@ -6799,7 +6798,7 @@ MaosongSun JieZhou XiangRen - 5776–5782 + 5777–5783 Existing event extraction methods classify each argument role independently, ignoring the conceptual correlations between different argument roles. In this paper, we propose a Hierarchical Modular Event Argument Extraction (HMEAE) model, to provide effective inductive bias from the concept hierarchy of event argument roles. Specifically, we design a neural module network for each basic unit of the concept hierarchy, and then hierarchically compose relevant unit modules with logical operations into a role-oriented modular network to classify a specific argument role. As many argument roles share the same high-level unit module, their correlation can be utilized to extract specific event arguments better. Experiments on real-world datasets show that HMEAE can effectively leverage useful knowledge from the concept hierarchy and significantly outperform the state-of-the-art baselines. The source code can be obtained from https://github.com/thunlp/HMEAE. D19-1584 D19-1584.Attachment.zip @@ -6811,7 +6810,7 @@ UlmeWennberg YiLuan HannanehHajishirzi - 5783–5788 + 5784–5789 We examine the capabilities of a unified, multi-task framework for three information extraction tasks: named entity recognition, relation extraction, and event extraction. Our framework (called DyGIE++) accomplishes all tasks by enumerating, refining, and scoring text spans designed to capture local (within-sentence) and global (cross-sentence) context. Our framework achieves state-of-the-art results across all tasks, on four datasets from a variety of domains. We perform experiments comparing different techniques to construct span representations. Contextualized embeddings like BERT perform well at capturing relationships among entities in the same or adjacent sentences, while dynamic span graph updates model long-range cross-sentence relationships. For instance, propagating span representations via predicted coreference links can enable the model to disambiguate challenging entity mentions. Our code is publicly available at https://github.com/dwadden/dygiepp and can be easily adapted for new tasks or datasets. D19-1585 D19-1585.Attachment.pdf @@ -6821,7 +6820,7 @@ Next Sentence Prediction helps Implicit Discourse Relation Classification within and across Domains WeiShi VeraDemberg - 5789–5795 + 5790–5796 Implicit discourse relation classification is one of the most difficult tasks in discourse parsing. Previous studies have generally focused on extracting better representations of the relational arguments. In order to solve the task, it is however additionally necessary to capture what events are expected to cause or follow each other. Current discourse relation classifiers fall short in this respect. We here show that this shortcoming can be effectively addressed by using the bidirectional encoder representation from transformers (BERT) proposed by Devlin et al. (2019), which were trained on a next-sentence prediction task, and thus encode a representation of likely next sentences. The BERT-based model outperforms the current state of the art in 11-way classification by 8% points on the standard PDTB dataset. Our experiments also demonstrate that the model can be successfully ported to other domains: on the BioDRB dataset, the model outperforms the state of the art system around 15% points. D19-1586 10.18653/v1/D19-1586 @@ -6834,7 +6833,7 @@ HidetakaKamigaito ManabuOkumura MasaakiNagata - 5796–5801 + 5797–5802 Rhetorical Structure Theory (RST) parsing is crucial for many downstream NLP tasks that require a discourse structure for a text. Most of the previous RST parsers have been based on supervised learning approaches. That is, they require an annotated corpus of sufficient size and quality, and heavily rely on the language and domain dependent corpus. In this paper, we present two language-independent unsupervised RST parsing methods based on dynamic programming. The first one builds the optimal tree in terms of a dissimilarity score function that is defined for splitting a text span into smaller ones. The second builds the optimal tree in terms of a similarity score function that is defined for merging two adjacent spans into a large one. Experimental results on English and German RST treebanks showed that our parser based on span merging achieved the best score, around 0.8 F_1 score, which is close to the scores of the previous supervised parsers. D19-1587 10.18653/v1/D19-1587 @@ -6845,7 +6844,7 @@ OmerLevy LukeZettlemoyer DanielWeld - 5802–5807 + 5803–5808 We apply BERT to coreference resolution, achieving a new state of the art on the GAP (+11.5 F1) and OntoNotes (+3.9 F1) benchmarks. A qualitative analysis of model predictions indicates that, compared to ELMo and BERT-base, BERT-large is particularly better at distinguishing between related but distinct entities (e.g., President and CEO), but that there is still room for improvement in modeling document-level context, conversations, and mention paraphrasing. We will release all code and trained models upon publication. D19-1588 10.18653/v1/D19-1588 @@ -6854,7 +6853,7 @@ Linguistic Versus Latent Relations for Modeling Coherent Flow in Paragraphs DongyeopKang EduardHovy - 5808–5814 + 5809–5815 Generating a long, coherent text such as a paragraph requires a high-level control of different levels of relations between sentences (e.g., tense, coreference). We call such a logical connection between sentences as a (paragraph) flow. In order to produce a coherent flow of text, we explore two forms of intersentential relations in a paragraph: one is a human-created linguistical relation that forms a structure (e.g., discourse tree) and the other is a relation from latent representation learned from the sentences themselves. Our two proposed models incorporate each form of relations into document-level language models: the former is a supervised model that jointly learns a language model as well as discourse relation prediction, and the latter is an unsupervised model that is hierarchically conditioned by a recurrent neural network (RNN) over the latent information. Our proposed models with both forms of relations outperform the baselines in partially conditioned paragraph generation task. Our codes and data are publicly available. D19-1589 D19-1589.Attachment.pdf @@ -6867,7 +6866,7 @@ KentaroTorisawa Jong-HoonOh JulienKloetzer - 5815–5821 + 5816–5822 We propose new BERT-based methods for recognizing event causality such as “smoke cigarettes” –> “die of lung cancer” written in web texts. In our methods, we grasp each annotator’s policy by training multiple classifiers, each of which predicts the labels given by a single annotator, and combine the resulting classifiers’ outputs to predict the final labels determined by majority vote. Furthermore, we investigate the effect of supplying background knowledge to our classifiers. Since BERT models are pre-trained with a large corpus, some sort of background knowledge for event causality may be learned during pre-training. Our experiments with a Japanese dataset suggest that this is actually the case: Performance improved when we pre-trained the BERT models with web texts containing a large number of event causalities instead of Wikipedia articles or randomly sampled web texts. However, this effect was limited. Therefore, we further improved performance by simply adding texts related to an input causality candidate as background knowledge to the input of the BERT models. We believe these findings indicate a promising future research direction. D19-1590 10.18653/v1/D19-1590 @@ -6877,7 +6876,7 @@ JiXin JimmyLin YaoliangYu - 5822–5829 + 5823–5830 Memory neurons of long short-term memory (LSTM) networks encode and process information in powerful yet mysterious ways. While there has been work to analyze their behavior in carrying low-level information such as linguistic properties, how they directly contribute to label prediction remains unclear. We find inspiration from biologists and study the affinity between individual neurons and labels, propose a novel metric to quantify the sensitivity of neurons to each label, and conduct experiments to show the validity of our proposed metric. We discover that some neurons are trained to specialize on a subset of labels, and while dropping an arbitrary neuron has little effect on the overall accuracy of the model, dropping label-specialized neurons predictably and significantly degrades prediction accuracy on the associated label. We further examine the consistency of neuron-label affinity across different models. These observations provide insight into the inner mechanisms of LSTMs. D19-1591 10.18653/v1/D19-1591 @@ -6887,7 +6886,7 @@ Martenvan Schijndel AaronMueller TalLinzen - 5830–5836 + 5831–5837 Recurrent neural networks can learn to predict upcoming words remarkably well on average; in syntactically complex contexts, however, they often assign unexpectedly high probabilities to ungrammatical words. We investigate to what extent these shortcomings can be mitigated by increasing the size of the network and the corpus on which it is trained. We find that gains from increasing network size are minimal beyond a certain point. Likewise, expanding the training corpus yields diminishing returns; we estimate that the training corpus would need to be unrealistically large for the models to match human performance. A comparison to GPT and BERT, Transformer-based models trained on billions of words, reveals that these models perform even more poorly than our LSTMs in some constructions. Our results make the case for more data efficient architectures. D19-1592 D19-1592.Attachment.zip @@ -6900,7 +6899,7 @@ FelixHill Daniel M.Low AndersSøgaard - 5837–5844 + 5838–5845 Representational Similarity Analysis (RSA) is a technique developed by neuroscientists for comparing activity patterns of different measurement modalities (e.g., fMRI, electrophysiology, behavior). As a framework, RSA has several advantages over existing approaches to interpretation of language encoders based on probing or diagnostic classification: namely, it does not require large training samples, is not prone to overfitting, and it enables a more transparent comparison between the representational geometries of different models and modalities. We demonstrate the utility of RSA by establishing a previously unknown correspondence between widely-employed pretrained language encoders and human processing difficulty via eye-tracking data, showcasing its potential in the interpretability toolbox for neural models. D19-1593 D19-1593.Attachment.pdf @@ -6913,7 +6912,7 @@ KeithHall ChrisDyer JonathanBrennan - 5845–5851 + 5846–5852 Domain-specific training typically makes NLP systems work better. We show that this extends to cognitive modeling as well by relating the states of a neural phrase-structure parser to electrophysiological measures from human participants. These measures were recorded as participants listened to a spoken recitation of the same literary text that was supplied as input to the neural parser. Given more training data, the system derives a better cognitive model — but only when the training examples come from the same textual genre. This finding is consistent with the idea that humans adapt syntactic expectations to particular genres during language comprehension (Kaan and Chun, 2018; Branigan and Pickering, 2017). D19-1594 10.18653/v1/D19-1594 @@ -6923,7 +6922,7 @@ StevenDerby PaulMiller BarryDevereux - 5852–5858 + 5853–5859 Feature norm datasets of human conceptual knowledge, collected in surveys of human volunteers, yield highly interpretable models of word meaning and play an important role in neurolinguistic research on semantic cognition. However, these datasets are limited in size due to practical obstacles associated with exhaustively listing properties for a large number of words. In contrast, the development of distributional modelling techniques and the availability of vast text corpora have allowed researchers to construct effective vector space models of word meaning over large lexicons. However, this comes at the cost of interpretable, human-like information about word meaning. We propose a method for mapping human property knowledge onto a distributional semantic space, which adapts the word2vec architecture to the task of modelling concept features. Our approach gives a measure of concept and feature affinity in a single semantic space, which makes for easy and efficient ranking of candidate human-derived semantic properties for arbitrary words. We compare our model with a previous approach, and show that it performs better on several evaluation tasks. Finally, we discuss how our method could be used to develop efficient sampling techniques to extend existing feature norm datasets in a reliable way. D19-1595 10.18653/v1/D19-1595 @@ -6935,7 +6934,7 @@ AjayDivakaran StefanLee GiedriusBurachas - 5859–5864 + 5860–5865 While models for Visual Question Answering (VQA) have steadily improved over the years, interacting with one quickly reveals that these models lack consistency. For instance, if a model answers “red” to “What color is the balloon?”, it might answer “no” if asked, “Is the balloon red?”. These responses violate simple notions of entailment and raise questions about how effectively VQA models ground language. In this work, we introduce a dataset, ConVQA, and metrics that enable quantitative evaluation of consistency in VQA. For a given observable fact in an image (e.g. the balloon’s color), we generate a set of logically consistent question-answer (QA) pairs (e.g. Is the balloon red?) and also collect a human-annotated set of common-sense based consistent QA pairs (e.g. Is the balloon the same color as tomato sauce?). Further, we propose a consistency-improving data augmentation module, a Consistency Teacher Module (CTM). CTM automatically generates entailed (or similar-intent) questions for a source QA pair and fine-tunes the VQA model if the VQA’s answer to the entailed question is consistent with the source QA pair. We demonstrate that our CTM-based training improves the consistency of VQA models on the Con-VQA datasets and is a strong baseline for further research. D19-1596 D19-1596.Attachment.pdf @@ -6951,7 +6950,7 @@ LinZhou XinyuDai YuzhongQu - 5865–5870 + 5866–5871 Scenario-based question answering (SQA) has attracted increasing research attention. It typically requires retrieving and integrating knowledge from multiple sources, and applying general knowledge to a specific case described by a scenario. SQA widely exists in the medical, geography, and legal domains—both in practice and in the exams. In this paper, we introduce the GeoSQA dataset. It consists of 1,981 scenarios and 4,110 multiple-choice questions in the geography domain at high school level, where diagrams (e.g., maps, charts) have been manually annotated with natural language descriptions to benefit NLP research. Benchmark results on a variety of state-of-the-art methods for question answering, textual entailment, and reading comprehension demonstrate the unique challenges presented by SQA for future research. D19-1597 10.18653/v1/D19-1597 @@ -6961,7 +6960,7 @@ MatthewLe Y-LanBoureau MaximilianNickel - 5871–5876 + 5872–5877 Theory of mind, i.e., the ability to reason about intents and beliefs of agents is an important task in artificial intelligence and central to resolving ambiguous references in natural language dialogue. In this work, we revisit the evaluation of theory of mind through question answering. We show that current evaluation methods are flawed and that existing benchmark tasks can be solved without theory of mind due to dataset biases. Based on prior work, we propose an improved evaluation protocol and dataset in which we explicitly control for data regularities via a careful examination of the answer space. We show that state-of-the-art methods which are successful on existing benchmarks fail to solve theory-of-mind tasks in our proposed approach. D19-1598 D19-1598.Attachment.pdf @@ -6974,7 +6973,7 @@ XiaofeiMa RameshNallapati BingXiang - 5877–5881 + 5878–5882 BERT model has been successfully applied to open-domain QA tasks. However, previous work trains BERT by viewing passages corresponding to the same question as independent training instances, which may cause incomparable scores for answers from different passages. To tackle this issue, we propose a multi-passage BERT model to globally normalize answer scores across all passages of the same question, and this change enables our QA model find better answers by utilizing more passages. In addition, we find that splitting articles into passages with the length of 100 words by sliding window improves performance by 4%. By leveraging a passage ranker to select high-quality passages, multi-passage BERT gains additional 2%. Experiments on four standard benchmarks showed that our multi-passage BERT outperforms all state-of-the-art models on all benchmarks. In particular, on the OpenSQuAD dataset, our model gains 21.4% EM and 21.5% F1 over all non-BERT models, and 5.8% EM and 6.5% F1 over BERT-based models. D19-1599 10.18653/v1/D19-1599 @@ -6989,7 +6988,7 @@ WentaoMa ShijinWang GuopingHu - 5882–5888 + 5883–5889 Machine Reading Comprehension (MRC) has become enormously popular recently and has attracted a lot of attention. However, the existing reading comprehension datasets are mostly in English. In this paper, we introduce a Span-Extraction dataset for Chinese machine reading comprehension to add language diversities in this area. The dataset is composed by near 20,000 real questions annotated on Wikipedia paragraphs by human experts. We also annotated a challenge set which contains the questions that need comprehensive understanding and multi-sentence inference throughout the context. We present several baseline systems as well as anonymous submissions for demonstrating the difficulties in this dataset. With the release of the dataset, we hosted the Second Evaluation Workshop on Chinese Machine Reading Comprehension (CMRC 2018). We hope the release of the dataset could further accelerate the Chinese machine reading comprehension research. Resources are available: https://github.com/ymcui/cmrc2018 D19-1600 10.18653/v1/D19-1600 @@ -7000,7 +6999,7 @@ SeungtaekChoi HaejuPark Seung-wonHwang - 5889–5894 + 5890–5895 This paper studies the problem of non-factoid question answering, where the answer may span over multiple sentences. Existing solutions can be categorized into representation- and interaction-focused approaches. We combine their complementary strength, by a hybrid approach allowing multi-granular interactions, but represented at word level, enabling an easy integration with strong word-level signals. Specifically, we propose MICRON: Multigranular Interaction for Contextualizing RepresentatiON, a novel approach which derives contextualized uni-gram representation from n-grams. Our contributions are as follows: First, we enable multi-granular matches between question and answer n-grams. Second, by contextualizing word representation with surrounding n-grams, MICRON can naturally utilize word-based signals for query term weighting, known to be effective in information retrieval. We validate MICRON in two public non-factoid question answering datasets: WikiPassageQA and InsuranceQA, showing our model achieves the state of the art among baselines with reported performances on both datasets. D19-1601 10.18653/v1/D19-1601 @@ -7015,7 +7014,7 @@ YajuanLyu KangLiu JunZhao - 5895–5900 + 5896–5901 Leveraging external knowledge is an emerging trend in machine comprehension task. Previous work usually utilizes knowledge graphs such as ConceptNet as external knowledge, and extracts triples from them to enhance the initial representation of the machine comprehension context. However, such method cannot capture the structural information in the knowledge graph. To this end, we propose a Structural Knowledge Graph-aware Network(SKG) model, constructing sub-graphs for entities in the machine comprehension context. Our method dynamically updates the representation of the knowledge according to the structural information of the constructed sub-graph. Experiments show that SKG achieves state-of-the-art performance on the ReCoRD dataset. D19-1602 10.18653/v1/D19-1602 @@ -7027,7 +7026,7 @@ PeterShaw MassimoNicosia YaseminAltun - 5901–5909 + 5902–5910 We present a novel approach to answering sequential questions based on structured objects such as knowledge bases or tables without using a logical form as an intermediate representation. We encode tables as graphs using a graph neural network model based on the Transformer architecture. The answers are then selected from the encoded graph using a pointer network. This model is appropriate for processing conversations around structured data, where the attention mechanism that selects the answers to a question can also be used to resolve conversational references. We demonstrate the validity of this approach with competitive results on the Sequential Question Answering (SQA) task. D19-1603 10.18653/v1/D19-1603 @@ -7038,7 +7037,7 @@ ShwetaGarg KartikMehta NikhilRasiwasia - 5910–5916 + 5911–5917 In this paper, we establish the effectiveness of using hard negatives, coupled with a siamese network and a suitable loss function, for the tasks of answer selection and answer triggering. We show that the choice of sampling strategy is key for achieving improved performance on these tasks. Evaluating on recent answer selection datasets - InsuranceQA, SelQA, and an internal QA dataset, we show that using hard negatives with relatively simple model architectures (bag of words and LSTM-CNN) drives significant performance gains. On InsuranceQA, this strategy alone improves over previously reported results by a minimum of 1.6 points in P@1. Using hard negatives with a Transformer encoder provides a further improvement of 2.3 points. Further, we propose to use quadruplet loss for answer triggering, with the aim of producing globally meaningful similarity scores. We show that quadruplet loss function coupled with the selection of hard negatives enables bag-of-words models to improve F1 score by 2.3 points over previous baselines, on SelQA answer triggering dataset. Our results provide key insights into answer selection and answer triggering tasks. D19-1604 D19-1604.Attachment.pdf @@ -7049,7 +7048,7 @@ AhmedElgohary DenisPeskov JordanBoyd-Graber - 5917–5923 + 5918–5924 Question answering is an AI-complete problem, but existing datasets lack key elements of language understanding such as coreference and ellipsis resolution. We consider sequential question answering: multiple questions are asked one-by-one in a conversation between a questioner and an answerer. Answering these questions is only possible through understanding the conversation history. We introduce the task of question-in-context rewriting: given the context of a conversation’s history, rewrite a context-dependent into a self-contained question with the same answer. We construct, CANARD, a dataset of 40,527 questions based on QuAC (Choi et al., 2018) and train Seq2Seq models for incorporating context into standalone questions. D19-1605 D19-1605.Attachment.zip @@ -7062,7 +7061,7 @@ AnaMarasović Noah A.Smith MattGardner - 5924–5931 + 5925–5932 Machine comprehension of texts longer than a single sentence often requires coreference resolution. However, most current reading comprehension benchmarks do not contain complex coreferential phenomena and hence fail to evaluate the ability of models to resolve coreference. We present a new crowdsourced dataset containing more than 24K span-selection questions that require resolving coreference among entities in over 4.7K English paragraphs from Wikipedia. Obtaining questions focused on such phenomena is challenging, because it is hard to avoid lexical cues that shortcut complex reasoning. We deal with this issue by using a strong baseline model as an adversary in the crowdsourcing loop, which helps crowdworkers avoid writing questions with exploitable surface cues. We show that state-of-the-art reading comprehension models perform significantly worse than humans on this benchmark—the best model performance is 70.5 F1, while the estimated human performance is 93.4 F1. D19-1606 10.18653/v1/D19-1606 @@ -7072,7 +7071,7 @@ Tsung-YuanHsu Chi-LiangLiu Hung-yiLee - 5932–5939 + 5933–5940 Because it is not feasible to collect training data for every language, there is a growing interest in cross-lingual transfer learning. In this paper, we systematically explore zero-shot cross-lingual transfer learning on reading comprehension tasks with language representation model pre-trained on multi-lingual corpus. The experimental results show that with pre-trained language representation zero-shot learning is feasible, and translating the source data into the target language is not necessary and even degrades the performance. We further explore what does the model learn in zero-shot setting. D19-1607 D19-1607.Attachment.pdf @@ -7084,7 +7083,7 @@ MattGardner KevinLin PeterClark - 5940–5945 + 5941–5946 We introduce the first open-domain dataset, called QuaRTz, for reasoning about textual qualitative relationships. QuaRTz contains general qualitative statements, e.g., “A sunscreen with a higher SPF protects the skin longer.”, twinned with 3864 crowdsourced situated questions, e.g., “Billy is wearing sunscreen with a lower SPF than Lucy. Who will be best protected from the sun?”, plus annotations of the properties being compared. Unlike previous datasets, the general knowledge is textual and not tied to a fixed set of relationships, and tests a system’s ability to comprehend and apply textual qualitative knowledge in a novel setting. We find state-of-the-art results are substantially (20%) below human performance, presenting an open challenge to the NLP community. D19-1608 10.18653/v1/D19-1608 @@ -7095,7 +7094,7 @@ LuhengHe KentonLee EmilyPitler - 5946–5951 + 5947–5952 Reading comprehension models have been successfully applied to extractive text answers, but it is unclear how best to generalize these models to abstractive numerical answers. We enable a BERT-based reading comprehension model to perform lightweight numerical reasoning. We augment the model with a predefined set of executable ‘programs’ which encompass simple arithmetic as well as extraction. Rather than having to learn to manipulate numbers directly, the model can pick a program and execute it. On the recent Discrete Reasoning Over Passages (DROP) dataset, designed to challenge reading comprehension models, we show a 33% absolute improvement by adding shallow programs. The model can learn to predict new operations when appropriate in a math word problem setting (Roy and Roth, 2015) with very few training examples. D19-1609 D19-1609.Attachment.zip @@ -7107,7 +7106,7 @@ Quan HungTran TrungBui DaisukeKihara - 5952–5958 + 5953–5959 Answer selection is an important research problem, with applications in many areas. Previous deep learning based approaches for the task mainly adopt the Compare-Aggregate architecture that performs word-level comparison followed by aggregation. In this work, we take a departure from the popular Compare-Aggregate architecture, and instead, propose a new gated self-attention memory network for the task. Combined with a simple transfer learning technique from a large-scale online corpus, our model outperforms previous methods by a large margin, achieving new state-of-the-art results on two standard answer selection datasets: TrecQA and WikiQA. D19-1610 10.18653/v1/D19-1610 @@ -7116,7 +7115,7 @@ Polly Want a Cracker: Analyzing Performance of Parroting on Paraphrase Generation Datasets Hong-RenMao Hung-YiLee - 5959–5967 + 5960–5968 Paraphrase generation is an interesting and challenging NLP task which has numerous practical applications. In this paper, we analyze datasets commonly used for paraphrase generation research, and show that simply parroting input sentences surpasses state-of-the-art models in the literature when evaluated on standard metrics. Our findings illustrate that a model could be seemingly adept at generating paraphrases, despite only making trivial changes to the input sentence or even none at all. D19-1611 D19-1611.Attachment.zip @@ -7126,7 +7125,7 @@ Query-focused Sentence Compression in Linear Time AbramHandler BrendanO’Connor - 5968–5974 + 5969–5975 Search applications often display shortened sentences which must contain certain query terms and must fit within the space constraints of a user interface. This work introduces a new transition-based sentence compression technique developed for such settings. Our query-focused method constructs length and lexically constrained compressions in linear time, by growing a subgraph in the dependency parse of a sentence. This theoretically efficient approach achieves an 11x empirical speedup over baseline ILP methods, while better reconstructing gold constrained shortenings. Such speedups help query-focused applications, because users are measurably hindered by interface lags. Additionally, our technique does not require an ILP solver or a GPU. D19-1612 D19-1612.Attachment.zip @@ -7138,7 +7137,7 @@ ShuyangLi JianmoNi JulianMcAuley - 5975–5981 + 5976–5982 Existing approaches to recipe generation are unable to create recipes for users with culinary preferences but incomplete knowledge of ingredients in specific dishes. We propose a new task of personalized recipe generation to help these users: expanding a name and incomplete ingredient details into complete natural-text instructions aligned with the user’s historical preferences. We attend on technique- and recipe-level representations of a user’s previously consumed recipes, fusing these ‘user-aware’ representations in an attention fusion layer to control recipe text generation. Experiments on a new dataset of 180K recipes and 700K interactions show our model’s ability to generate plausible and personalized recipes compared to non-personalized baselines. D19-1613 D19-1613.Attachment.pdf @@ -7148,7 +7147,7 @@ Generating Highly Relevant Questions JiazuoQiu DeyiXiong - 5982–5986 + 5983–5987 The neural seq2seq based question generation (QG) is prone to generating generic and undiversified questions that are poorly relevant to the given passage and target answer. In this paper, we propose two methods to address the issue. (1) By a partial copy mechanism, we prioritize words that are morphologically close to words in the input passage when generating questions; (2) By a QA-based reranker, from the n-best list of question candidates, we select questions that are preferred by both the QA and QG model. Experiments and analyses demonstrate that the proposed two methods substantially improve the relevance of generated questions to passages and answers. D19-1614 10.18653/v1/D19-1614 @@ -7159,7 +7158,7 @@ Bodhisattwa PrasadMajumder JulianMcAuley GarrisonCottrell - 5987–5992 + 5988–5993 Stories generated with neural language models have shown promise in grammatical and stylistic consistency. However, the generated stories are still lacking in common sense reasoning, e.g., they often contain sentences deprived of world knowledge. We propose a simple multi-task learning scheme to achieve quantitatively better common sense reasoning in language models by leveraging auxiliary training signals from datasets designed to provide common sense grounding. When combined with our two-stage fine-tuning pipeline, our method achieves improved common sense reasoning and state-of-the-art perplexity on the WritingPrompts (Fan et al., 2018) story generation dataset. D19-1615 D19-1615.Attachment.pdf @@ -7169,7 +7168,7 @@ Abstract Text Summarization: A Low Resource Challenge ShantipriyaParida PetrMotlicek - 5993–5997 + 5994–5998 Text summarization is considered as a challenging task in the NLP community. The availability of datasets for the task of multilingual text summarization is rare, and such datasets are difficult to construct. In this work, we build an abstract text summarizer for the German language text using the state-of-the-art “Transformer” model. We propose an iterative data augmentation approach which uses synthetic data along with the real summarization data for the German language. To generate synthetic data, the Common Crawl (German) dataset is exploited, which covers different domains. The synthetic data is effective for the low resource condition and is particularly helpful for our multilingual scenario where availability of summarizing data is still a challenging issue. The data are also useful in deep learning scenarios where the neural models require a large amount of training data for utilization of its capacity. The obtained summarization performance is measured in terms of ROUGE and BLEU score. We achieve an absolute improvement of +1.5 and +16.0 in ROUGE1 F1 (R1_F1) on the development and test sets, respectively, compared to the system which does not rely on data augmentation. D19-1616 10.18653/v1/D19-1616 @@ -7178,7 +7177,7 @@ Generating Modern Poetry Automatically in <fixed-case>F</fixed-case>innish MikaHämäläinen KhalidAlnajjar - 5998–6003 + 5999–6004 We present a novel approach for generating poetry automatically for the morphologically rich Finnish language by using a genetic algorithm. The approach improves the state of the art of the previous Finnish poem generators by introducing a higher degree of freedom in terms of structural creativity. Our approach is evaluated and described within the paradigm of computational creativity, where the fitness functions of the genetic algorithm are assimilated with the notion of aesthetics. The output is considered to be a poem 81.5% of the time by human evaluators. D19-1617 10.18653/v1/D19-1617 @@ -7189,7 +7188,7 @@ ProdromosMalakasiotis MariannaApidianaki IonAndroutsopoulos - 6004–6010 + 6005–6011 We propose SUM-QE, a novel Quality Estimation model for summarization based on BERT. The model addresses linguistic quality aspects that are only indirectly captured by content-based approaches to summary evaluation, without involving comparison with human references. SUM-QE achieves very high correlations with human ratings, outperforming simpler models addressing these linguistic aspects. Predictions of the SUM-QE model can be used for system development, and to inform users of the quality of automatically produced summaries and other types of generated text. D19-1618 D19-1618.Attachment.zip @@ -7199,7 +7198,7 @@ An Empirical Comparison on Imitation Learning and Reinforcement Learning for Paraphrase Generation WanyuDu YangfengJi - 6011–6017 + 6012–6018 Generating paraphrases from given sentences involves decoding words step by step from a large vocabulary. To learn a decoder, supervised learning which maximizes the likelihood of tokens always suffers from the exposure bias. Although both reinforcement learning (RL) and imitation learning (IL) have been widely used to alleviate the bias, the lack of direct comparison leads to only a partial image on their benefits. In this work, we present an empirical study on how RL and IL can help boost the performance of generating paraphrases, with the pointer-generator as a base model. Experiments on the benchmark datasets show that (1) imitation learning is constantly better than reinforcement learning; and (2) the pointer-generator models with imitation learning outperform the state-of-the-art methods with a large margin. D19-1619 D19-1619.Attachment.zip @@ -7211,7 +7210,7 @@ YueDong Jackie Chi KitCheung AnnieLouis - 6018–6023 + 6019–6024 Sentence position is a strong feature for news summarization, since the lead often (but not always) summarizes the key points of the article. In this paper, we show that recent neural systems excessively exploit this trend, which although powerful for many inputs, is also detrimental when summarizing documents where important content should be extracted from later parts of the article. We propose two techniques to make systems sensitive to the importance of content in different parts of the article. The first technique employs ‘unbiased’ data; i.e., randomly shuffled sentences of the source document, to pretrain the model. The second technique uses an auxiliary ROUGE-based loss that encourages the model to distribute importance scores throughout a document by mimicking sentence-level ROUGE scores on the training data. We show that these techniques significantly improve the performance of a competitive reinforcement learning based extractive system, with the auxiliary loss being more powerful than pretraining. D19-1620 D19-1620.Attachment.zip @@ -7223,7 +7222,7 @@ Sanket VaibhavMehta JaimeCarbonell TaylorBerg-Kirkpatrick - 6024–6030 + 6025–6031 Existing recurrent neural language models often fail to capture higher-level structure present in text: for example, rhyming patterns present in poetry. Much prior work on poetry generation uses manually defined constraints which are satisfied during decoding using either specialized decoding procedures or rejection sampling. The rhyming constraints themselves are typically not learned by the generator. We propose an alternate approach that uses a structured discriminator to learn a poetry generator that directly captures rhyming constraints in a generative adversarial setup. By causing the discriminator to compare poems based only on a learned similarity matrix of pairs of line ending words, the proposed approach is able to successfully learn rhyming patterns in two different English poetry datasets (Sonnet and Limerick) without explicitly being provided with any phonetic information D19-1621 10.18653/v1/D19-1621 @@ -7233,7 +7232,7 @@ WenjieZhou MinghuaZhang YunfangWu - 6031–6036 + 6032–6037 Question generation is a challenging task which aims to ask a question based on an answer and relevant context. The existing works suffer from the mismatching between question type and answer, i.e. generating a question with type how while the answer is a personal name. We propose to automatically predict the question type based on the input answer and context. Then, the question type is fused into a seq2seq model to guide the question generation, so as to deal with the mismatching problem. We achieve significant improvement on the accuracy of question type prediction and finally obtain state-of-the-art results for question generation on both SQuAD and MARCO datasets. D19-1622 10.18653/v1/D19-1622 @@ -7244,7 +7243,7 @@ DerenLei PengdaQin William YangWang - 6037–6043 + 6038–6044 Deep reinforcement learning (RL) has been a commonly-used strategy for the abstractive summarization task to address both the exposure bias and non-differentiable task issues. However, the conventional reward Rouge-L simply looks for exact n-grams matches between candidates and annotated references, which inevitably makes the generated sentences repetitive and incoherent. In this paper, instead of Rouge-L, we explore the practicability of utilizing the distributional semantics to measure the matching degrees. With distributional semantics, sentence-level evaluation can be obtained, and semantically-correct phrases can also be generated without being limited to the surface form of the reference sentences. Human judgments on Gigaword and CNN/Daily Mail datasets show that our proposed distributional semantics reward (DSR) has distinct superiority in capturing the lexical and compositional diversity of natural language. D19-1623 10.18653/v1/D19-1623 @@ -7252,7 +7251,7 @@ Clause-Wise and Recursive Decoding for Complex and Cross-Domain Text-to-<fixed-case>SQL</fixed-case> Generation DongjunLee - 6044–6050 + 6045–6051 Most deep learning approaches for text-to-SQL generation are limited to the WikiSQL dataset, which only supports very simple queries over a single table. We focus on the Spider dataset, a complex and cross-domain text-to-SQL task, which includes complex queries over multiple tables. In this paper, we propose a SQL clause-wise decoding neural architecture with a self-attention based database schema encoder to address the Spider task. Each of the clause-specific decoders consists of a set of sub-modules, which is defined by the syntax of each clause. Additionally, our model works recursively to support nested queries. When evaluated on the Spider dataset, our approach achieves 4.6% and 9.8% accuracy gain in the test and dev sets, respectively. In addition, we show that our model is significantly more effective at predicting complex and nested queries than previous work. D19-1624 10.18653/v1/D19-1624 @@ -7263,7 +7262,7 @@ JonathanGordon NanyunPeng JonathanMay - 6051–6057 + 6052–6058 How do adjectives project from a noun to its parts? If a motorcycle is red, are its wheels red? Is a nuclear submarine’s captain nuclear? These questions are easy for humans to judge using our commonsense understanding of the world, but are difficult for computers. To attack this challenge, we crowdsource a set of human judgments that answer the English-language question “Given a whole described by an adjective, does the adjective also describe a given part?” We build strong baselines for this task with a classification approach. Our findings indicate that, despite the recent successes of large language models on tasks aimed to assess commonsense knowledge, these models do not greatly outperform simple word-level models based on pre-trained word embeddings. This provides evidence that the amount of commonsense knowledge encoded in these language models does not extend far beyond that already baked into the word embeddings. Our dataset will serve as a useful testbed for future research in commonsense reasoning, especially as it relates to adjectives and objects D19-1625 D19-1625.Attachment.pdf @@ -7276,7 +7275,7 @@ WeizhenQi NanDuan XiaolaLin - 6058–6062 + 6059–6063 In this work, we propose an aggregation method to combine the Bidirectional Encoder Representations from Transformer (BERT) with a MatchLSTM layer for Sequence Matching. Given a sentence pair, we extract the output representations of it from BERT. Then we extend BERT with a MatchLSTM layer to get further interaction of the sentence pair for sequence matching tasks. Taking natural language inference as an example, we split BERT output into two parts, which is from premise sentence and hypothesis sentence. At each position of the hypothesis sentence, both the weighted representation of the premise sentence and the representation of the current token are fed into LSTM. We jointly train the aggregation layer and pre-trained layer for sequence matching. We conduct an experiment on two publicly available datasets, WikiQA and SNLI. Experiments show that our model achieves significantly improvement compared with state-of-the-art methods on both datasets. D19-1626 10.18653/v1/D19-1626 @@ -7285,7 +7284,7 @@ What Does This Word Mean? Explaining Contextualized Embeddings with Natural Language Definition Ting-YunChang Yun-NungChen - 6063–6069 + 6064–6070 Contextualized word embeddings have boosted many NLP tasks compared with traditional static word embeddings. However, the word with a specific sense may have different contextualized embeddings due to its various contexts. To further investigate what contextualized word embeddings capture, this paper analyzes whether they can indicate the corresponding sense definitions and proposes a general framework that is capable of explaining word meanings given contextualized word embeddings for better interpretation. The experiments show that both ELMo and BERT embeddings can be well interpreted via a readable textual form, and the findings may benefit the research community for a better understanding of what the embeddings capture. D19-1627 D19-1627.Attachment.zip @@ -7299,7 +7298,7 @@ TengfeiMa VinayReddy RishiArora - 6070–6074 + 6071–6075 Pre-trained BERT contextualized representations have achieved state-of-the-art results on multiple downstream NLP tasks by fine-tuning with task-specific data. While there has been a lot of focus on task-specific fine-tuning, there has been limited work on improving the pre-trained representations. In this paper, we explore ways of improving the pre-trained contextual representations for the task of automatic short answer grading, a critical component of intelligent tutoring systems. We show that the pre-trained BERT model can be improved by augmenting data from the domain-specific resources like textbooks. We also present a new approach to use labeled short answering grading data for further enhancement of the language model. Empirical evaluation on multi-domain datasets shows that task-specific fine-tuning on the enhanced pre-trained language model achieves superior performance for short answer grading. D19-1628 10.18653/v1/D19-1628 @@ -7311,7 +7310,7 @@ KeisukeSakaguchi PeterClark AntoineBosselut - 6075–6084 + 6076–6085 We introduce WIQA, the first large-scale dataset of “What if...” questions over procedural text. WIQA contains a collection of paragraphs, each annotated with multiple influence graphs describing how one change affects another, and a large (40k) collection of “What if...?” multiple-choice questions derived from these. For example, given a paragraph about beach erosion, would stormy weather hasten or decelerate erosion? WIQA contains three kinds of questions: perturbations to steps mentioned in the paragraph; external (out-of-paragraph) perturbations requiring commonsense knowledge; and irrelevant (no effect) perturbations. We find that state-of-the-art models achieve 73.8% accuracy, well below the human performance of 96.3%. We analyze the challenges, in particular tracking chains of influences, and present the dataset as an open challenge to the community. D19-1629 D19-1629.Attachment.zip @@ -7321,7 +7320,7 @@ Evaluating <fixed-case>BERT</fixed-case> for natural language inference: A case study on the <fixed-case>C</fixed-case>ommitment<fixed-case>B</fixed-case>ank NanjiangJiang Marie-Catherinede Marneffe - 6085–6090 + 6086–6091 Natural language inference (NLI) datasets (e.g., MultiNLI) were collected by soliciting hypotheses for a given premise from annotators. Such data collection led to annotation artifacts: systems can identify the premise-hypothesis relationship without observing the premise (e.g., negation in hypothesis being indicative of contradiction). We address this problem by recasting the CommitmentBank for NLI, which contains items involving reasoning over the extent to which a speaker is committed to complements of clause-embedding verbs under entailment-canceling environments (conditional, negation, modal and question). Instead of being constructed to stand in certain relationships with the premise, hypotheses in the recast CommitmentBank are the complements of the clause-embedding verb in each premise, leading to no annotation artifacts in the hypothesis. A state-of-the-art BERT-based model performs well on the CommitmentBank with 85% F1. However analysis of model behavior shows that the BERT models still do not capture the full complexity of pragmatic reasoning, nor encode some of the linguistic generalizations, highlighting room for improvement. D19-1630 10.18653/v1/D19-1630 @@ -7334,7 +7333,7 @@ SantoshTokala NiloyGanguly PawanGoyal - 6091–6096 + 6092–6097 Recently, biomedical version of embeddings obtained from language models such as BioELMo have shown state-of-the-art results for the textual inference task in the medical domain. In this paper, we explore how to incorporate structured domain knowledge, available in the form of a knowledge graph (UMLS), for the Medical NLI task. Specifically, we experiment with fusing embeddings obtained from knowledge graph with the state-of-the-art approaches for NLI task (ESIM model). We also experiment with fusing the domain-specific sentiment information for the task. Experiments conducted on MedNLI dataset clearly show that this strategy improves the baseline BioELMo architecture for the Medical NLI task. D19-1631 D19-1631.Attachment.pdf @@ -7350,7 +7349,7 @@ PhilippKoehn VishravChaudhary Marc’AurelioRanzato - 6097–6110 + 6098–6111 For machine translation, a vast majority of language pairs in the world are considered low-resource because they have little parallel data available. Besides the technical challenges of learning with limited supervision, it is difficult to evaluate methods trained on low-resource language pairs because of the lack of freely and publicly available benchmarks. In this work, we introduce the FLORES evaluation datasets for Nepali–English and Sinhala– English, based on sentences translated from Wikipedia. Compared to English, these are languages with very different morphology and syntax, for which little out-of-domain parallel data is available and for which relatively large amounts of monolingual data are freely available. We describe our process to collect and cross-check the quality of translations, and we report baseline performance using several learning settings: fully supervised, weakly supervised, semi-supervised, and fully unsupervised. Our experiments demonstrate that current state-of-the-art methods perform rather poorly on this benchmark, posing a challenge to the research community working on low-resource MT. Data and code to reproduce our experiments are available at https://github.com/facebookresearch/flores. D19-1632 D19-1632.Attachment.zip @@ -7362,7 +7361,7 @@ OmerLevy YinhanLiu LukeZettlemoyer - 6111–6120 + 6112–6121 Most machine translation systems generate text autoregressively from left to right. We, instead, use a masked language modeling objective to train a model to predict any subset of the target words, conditioned on both the input text and a partially masked target translation. This approach allows for efficient iterative decoding, where we first predict all of the target words non-autoregressively, and then repeatedly mask out and regenerate the subset of words that the model is least confident about. By applying this strategy for a constant number of iterations, our model improves state-of-the-art performance levels for non-autoregressive and parallel decoding translation models by over 4 BLEU on average. It is also able to reach within about 1 BLEU point of a typical left-to-right transformer model, while decoding significantly faster. D19-1633 10.18653/v1/D19-1633 @@ -7374,7 +7373,7 @@ HuanboLuan JingfangXu MaosongSun - 6121–6131 + 6122–6132 Automatic post-editing (APE), which aims to correct errors in the output of machine translation systems in a post-processing step, is an important task in natural language processing. While recent work has achieved considerable performance gains by using neural networks, how to model the copying mechanism for APE remains a challenge. In this work, we propose a new method for modeling copying for APE. To better identify translation errors, our method learns the representations of source sentences and system outputs in an interactive way. These representations are used to explicitly indicate which words in the system outputs should be copied. Finally, CopyNet (Gu et.al., 2016) can be combined with our method to place the copied words in correct positions in post-edited translations. Experiments on the datasets of the WMT 2016-2017 APE shared tasks show that our approach outperforms all best published results. D19-1634 10.18653/v1/D19-1634 @@ -7384,7 +7383,7 @@ YupeiDu YuanbinWu ManLan - 6132–6142 + 6133–6143 Word embeddings have been widely used to study gender stereotypes in texts. One key problem regarding existing bias scores is to evaluate their validities: do they really reflect true bias levels? For a small set of words (e.g. occupations), we can rely on human annotations or external data. However, for most words, evaluating the correctness of them is still an open problem. In this work, we utilize word association test, which contains rich types of word connections annotated by human participants, to explore how gender stereotypes spread within our minds. Specifically, we use random walk on word association graph to derive bias scores for a large amount of words. Experiments show that these bias scores correlate well with bias in the real world. More importantly, comparing with word-embedding-based bias scores, it provides a different perspective on gender stereotypes in words. D19-1635 10.18653/v1/D19-1635 @@ -7394,7 +7393,7 @@ AbhijitMishra TarunTater KarthikSankaranarayanan - 6143–6153 + 6144–6154 In this paper, we propose a novel framework for sarcasm generation; the system takes a literal negative opinion as input and translates it into a sarcastic version. Our framework does not require any paired data for training. Sarcasm emanates from context-incongruity which becomes apparent as the sentence unfolds. Our framework introduces incongruity into the literal input version through modules that: (a) filter factual content from the input opinion, (b) retrieve incongruous phrases related to the filtered facts and (c) synthesize sarcastic text from the incongruous filtered and incongruous phrases. The framework employs reinforced neural sequence to sequence learning and information retrieval and is trained only using unlabeled non-sarcastic and sarcastic opinions. Since no labeled dataset exists for such a task, for evaluation, we manually prepare a benchmark dataset containing literal opinions and their sarcastic paraphrases. Qualitative and quantitative performance analyses on the data reveal our system’s superiority over baselines built using known unsupervised statistical and neural machine translation and style transfer techniques. D19-1636 10.18653/v1/D19-1636 @@ -7408,7 +7407,7 @@ WeijiangFeng Elena Suet-YingChiu HongYu - 6154–6163 + 6155–6164 Classical Chinese poetry is a jewel in the treasure house of Chinese culture. Previous poem generation models only allow users to employ keywords to interfere the meaning of generated poems, leaving the dominion of generation to the model. In this paper, we propose a novel task of generating classical Chinese poems from vernacular, which allows users to have more control over the semantic of generated poems. We adapt the approach of unsupervised machine translation (UMT) to our task. We use segmentation-based padding and reinforcement learning to address under-translation and over-translation respectively. According to experiments, our approach significantly improve the perplexity and BLEU compared with typical UMT models. Furthermore, we explored guidelines on how to write the input vernacular to generate better poems. Human evaluation showed our approach can generate high-quality poems which are comparable to amateur poems. D19-1637 D19-1637.Attachment.pdf @@ -7418,7 +7417,7 @@ Set to Ordered Text: Generating Discharge Instructions from Medical Billing Codes LittonJ Kurisinkel NancyChen - 6164–6174 + 6165–6175 We present set to ordered text, a natural language generation task applied to automatically generating discharge instructions from admission ICD (International Classification of Diseases) codes. This task differs from other natural language generation tasks in the following ways: (1) The input is a set of identifiable entities (ICD codes) where the relations between individual entity are not explicitly specified. (2) The output text is not a narrative description (e.g. news articles) composed from the input. Rather, inferences are made from the input (symptoms specified in ICD codes) to generate the output (instructions). (3) There is an optimal order in which each sentence (instruction) should appear in the output. Unlike most other tasks, neither the input (ICD codes) nor their corresponding symptoms appear in the output, so the ordering of the output instructions needs to be learned in an unsupervised fashion. Based on clinical intuition, we hypothesize that each instruction in the output is mapped to a subset of ICD codes specified in the input. We propose a neural architecture that jointly models (a) subset selection: choosing relevant subsets from a set of input entities; (b) content ordering: learning the order of instructions; and (c) text generation: representing the instructions corresponding to the selected subsets in natural language. In addition, we penalize redundancy during beam search to improve tractability for long text generation. Our model outperforms baseline models in BLEU scores and human evaluation. We plan to extend this work to other tasks such as recipe generation from ingredients. D19-1638 10.18653/v1/D19-1638 @@ -7431,7 +7430,7 @@ EricMeinhardt EricBakovic LeonBergen - 6175–6185 + 6176–6186 Phonological processes are context-dependent sound changes in natural languages. We present an unsupervised approach to learning human-readable descriptions of phonological processes from collections of related utterances. Our approach builds upon a technique from the programming languages community called *constraint-based program synthesis*. We contribute a novel encoding of the learning problem into Boolean Satisfiability constraints, which enables both data efficiency and fast inference. We evaluate our system on textbook phonology problems and datasets from the literature, and show that it achieves high accuracy at interactive speeds. D19-1639 10.18653/v1/D19-1639 @@ -7446,7 +7445,7 @@ QiongZhang LuoSi XiaozhongLiu - 6186–6195 + 6187–6196 The task of Chinese text spam detection is very challenging due to both glyph and phonetic variations of Chinese characters. This paper proposes a novel framework to jointly model Chinese variational, semantic, and contextualized representations for Chinese text spam detection task. In particular, a Variation Family-enhanced Graph Embedding (VFGE) algorithm is designed based on a Chinese character variation graph. The VFGE can learn both the graph embeddings of the Chinese characters (local) and the latent variation families (global). Furthermore, an enhanced bidirectional language model, with a combination gate function and an aggregation learning function, is proposed to integrate the graph and text information while capturing the sequential information. Extensive experiments have been conducted on both SMS and review datasets, to show the proposed method outperforms a series of state-of-the-art models for Chinese spam detection. D19-1640 10.18653/v1/D19-1640 @@ -7455,7 +7454,7 @@ An Attentive Fine-Grained Entity Typing Model with Latent Type Representation YingLin HengJi - 6196–6201 + 6197–6202 We propose a fine-grained entity typing model with a novel attention mechanism and a hybrid type classifier. We advance existing methods in two aspects: feature extraction and type prediction. To capture richer contextual information, we adopt contextualized word representations instead of fixed word embeddings used in previous work. In addition, we propose a two-step mention-aware attention mechanism to enable the model to focus on important words in mentions and contexts. We also present a hybrid classification method beyond binary relevance to exploit type inter-dependency with latent type representation. Instead of independently predicting each type, we predict a low-dimensional vector that encodes latent type features and reconstruct the type vector from this latent representation. Experiment results on multiple data sets show that our model significantly advances the state-of-the-art on fine-grained entity typing, obtaining up to 6.1% and 5.5% absolute gains in macro averaged F-score and micro averaged F-score respectively. D19-1641 10.18653/v1/D19-1641 @@ -7465,7 +7464,7 @@ QiangNing SanjaySubramanian DanRoth - 6202–6208 + 6203–6209 Determining temporal relations (e.g., before or after) between events has been a challenging natural language understanding task, partly due to the difficulty to generate large amounts of high-quality training data. Consequently, neural approaches have not been widely used on it, or showed only moderate improvements. This paper proposes a new neural system that achieves about 10% absolute improvement in accuracy over the previous best system (25% error reduction) on two benchmark datasets. The proposed system is trained on the state-of-the-art MATRES dataset and applies contextualized word embeddings, a Siamese encoder of a temporal common sense knowledge base, and global inference via integer linear programming (ILP). We suggest that the new approach could serve as a strong baseline for future research in this area. D19-1642 D19-1642.Attachment.zip @@ -7477,7 +7476,7 @@ DonghongDu XinLi YangqiuSong - 6209–6214 + 6210–6215 Fine-grained entity typing is a challenging problem since it usually involves a relatively large tag set and may require to understand the context of the entity mention. In this paper, we use entity linking to help with the fine-grained entity type classification process. We propose a deep neural model that makes predictions based on both the context and the information obtained from entity linking results. Experimental results on two commonly used datasets demonstrates the effectiveness of our approach. On both datasets, it achieves more than 5% absolute strict accuracy improvement over the state of the art. D19-1643 10.18653/v1/D19-1643 @@ -7486,7 +7485,7 @@ Combining Spans into Entities: A Neural Two-Stage Approach for Recognizing Discontiguous Entities BailinWang WeiLu - 6215–6223 + 6216–6224 In medical documents, it is possible that an entity of interest not only contains a discontiguous sequence of words but also overlaps with another entity. Entities of such structures are intrinsically hard to recognize due to the large space of possible entity combinations. In this work, we propose a neural two-stage approach to recognizing discontiguous and overlapping entities by decomposing this problem into two subtasks: 1) it first detects all the overlapping spans that either form entities on their own or present as segments of discontiguous entities, based on the representation of segmental hypergraph, 2) next it learns to combine these segments into discontiguous entities with a classifier, which filters out other incorrect combinations of segments. Two neural components are designed for these subtasks respectively and they are learned jointly using a shared encoder for text. Our model achieves the state-of-the-art performance in a standard dataset, even in the absence of external features that previous methods used. D19-1644 10.18653/v1/D19-1644 @@ -7497,7 +7496,7 @@ TakuyaHiraoka KunihikoSadamasa MathiasNiepert - 6224–6230 + 6225–6231 Most existing relation extraction approaches exclusively target binary relations, and n-ary relation extraction is relatively unexplored. Current state-of-the-art n-ary relation extraction method is based on a supervised learning approach and, therefore, may suffer from the lack of sufficient relation labels. In this paper, we propose a novel approach to cross-sentence n-ary relation extraction based on universal schemas. To alleviate the sparsity problem and to leverage inherent decomposability of n-ary relations, we propose to learn relation representations of lower-arity facts that result from decomposing higher-arity facts. The proposed method computes a score of a new n-ary fact by aggregating scores of its decomposed lower-arity facts. We conduct experiments with datasets for ternary relation extraction and empirically show that our method improves the n-ary relation extraction performance compared to previous methods. D19-1645 D19-1645.Attachment.zip @@ -7511,7 +7510,7 @@ LeSun BinDong ShanshanJiang - 6231–6236 + 6232–6237 Current region-based NER models only rely on fully-annotated training data to learn effective region encoder, which often face the training data bottleneck. To alleviate this problem, this paper proposes Gazetteer-Enhanced Attentive Neural Networks, which can enhance region-based NER by learning name knowledge of entity mentions from easily-obtainable gazetteers, rather than only from fully-annotated data. Specially, we first propose an attentive neural network (ANN), which explicitly models the mention-context association and therefore is convenient for integrating externally-learned knowledge. Then we design an auxiliary gazetteer network, which can effectively encode name regularity of mentions only using gazetteers. Finally, the learned gazetteer network is incorporated into ANN for better NER. Experiments show that our ANN can achieve the state-of-the-art performance on ACE2005 named entity recognition benchmark. Besides, incorporating gazetteer network can further improve the performance and significantly reduce the requirement of training data. D19-1646 10.18653/v1/D19-1646 @@ -7522,7 +7521,7 @@ RobertJäschke FrankFischer JannikStrötgen - 6237–6242 + 6238–6243 Attributing a particular property to a person by naming another person, who is typically wellknown for the respective property, is called a Vossian Antonomasia (VA). This subtpye of metonymy, which overlaps with metaphor, has a specific syntax and is especially frequent in journalistic texts. While identifying Vossian Antonomasia is of particular interest in the study of stylistics, it is also a source of errors in relation and fact extraction as an explicitly mentioned entity occurs only metaphorically and should not be associated with respective contexts. Despite rather simple syntactic variations, the automatic extraction of VA was never addressed as yet since it requires a deeper semantic understanding of mentioned entities and underlying relations. In this paper, we propose a first method for the extraction of VAs that works completely automatically. Our approaches use named entity recognition, distant supervision based on Wikidata, and a bi-directional LSTM for postprocessing. The evaluation on 1.8 million articles of the New York Times corpus shows that our approach significantly outperforms the only existing semi-automatic approach for VA identification by more than 30 percentage points in precision. D19-1647 10.18653/v1/D19-1647 @@ -7534,7 +7533,7 @@ TakashiNinomiya TakuyaMakino TomoyaIwakura - 6243–6248 + 6244–6249 We propose a method to improve named entity recognition (NER) for chemical compounds using multi-task learning by jointly training a chemical NER model and a chemical com- pound paraphrase model. Our method en- ables the long short-term memory (LSTM) of the NER model to capture chemical com- pound paraphrases by sharing the parameters of the LSTM and character embeddings be- tween the two models. The experimental re- sults on the BioCreative IV’s CHEMDNER task show that our method improves chemi- cal NER and achieves state-of-the-art perfor- mance. D19-1648 10.18653/v1/D19-1648 @@ -7548,7 +7547,7 @@ PengLi MaosongSun JieZhou - 6249–6254 + 6250–6255 We present FewRel 2.0, a more challenging task to investigate two aspects of few-shot relation classification models: (1) Can they adapt to a new domain with only a handful of instances? (2) Can they detect none-of-the-above (NOTA) relations? To construct FewRel 2.0, we build upon the FewRel dataset by adding a new test set in a quite different domain, and a NOTA relation choice. With the new dataset and extensive experimental analysis, we found (1) that the state-of-the-art few-shot relation classification models struggle on these two aspects, and (2) that the commonly-used techniques for domain adaptation and NOTA detection still cannot handle the two challenges well. Our research calls for more attention and further efforts to these two real-world issues. All details and resources about the dataset and baselines are released at https://github.com/thunlp/fewrel. D19-1649 10.18653/v1/D19-1649 @@ -7558,7 +7557,7 @@ StephenMayhew TatianaTsygankova DanRoth - 6255–6260 + 6256–6261 For those languages which use it, capitalization is an important signal for the fundamental NLP tasks of Named Entity Recognition (NER) and Part of Speech (POS) tagging. In fact, it is such a strong signal that model performance on these tasks drops sharply in common lowercased scenarios, such as noisy web text or machine translation outputs. In this work, we perform a systematic analysis of solutions to this problem, modifying only the casing of the train or test data using lowercasing and truecasing methods. While prior work and first impressions might suggest training a caseless model, or using a truecaser at test time, we show that the most effective strategy is a concatenation of cased and lowercased training data, producing a single model with high performance on both cased and uncased text. As shown in our experiments, this result holds across tasks and input representations. Finally, we show that our proposed solution gives an 8% F1 improvement in mention detection on noisy out-of-domain Twitter data. D19-1650 10.18653/v1/D19-1650 @@ -7568,7 +7567,7 @@ SangnieBhardwaj SamarthAggarwal MausamMausam - 6261–6266 + 6262–6267 Open Information Extraction (Open IE) systems have been traditionally evaluated via manual annotation. Recently, an automated evaluator with a benchmark dataset (OIE2016) was released – it scores Open IE systems automatically by matching system predictions with predictions in the benchmark dataset. Unfortunately, our analysis reveals that its data is rather noisy, and the tuple matching in the evaluator has issues, making the results of automated comparisons less trustworthy. We contribute CaRB, an improved dataset and framework for testing Open IE systems. To the best of our knowledge, CaRB is the first crowdsourced Open IE dataset and it also makes substantive changes in the matching code and metrics. NLP experts annotate CaRB’s dataset to be more accurate than OIE2016. Moreover, we find that on one pair of Open IE systems, CaRB framework provides contradictory results to OIE2016. Human assessment verifies that CaRB’s ranking of the two systems is the accurate ranking. We release the CaRB framework along with its crowdsourced dataset. D19-1651 D19-1651.Attachment.zip @@ -7578,7 +7577,7 @@ Weakly Supervised Attention Networks for Entity Recognition BarunPatra Joel Ruben AntonyMoniz - 6267–6272 + 6268–6273 The task of entity recognition has traditionally been modelled as a sequence labelling task. However, this usually requires a large amount of fine-grained data annotated at the token level, which in turn can be expensive and cumbersome to obtain. In this work, we aim to circumvent this requirement of word-level annotated data. To achieve this, we propose a novel architecture for entity recognition from a corpus containing weak binary presence/absence labels, which are relatively easier to obtain. We show that our proposed weakly supervised model, trained solely on a multi-label classification task, performs reasonably well on the task of entity recognition, despite not having access to any token-level ground truth data. D19-1652 D19-1652.Attachment.pdf @@ -7589,7 +7588,7 @@ GakuMorio RyoEgawa KatsuhideFujita - 6273–6278 + 6274–6279 In online arguments, identifying how users construct their arguments to persuade others is important in order to understand a persuasive strategy directly. However, existing research lacks empirical investigations on highly semantic aspects of elementary units (EUs), such as propositions for a persuasive online argument. Therefore, this paper focuses on a pilot study, revealing a persuasion strategy using EUs. Our contributions are as follows: (1) annotating five types of EUs in a persuasive forum, the so-called ChangeMyView, (2) revealing both intuitive and non-intuitive strategic insights for the persuasion by analyzing 4612 annotated EUs, and (3) proposing baseline neural models that identify the EU boundary and type. Our observations imply that EUs definitively characterize online persuasion strategies. D19-1653 10.18653/v1/D19-1653 @@ -7601,7 +7600,7 @@ RuifengXu XiangAo MinYang - 6279–6284 + 6280–6285 Aspect-based sentiment analysis (ABSA) has attracted increasing attention recently due to its broad applications. In existing ABSA datasets, most sentences contain only one aspect or multiple aspects with the same sentiment polarity, which makes ABSA task degenerate to sentence-level sentiment analysis. In this paper, we present a new large-scale Multi-Aspect Multi-Sentiment (MAMS) dataset, in which each sentence contains at least two different aspects with different sentiment polarities. The release of this dataset would push forward the research in this field. In addition, we propose simple yet effective CapsNet and CapsNet-BERT models which combine the strengths of recent NLP advances. Experiments on our new dataset show that the proposed model significantly outperforms the state-of-the-art baseline methods D19-1654 10.18653/v1/D19-1654 @@ -7613,7 +7612,7 @@ ChaozhuoLi YanYang TianruiLi - 6285–6291 + 6286–6292 Deep neural networks (DNNs) can fit (or even over-fit) the training data very well. If a DNN model is trained using data with noisy labels and tested on data with clean labels, the model may perform poorly. This paper studies the problem of learning with noisy labels for sentence-level sentiment classification. We propose a novel DNN model called NetAb (as shorthand for convolutional neural Networks with Ab-networks) to handle noisy labels during training. NetAb consists of two convolutional neural networks, one with a noise transition layer for dealing with the input noisy labels and the other for predicting ‘clean’ labels. We train the two networks using their respective loss functions in a mutual reinforcement manner. Experimental results demonstrate the effectiveness of the proposed model. D19-1655 10.18653/v1/D19-1655 @@ -7623,7 +7622,7 @@ ChenLiu MuhammadOsama AndersonDe Andrade - 6292–6297 + 6293–6298 We introduce a new dataset for multi-class emotion analysis from long-form narratives in English. The Dataset for Emotions of Narrative Sequences (DENS) was collected from both classic literature available on Project Gutenberg and modern online narratives avail- able on Wattpad, annotated using Amazon Mechanical Turk. A number of statistics and baseline benchmarks are provided for the dataset. Of the tested techniques, we find that the fine-tuning of a pre-trained BERT model achieves the best results, with an average micro-F1 score of 60.4%. Our results show that the dataset provides a novel opportunity in emotion analysis that requires moving beyond existing sentence-level techniques. D19-1656 D19-1656.Attachment.zip @@ -7633,7 +7632,7 @@ Multi-Task Stance Detection with Sentiment and Stance Lexicons YingjieLi CorneliaCaragea - 6298–6304 + 6299–6305 Stance detection aims to detect whether the opinion holder is in support of or against a given target. Recent works show improvements in stance detection by using either the attention mechanism or sentiment information. In this paper, we propose a multi-task framework that incorporates target-specific attention mechanism and at the same time takes sentiment classification as an auxiliary task. Moreover, we used a sentiment lexicon and constructed a stance lexicon to provide guidance for the attention layer. Experimental results show that the proposed model significantly outperforms state-of-the-art deep learning methods on the SemEval-2016 dataset. D19-1657 10.18653/v1/D19-1657 @@ -7642,7 +7641,7 @@ A Robust Self-Learning Framework for Cross-Lingual Text Classification XinDong Gerardde Melo - 6305–6309 + 6306–6310 Based on massive amounts of data, recent pretrained contextual representation models have made significant strides in advancing a number of different English NLP tasks. However, for other languages, relevant training data may be lacking, while state-of-the-art deep learning methods are known to be data-hungry. In this paper, we present an elegantly simple robust self-learning framework to include unlabeled non-English samples in the fine-tuning process of pretrained multilingual representation models. We leverage a multilingual model’s own predictions on unlabeled non-English data in order to obtain additional information that can be used during further fine-tuning. Compared with original multilingual models and other cross-lingual classification models, we observe significant gains in effectiveness on document and sentiment classification for a range of diverse languages. D19-1658 10.18653/v1/D19-1658 @@ -7650,7 +7649,7 @@ Learning to Flip the Sentiment of Reviews from Non-Parallel Corpora CanasaiKruengkrai - 6310–6315 + 6311–6316 Flipping sentiment while preserving sentence meaning is challenging because parallel sentences with the same content but different sentiment polarities are not always available for model learning. We introduce a method for acquiring imperfectly aligned sentences from non-parallel corpora and propose a model that learns to minimize the sentiment and content losses in a fully end-to-end manner. Our model is simple and offers well-balanced results across two domains: Yelp restaurant and Amazon product reviews. D19-1659 10.18653/v1/D19-1659 @@ -7662,7 +7661,7 @@ YukaTakei HirokiOkamoto JunGoto - 6316–6321 + 6317–6322 Twitter is used for various applications such as disaster monitoring and news material gathering. In these applications, each Tweet is classified into pre-defined classes. These classes have a semantic relationship with each other and can be classified into a hierarchical structure, which is regarded as important information. Label texts of pre-defined classes themselves also include important clues for classification. Therefore, we propose a method that can consider the hierarchical structure of labels and label texts themselves. We conducted evaluation over the Text REtrieval Conference (TREC) 2018 Incident Streams (IS) track dataset, and we found that our method outperformed the methods of the conference participants. D19-1660 10.18653/v1/D19-1660 @@ -7672,7 +7671,7 @@ MiriamHurtado Bodell MartinArvidsson MånsMagnusson - 6322–6328 + 6323–6329 Word embeddings have demonstrated strong performance on NLP tasks. However, lack of interpretability and the unsupervised nature of word embeddings have limited their use within computational social science and digital humanities. We propose the use of informative priors to create interpretable and domain-informed dimensions for probabilistic word embeddings. Experimental results show that sensible priors can capture latent semantic concepts better than or on-par with the current state of the art, while retaining the simplicity and generalizability of using priors. D19-1661 D19-1661.Attachment.zip @@ -7685,7 +7684,7 @@ YanaiElazar DesmondElliott AndersSøgaard - 6329–6334 + 6330–6335 Elazar and Goldberg (2018) showed that protected attributes can be extracted from the representations of a debiased neural network for mention detection at above-chance levels, by evaluating a diagnostic classifier on a held-out subsample of the data it was trained on. We revisit their experiments and conduct a series of follow-up experiments showing that, in fact, the diagnostic classifier generalizes poorly to both new in-domain samples and new domains, indicating that it relies on correlations specific to their particular data sample. We further show that a diagnostic classifier trained on the biased baseline neural network also does not generalize to new samples. In other words, the biases detected in Elazar and Goldberg (2018) seem restricted to their particular data sample, and would therefore not bias the decisions of the model on new samples, whether in-domain or out-of-domain. In light of this, we discuss better methodologies for detecting bias in our models. D19-1662 D19-1662.Attachment.pdf @@ -7696,7 +7695,7 @@ JasabantaPatro SrijanBansal AnimeshMukherjee - 6335–6341 + 6336–6342 In this paper we propose a deep learning framework for sarcasm target detection in predefined sarcastic texts. Identification of sarcasm targets can help in many core natural language processing tasks such as aspect based sentiment analysis, opinion mining etc. To begin with, we perform an empirical study of the socio-linguistic features and identify those that are statistically significant in indicating sarcasm targets (p-values in the range(0.05,0.001)). Finally, we present a deep-learning framework augmented with socio-linguistic features to detect sarcasm targets in sarcastic book-snippets and tweets.We achieve a huge improvement in the performance in terms of exact match and dice scores compared to the current state-of-the-art baseline. D19-1663 10.18653/v1/D19-1663 @@ -7710,7 +7709,7 @@ Prafulla KumarChoubey RuihongHuang LuWang - 6342–6348 + 6343–6349 The increasing prevalence of political bias in news media calls for greater public awareness of it, as well as robust methods for its detection. While prior work in NLP has primarily focused on the lexical bias captured by linguistic attributes such as word choice and syntax, other types of bias stem from the actual content selected for inclusion in the text. In this work, we investigate the effects of informational bias: factual content that can nevertheless be deployed to sway reader opinion. We first produce a new dataset, BASIL, of 300 news articles annotated with 1,727 bias spans and find evidence that informational bias appears in news articles more frequently than lexical bias. We further study our annotations to observe how informational bias surfaces in news articles by different media outlets. Lastly, a baseline model for informational bias prediction is presented by fine-tuning BERT on our labeled data, indicating the challenges of the task and future directions. D19-1664 D19-1664.Attachment.zip @@ -7720,7 +7719,7 @@ Incorporating Label Dependencies in Multilabel Stance Detection WilliamFerreira AndreasVlachos - 6349–6353 + 6350–6354 Stance detection in social media is a well-studied task in a variety of domains. Nevertheless, previous work has mostly focused on multiclass versions of the problem, where the labels are mutually exclusive, and typically positive, negative or neutral. In this paper, we address versions of the task in which an utterance can have multiple labels, thus corresponding to multilabel classification. We propose a method that explicitly incorporates label dependencies in the training objective and compare it against a variety of baselines, as well as a reduction of multilabel to multiclass learning. In experiments with three datasets, we find that our proposed method improves upon all baselines on two out of three datasets. We also show that the reduction of multilabel to multiclass classification can be very competitive, especially in cases where the output consists of a small number of labels and one can enumerate over all label combinations. D19-1665 D19-1665.Attachment.zip @@ -7734,7 +7733,7 @@ AlvinGrissom II BrendanO’Connor MohitIyyer - 6354–6360 + 6355–6361 Sports broadcasters inject drama into play-by-play commentary by building team and player narratives through subjective analyses and anecdotes. Prior studies based on small datasets and manual coding show that such theatrics evince commentator bias in sports broadcasts. To examine this phenomenon, we assemble FOOTBALL, which contains 1,455 broadcast transcripts from American football games across six decades that are automatically annotated with 250K player mentions and linked with racial metadata. We identify major confounding factors for researchers examining racial bias in FOOTBALL, and perform a computational analysis that supports conclusions from prior social science studies. D19-1666 D19-1666.Attachment.pdf @@ -7747,7 +7746,7 @@ WeiDai ZehuiDai YadongDing - 6361–6366 + 6362–6367 Judgment prediction for legal cases has attracted much research efforts for its practice use, of which the ultimate goal is prison term prediction. While existing work merely predicts the total prison term, in reality a defendant is often charged with multiple crimes. In this paper, we argue that charge-based prison term prediction (CPTP) not only better fits realistic needs, but also makes the total prison term prediction more accurate and interpretable. We collect the first large-scale structured data for CPTP and evaluate several competitive baselines. Based on the observation that fine-grained feature selection is the key to achieving good performance, we propose the Deep Gating Network (DGN) for charge-specific feature selection and aggregation. Experiments show that DGN achieves the state-of-the-art performance. D19-1667 10.18653/v1/D19-1667 @@ -7757,7 +7756,7 @@ YannisAssael TheaSommerschield JonathanPrag - 6367–6374 + 6368–6375 Ancient History relies on disciplines such as Epigraphy, the study of ancient inscribed texts, for evidence of the recorded past. However, these texts, “inscriptions”, are often damaged over the centuries, and illegible parts of the text must be restored by specialists, known as epigraphists. This work presents Pythia, the first ancient text restoration model that recovers missing characters from a damaged text input using deep neural networks. Its architecture is carefully designed to handle long-term context information, and deal efficiently with missing or corrupted character and word representations. To train it, we wrote a non-trivial pipeline to convert PHI, the largest digital corpus of ancient Greek inscriptions, to machine actionable text, which we call PHI-ML. On PHI-ML, Pythia’s predictions achieve a 30.1% character error rate, compared to the 57.3% of human epigraphists. Moreover, in 73.5% of cases the ground-truth sequence was among the Top-20 hypotheses of Pythia, which effectively demonstrates the impact of this assistive method on the field of digital epigraphy, and sets the state-of-the-art in ancient text restoration. D19-1668 D19-1668.Attachment.zip @@ -7769,7 +7768,7 @@ AndrewCattle EvangelosPapalexakis XiaojuanMa - 6375–6380 + 6376–6381 We propose a novel tensor embedding method that can effectively extract lexical features for humor recognition. Specifically, we use word-word co-occurrence to encode the contextual content of documents, and then decompose the tensor to get corresponding vector representations. We show that this simple method can capture features of lexical humor effectively for continuous humor recognition. In particular, we achieve a distance of 0.887 on a global humor ranking task, comparable to the top performing systems from SemEval 2017 Task 6B (Potash et al., 2017) but without the need for any external training corpus. In addition, we further show that this approach is also beneficial for small sample humor recognition tasks through a semi-supervised label propagation procedure, which achieves about 0.7 accuracy on the 16000 One-Liners (Mihalcea and Strapparava, 2005) and Pun of the Day (Yang et al., 2015) humour classification datasets using only 10% of known labels. D19-1669 10.18653/v1/D19-1669 @@ -7778,7 +7777,7 @@ <fixed-case>EDA</fixed-case>: Easy Data Augmentation Techniques for Boosting Performance on Text Classification Tasks JasonWei KaiZou - 6381–6387 + 6382–6388 We present EDA: easy data augmentation techniques for boosting performance on text classification tasks. EDA consists of four simple but powerful operations: synonym replacement, random insertion, random swap, and random deletion. On five text classification tasks, we show that EDA improves performance for both convolutional and recurrent neural networks. EDA demonstrates particularly strong results for smaller datasets; on average, across five datasets, training with EDA while using only 50% of the available training set achieved the same accuracy as normal training with all available data. We also performed extensive ablation studies and suggest parameters for practical use. D19-1670 D19-1670.Attachment.zip @@ -7792,7 +7791,7 @@ TaoQi YongfengHuang XingXie - 6388–6393 + 6389–6394 News recommendation can help users find interested news and alleviate information overload. Precisely modeling news and users is critical for news recommendation, and capturing the contexts of words and news is important to learn news and user representations. In this paper, we propose a neural news recommendation approach with multi-head self-attention (NRMS). The core of our approach is a news encoder and a user encoder. In the news encoder, we use multi-head self-attentions to learn news representations from news titles by modeling the interactions between words. In the user encoder, we learn representations of users from their browsed news and use multi-head self-attention to capture the relatedness between the news. Besides, we apply additive attention to learn more informative news and user representations by selecting important words and news. Experiments on a real-world dataset validate the effectiveness and efficiency of our approach. D19-1671 10.18653/v1/D19-1671 @@ -7802,7 +7801,7 @@ XiaoleiHuang JonathanMay NanyunPeng - 6394–6400 + 6395–6401 Building named entity recognition (NER) models for languages that do not have much training data is a challenging task. While recent work has shown promising results on cross-lingual transfer from high-resource languages, it is unclear what knowledge is transferred. In this paper, we first propose a simple and efficient neural architecture for cross-lingual NER. Experiments show that our model achieves competitive performance with the state-of-the-art. We further explore how transfer learning works for cross-lingual NER on two transferable factors: sequential order and multilingual embedding. Our results shed light on future research for improving cross-lingual NER. D19-1672 10.18653/v1/D19-1672 @@ -7814,7 +7813,7 @@ XikaiLiu HongfeiLin FengXia - 6401–6406 + 6402–6407 Humor plays important role in human communication, which makes it important problem for natural language processing. Prior work on the analysis of humor focuses on whether text is humorous or not, or the degree of funniness, but this is insufficient to explain why it is funny. We therefore create a dataset on humor with 9,123 manually annotated jokes in Chinese. We propose a novel annotation scheme to give scenarios of how humor arises in text. Specifically, our annotations of linguistic humor not only contain the degree of funniness, like previous work, but they also contain key words that trigger humor as well as character relationship, scene, and humor categories. We report reasonable agreement between annota-tors. We also conduct an analysis and exploration of the dataset. To the best of our knowledge, we are the first to approach humor annotation for exploring the underlying mechanism of the use of humor, which may contribute to a significantly deeper analysis of humor. We also contribute with a scarce and valuable dataset, which we will release publicly. D19-1673 10.18653/v1/D19-1673 @@ -7825,7 +7824,7 @@ ShoTakase TsutomuHirao MasaakiNagata - 6407–6411 + 6408–6412 An anagram is a sentence or a phrase that is made by permutating the characters of an input sentence or a phrase. For example, “Trims cash” is an anagram of “Christmas”. Existing automatic anagram generation methods can find possible combinations of words form an anagram. However, they do not pay much attention to the naturalness of the generated anagrams. In this paper, we show that simple depth-first search can yield natural anagrams when it is combined with modern neural language models. Human evaluation results show that the proposed method can generate significantly more natural anagrams than baseline methods. D19-1674 10.18653/v1/D19-1674 @@ -7836,7 +7835,7 @@ SubhabrataMukherjee AndrewYates GerhardWeikum - 6412–6417 + 6413–6418 Controversial claims are abundant in online media and discussion forums. A better understanding of such claims requires analyzing them from different perspectives. Stance classification is a necessary step for inferring these perspectives in terms of supporting or opposing the claim. In this work, we present a neural network model for stance classification leveraging BERT representations and augmenting them with a novel consistency constraint. Experiments on the Perspectrum dataset, consisting of claims and users’ perspectives from various debate websites, demonstrate the effectiveness of our approach over state-of-the-art baselines. D19-1675 10.18653/v1/D19-1675 @@ -7848,7 +7847,7 @@ Georgios ChristosChouliaras AminaKeldibek MaartenVersteegh - 6418–6423 + 6419–6424 This paper explores different approaches to multilingual intent classification in a low resource setting. Recent advances in multilingual text representations promise cross-lingual transfer for classifiers. We investigate the potential for this transfer in an applied industrial setting and compare to multilingual classification using machine translated text. Our results show that while the recently developed methods show promise, practical application calls for a combination of techniques for useful results. D19-1676 10.18653/v1/D19-1676 @@ -7859,7 +7858,7 @@ Sang-KiKo MarcoCognetta Yo-SubHan - 6424–6430 + 6425–6431 We continue the study of generating se-mantically correct regular expressions from natural language descriptions (NL). The current state-of-the-art model SemRegex produces regular expressions from NLs by rewarding the reinforced learning based on the semantic (rather than syntactic) equivalence between two regular expressions. Since the regular expression equivalence problem is PSPACE-complete, we introduce the EQ_Reg model for computing the simi-larity of two regular expressions using deep neural networks. Our EQ_Reg mod-el essentially softens the equivalence of two regular expressions when used as a reward function. We then propose a new regex generation model, SoftRegex, us-ing the EQ_Reg model, and empirically demonstrate that SoftRegex substantially reduces the training time (by a factor of at least 3.6) and produces state-of-the-art results on three benchmark datasets. D19-1677 10.18653/v1/D19-1677 @@ -7870,7 +7869,7 @@ KaranAggarwal ShafiqJoty JaideepSrivastava - 6431–6436 + 6432–6437 Monitoring patients in ICU is a challenging and high-cost task. Hence, predicting the condition of patients during their ICU stay can help provide better acute care and plan the hospital’s resources. There has been continuous progress in machine learning research for ICU management, and most of this work has focused on using time series signals recorded by ICU instruments. In our work, we show that adding clinical notes as another modality improves the performance of the model for three benchmark tasks: in-hospital mortality prediction, modeling decompensation, and length of stay forecasting that play an important role in ICU management. While the time-series data is measured at regular intervals, doctor notes are charted at irregular times, making it challenging to model them together. We propose a method to model them jointly, achieving considerable improvement across benchmark tasks over baseline time-series model. D19-1678 10.18653/v1/D19-1678 @@ -7880,7 +7879,7 @@ AdithyaRenduchintala PhilippKoehn JasonEisner - 6437–6442 + 6438–6443 We present a machine foreign-language teacher that modifies text in a student’s native language (L1) by replacing some word tokens with glosses in a foreign language (L2), in such a way that the student can acquire L2 vocabulary simply by reading the resulting macaronic text. The machine teacher uses no supervised data from human students. Instead, to guide the machine teacher’s choice of which words to replace, we equip a cloze language model with a training procedure that can incrementally learn representations for novel words, and use this model as a proxy for the word guessing and learning ability of real human students. We use Mechanical Turk to evaluate two variants of the student model: (i) one that generates a representation for a novel word using only surrounding context and (ii) an extension that also uses the spelling of the novel word. D19-1679 D19-1679.Attachment.pdf @@ -7892,7 +7891,7 @@ Yee SengChan HaolingQiu JoshuaFasching - 6443–6447 + 6444–6448 Solving long-lasting problems such as food insecurity requires a comprehensive understanding of interventions applied by governments and international humanitarian assistance organizations, and their results and consequences. Towards achieving this grand goal, a crucial first step is to extract past interventions and when and where they have been applied, from hundreds of thousands of reports automatically. In this paper, we developed a corpus annotated with interventions to foster research, and developed an information extraction system for extracting interventions and their location and time from text. We demonstrate early, very encouraging results on extracting interventions. D19-1680 10.18653/v1/D19-1680 @@ -7901,7 +7900,7 @@ <fixed-case>RUN</fixed-case> through the Streets: A New Dataset and Baseline Models for Realistic Urban Navigation TzufPaz-Argaman ReutTsarfaty - 6448–6454 + 6449–6455 Following navigation instructions in natural language (NL) requires a composition of language, action, and knowledge of the environment. Knowledge of the environment may be provided via visual sensors or as a symbolic world representation referred to as a map. Previous work on map-based NL navigation relied on small artificial worlds with a fixed set of entities known in advance. Here we introduce the Realistic Urban Navigation (RUN) task, aimed at interpreting NL navigation instructions based on a real, dense, urban map. Using Amazon Mechanical Turk, we collected a dataset of 2515 instructions aligned with actual routes over three regions of Manhattan. We then empirically study which aspects of a neural architecture are important for the RUN success, and empirically show that entity abstraction, attention over words and worlds, and a constantly updating world-state, significantly contribute to task accuracy. D19-1681 D19-1681.Attachment.zip @@ -7917,7 +7916,7 @@ XiaoxiaoGuo ShiyuChang MoYu - 6455–6460 + 6456–6461 In multi-party chat, it is common for multiple conversations to occur concurrently, leading to intermingled conversation threads in chat logs. In this work, we propose a novel Context-Aware Thread Detection (CATD) model that automatically disentangles these conversation threads. We evaluate our model on four real-world datasets and demonstrate an overall im-provement in thread detection accuracy over state-of-the-art benchmarks. D19-1682 D19-1682.Attachment.zip diff --git a/data/xml/R19.xml b/data/xml/R19.xml new file mode 100644 index 0000000000..327aac62c8 --- /dev/null +++ b/data/xml/R19.xml @@ -0,0 +1,1751 @@ + + + + + Natural Language Processing in a Deep Learning World + R19-1 + RuslanMitkov + GaliaAngelova + INCOMA Ltd. +
Varna, Bulgaria
+ September + 2019 + + + R19-1000 + + + Table Structure Recognition Based on Cell Relationship, a Bottom-Up Approach + DarshanAdiga + Shabir AhmadBhat + Muzaffar BashirShah + VivekaVyeth + 1–8 + In this paper, we present a relationship extraction based methodology for table structure recognition in PDF documents. The proposed deep learning-based method takes a bottom-up approach to table recognition in PDF documents. We outline the shortcomings of conventional approaches based on heuristics and machine learning-based top-down approaches. In this work, we explain how the task of table structure recognition can be modeled as a cell relationship extraction task and the importance of the bottom-up approach in recognizing the table cells. We use Multilayer Feedforward Neural Network for table structure recognition and compare the results of three feature sets. To gauge the performance of the proposed method, we prepared a training dataset using 250 tables in PDF documents, carefully selecting the table structures that are most commonly found in the documents. Our model achieves an overall accuracy of 97.95% and an F1-Score of 92.62% on the test dataset. + R19-1001 + 10.26615/978-954-452-056-4_001 + + + Identification of Good and Bad News on <fixed-case>T</fixed-case>witter + PiushAggarwal + AhmetAker + 9–17 + Social media plays a great role in news dissemination which includes good and bad news. However, studies show that news, in general, has a significant impact on our mental stature and that this influence is more in bad news. An ideal situation would be that we have a tool that can help to filter out the type of news we do not want to consume. In this paper, we provide the basis for such a tool. In our work, we focus on Twitter. We release a manually annotated dataset containing 6,853 tweets from 5 different topical categories. Each tweet is annotated with good and bad labels. We also investigate various machine learning systems and features and evaluate their performance on the newly generated dataset. We also perform a comparative analysis with sentiments showing that sentiment alone is not enough to distinguish between good and bad news. + R19-1002 + 10.26615/978-954-452-056-4_002 + + + Bilingual Low-Resource Neural Machine Translation with Round-Tripping: The Case of <fixed-case>P</fixed-case>ersian-<fixed-case>S</fixed-case>panish + BenyaminAhmadnia + BonnieDorr + 18–24 + The quality of Neural Machine Translation (NMT), as a data-driven approach, massively depends on quantity, quality, and relevance of the training dataset. Such approaches have achieved promising results for bilingually high-resource scenarios but are inadequate for low-resource conditions. This paper describes a round-trip training approach to bilingual low-resource NMT that takes advantage of monolingual datasets to address training data scarcity, thus augmenting translation quality. We conduct detailed experiments on Persian-Spanish as a bilingually low-resource scenario. Experimental results demonstrate that this competitive approach outperforms the baselines. + R19-1003 + 10.26615/978-954-452-056-4_003 + + + Enhancing Phrase-Based Statistical Machine Translation by Learning Phrase Representations Using Long Short-Term Memory Network + BenyaminAhmadnia + BonnieDorr + 25–32 + Phrases play a key role in Machine Translation (MT). In this paper, we apply a Long Short-Term Memory (LSTM) model over conventional Phrase-Based Statistical MT (PBSMT). The core idea is to use an LSTM encoder-decoder to score the phrase table generated by the PBSMT decoder. Given a source sequence, the encoder and decoder are jointly trained in order to maximize the conditional probability of a target sequence. Analytically, the performance of a PBSMT system is enhanced by using the conditional probabilities of phrase pairs computed by an LSTM encoder-decoder as an additional feature in the existing log-linear model. We compare the performance of the phrase tables in the PBSMT to the performance of the proposed LSTM and observe its positive impact on translation quality. We construct a PBSMT model using the Moses decoder and enrich the Language Model (LM) utilizing an external dataset. We then rank the phrase tables using an LSTM-based encoder-decoder. This method produces a gain of up to 3.14 BLEU score on the test set. + R19-1004 + 10.26615/978-954-452-056-4_004 + + + Automatic <fixed-case>P</fixed-case>ropbank Generation for <fixed-case>T</fixed-case>urkish + KorayAK + Olcay TanerYıldız + 33–41 + Semantic role labeling (SRL) is an important task for understanding natural languages, where the objective is to analyse propositions expressed by the verb and to identify each word that bears a semantic role. It provides an extensive dataset to enhance NLP applications such as information retrieval, machine translation, information extraction, and question answering. However, creating SRL models are difficult. Even in some languages, it is infeasible to create SRL models that have predicate-argument structure due to lack of linguistic resources. In this paper, we present our method to create an automatic Turkish PropBank by exploiting parallel data from the translated sentences of English PropBank. Experiments show that our method gives promising results. + R19-1005 + 10.26615/978-954-452-056-4_005 + + + Multilingual sentence-level bias detection in <fixed-case>W</fixed-case>ikipedia + DesislavaAleksandrova + FrançoisLareau + Pierre AndréMénard + 42–51 + We propose a multilingual method for the extraction of biased sentences from Wikipedia, and use it to create corpora in Bulgarian, French and English. Sifting through the revision history of the articles that at some point had been considered biased and later corrected, we retrieve the last tagged and the first untagged revisions as the before/after snapshots of what was deemed a violation of Wikipedia’s neutral point of view policy. We extract the sentences that were removed or rewritten in that edit. The approach yields sufficient data even in the case of relatively small Wikipedias, such as the Bulgarian one, where 62k articles produced 5k biased sentences. We evaluate our method by manually annotating 520 sentences for Bulgarian and French, and 744 for English. We assess the level of noise and analyze its sources. Finally, we exploit the data with well-known classification methods to detect biased sentences. Code and datasets are hosted at https://github.com/crim-ca/wiki-bias. + R19-1006 + 10.26615/978-954-452-056-4_006 + + + Supervised Morphological Segmentation Using Rich Annotated Lexicon + EbrahimAnsari + ZdeněkŽabokrtský + MohammadMahmoudi + HamidHaghdoost + JonášVidra + 52–61 + Morphological segmentation of words is the process of dividing a word into smaller units called morphemes; it is tricky especially when a morphologically rich or polysynthetic language is under question. In this work, we designed and evaluated several Recurrent Neural Network (RNN) based models as well as various other machine learning based approaches for the morphological segmentation task. We trained our models using annotated segmentation lexicons. To evaluate the effect of the training data size on our models, we decided to create a large hand-annotated morphologically segmented corpus of Persian words, which is, to the best of our knowledge, the first and the only segmentation lexicon for the Persian language. In the experimental phase, using the hand-annotated Persian lexicon and two smaller similar lexicons for Czech and Finnish languages, we evaluated the effect of the training data size, different hyper-parameters settings as well as different RNN-based models. + R19-1007 + 10.26615/978-954-452-056-4_007 + + + Combining Lexical Substitutes in Neural Word Sense Induction + NikolayArefyev + BorisSheludko + AlexanderPanchenko + 62–70 + Word Sense Induction (WSI) is the task of grouping of occurrences of an ambiguous word according to their meaning. In this work, we improve the approach to WSI proposed by Amrami and Goldberg (2018) based on clustering of lexical substitutes for an ambiguous word in a particular context obtained from neural language models. Namely, we propose methods for combining information from left and right context and similarity to the ambiguous word, which result in generating more accurate substitutes than the original approach. Our simple yet efficient improvement establishes a new state-of-the-art on WSI datasets for two languages. Besides, we show improvements to the original approach on a lexical substitution dataset. + R19-1008 + 10.26615/978-954-452-056-4_008 + + + Detecting Clitics Related Orthographic Errors in <fixed-case>T</fixed-case>urkish + UgurcanArikan + OnurGungor + SuzanUskudarli + 71–76 + For the spell correction task, vocabulary based methods have been replaced with methods that take morphological and grammar rules into account. However, such tools are fairly immature, and, worse, non-existent for many low resource languages. Checking only if a word is well-formed with respect to the morphological rules of a language may produce false negatives due to the ambiguity resulting from the presence of numerous homophonic words. In this work, we propose an approach to detect and correct the “de/da” clitic errors in Turkish text. Our model is a neural sequence tagger trained with a synthetically constructed dataset consisting of positive and negative samples. The model’s performance with this dataset is presented according to different word embedding configurations. The model achieved an F1 score of 86.67% on a synthetically constructed dataset. We also compared the model’s performance on a manually curated dataset of challenging samples that proved superior to other spelling correctors with 71% accuracy compared to the second-best (Google Docs) with and accuracy of 34%. + R19-1009 + 10.26615/978-954-452-056-4_009 + + + Benchmark Dataset for Propaganda Detection in <fixed-case>C</fixed-case>zech Newspaper Texts + VítBaisa + OndřejHerman + AlesHorak + 77–83 + Propaganda of various pressure groups ranging from big economies to ideological blocks is often presented in a form of objective newspaper texts. However, the real objectivity is here shaded with the support of imbalanced views and distorted attitudes by means of various manipulative stylistic techniques. In the project of Manipulative Propaganda Techniques in the Age of Internet, a new resource for automatic analysis of stylistic mechanisms for influencing the readers’ opinion is developed. In its current version, the resource consists of 7,494 newspaper articles from four selected Czech digital news servers annotated for the presence of specific manipulative techniques. In this paper, we present the current state of the annotations and describe the structure of the dataset in detail. We also offer an evaluation of bag-of-words classification algorithms for the annotated manipulative techniques. + R19-1010 + 10.26615/978-954-452-056-4_010 + + + Diachronic Analysis of Entities by Exploiting <fixed-case>W</fixed-case>ikipedia Page revisions + PierpaoloBasile + AnnalinaCaputo + SeamusLawless + GiovanniSemeraro + 84–91 + In the last few years, the increasing availability of large corpora spanning several time periods has opened new opportunities for the diachronic analysis of language. This type of analysis can bring to the light not only linguistic phenomena related to the shift of word meanings over time, but it can also be used to study the impact that societal and cultural trends have on this language change. This paper introduces a new resource for performing the diachronic analysis of named entities built upon Wikipedia page revisions. This resource enables the analysis over time of changes in the relations between entities (concepts), surface forms (words), and the contexts surrounding entities and surface forms, by analysing the whole history of Wikipedia internal links. We provide some useful use cases that prove the impact of this resource on diachronic studies and delineate some possible future usage. + R19-1011 + 10.26615/978-954-452-056-4_011 + + + Using a Lexical Semantic Network for the Ontology Building + NadiaBebeshina-Clairet + SylvieDespres + MathieuLafourcade + 92–101 + Building multilingual ontologies is a hard task as ontologies are often data-rich resources. We introduce an approach which allows exploiting structured lexical semantic knowledge for the ontology building. Given a multilingual lexical semantic (non ontological) resource and an ontology model, it allows mining relevant semantic knowledge and make the ontology building and enhancement process faster. + R19-1012 + 10.26615/978-954-452-056-4_012 + + + Naive Regularizers for Low-Resource Neural Machine Translation + MeriemBeloucif + Ana ValeriaGonzalez + MarcelBollmann + AndersSøgaard + 102–111 + Neural machine translation models have little inductive bias, which can be a disadvantage in low-resource scenarios. Neural models have to be trained on large amounts of data and have been shown to perform poorly when only limited data is available. We show that using naive regularization methods, based on sentence length, punctuation and word frequencies, to penalize translations that are very different from the input sentences, consistently improves the translation quality across multiple low-resource languages. We experiment with 12 language pairs, varying the training data size between 17k to 230k sentence pairs. Our best regularizer achieves an average increase of 1.5 BLEU score and 1.0 TER score across all the language pairs. For example, we achieve a BLEU score of 26.70 on the IWSLT15 English–Vietnamese translation task simply by using relative differences in punctuation as a regularizer. + R19-1013 + 10.26615/978-954-452-056-4_013 + + + Exploring Graph-Algebraic <fixed-case>CCG</fixed-case> Combinators for Syntactic-Semantic <fixed-case>AMR</fixed-case> Parsing + SebastianBeschke + 112–121 + We describe a new approach to semantic parsing based on Combinatory Categorial Grammar (CCG). The grammar’s semantic construction operators are defined in terms of a graph algebra, which allows our system to induce a compact CCG lexicon. We introduce an expectation maximisation algorithm which we use to filter our lexicon down to 2500 lexical templates. Our system achieves a semantic triple (Smatch) precision that is competitive with other CCG-based AMR parsing approaches. + R19-1014 + 10.26615/978-954-452-056-4_014 + + + Quasi Bidirectional Encoder Representations from Transformers for Word Sense Disambiguation + MicheleBevilacqua + RobertoNavigli + 122–131 + While contextualized embeddings have produced performance breakthroughs in many Natural Language Processing (NLP) tasks, Word Sense Disambiguation (WSD) has not benefited from them yet. In this paper, we introduce QBERT, a Transformer-based architecture for contextualized embeddings which makes use of a co-attentive layer to produce more deeply bidirectional representations, better-fitting for the WSD task. As a result, we are able to train a WSD system that beats the state of the art on the concatenation of all evaluation datasets by over 3 points, also outperforming a comparable model using ELMo. + R19-1015 + 10.26615/978-954-452-056-4_015 + + + Evaluating the Consistency of Word Embeddings from Small Data + JelkeBloem + AntskeFokkens + AurélieHerbelot + 132–141 + In this work, we address the evaluation of distributional semantic models trained on smaller, domain-specific texts, specifically, philosophical text. Specifically, we inspect the behaviour of models using a pre-trained background space in learning. We propose a measure of consistency which can be used as an evaluation metric when no in-domain gold-standard data is available. This measure simply computes the ability of a model to learn similar embeddings from different parts of some homogeneous data. We show that in spite of being a simple evaluation, consistency actually depends on various combinations of factors, including the nature of the data itself, the model used to train the semantic space, and the frequency of the learnt terms, both in the background space and in the in-domain data of interest. + R19-1016 + 10.26615/978-954-452-056-4_016 + + + Cross-Domain Training for Goal-Oriented Conversational Agents + Alexandra MariaBodîrlău + StefaniaBudulan + TraianRebedea + 142–150 + Goal-Oriented Chatbots in fields such as customer support, providing certain information or general help with bookings or reservations, suffer from low performance partly due to the difficulty of obtaining large domain-specific annotated datasets. Given that the problem is closely related to the domain of the conversational agent and the data belonging to a specific domain is difficult to annotate, there have been some attempts at surpassing these challenges such as unsupervised pre-training or transfer learning between different domains. A more thorough analysis of the transfer learning mechanism is justified by the significant improvement of the results demonstrated in the results section. We describe extensive experiments using transfer learning and warm-starting techniques with improvements of more than 5% in relative percentage of success rate in the majority of cases, and up to 10x faster convergence as opposed to training the system without them. + R19-1017 + 10.26615/978-954-452-056-4_017 + + + Learning Sentence Embeddings for Coherence Modelling and Beyond + TannerBohn + YiningHu + JinhangZhang + CharlesLing + 151–160 + We present a novel and effective technique for performing text coherence tasks while facilitating deeper insights into the data. Despite obtaining ever-increasing task performance, modern deep-learning approaches to NLP tasks often only provide users with the final network decision and no additional understanding of the data. In this work, we show that a new type of sentence embedding learned through self-supervision can be applied effectively to text coherence tasks while serving as a window through which deeper understanding of the data can be obtained. To produce these sentence embeddings, we train a recurrent neural network to take individual sentences and predict their location in a document in the form of a distribution over locations. We demonstrate that these embeddings, combined with simple visual heuristics, can be used to achieve performance competitive with state-of-the-art on multiple text coherence tasks, outperforming more complex and specialized approaches. Additionally, we demonstrate that these embeddings can provide insights useful to writers for improving writing quality and informing document structuring, and assisting readers in summarizing and locating information. + R19-1018 + 10.26615/978-954-452-056-4_018 + + + Risk Factors Extraction from Clinical Texts based on Linked Open Data + SvetlaBoytcheva + GaliaAngelova + ZhivkoAngelov + 161–167 + This paper presents experiments in risk factors analysis based on clinical texts enhanced with Linked Open Data (LOD). The idea is to determine whether a patient has risk factors for a specific disease analyzing only his/her outpatient records. A semantic graph of “meta-knowledge” about a disease of interest is constructed, with integrated multilingual terms (labels) of symptoms, risk factors etc. coming from Wikidata, PubMed, Wikipedia and MESH, and linked to clinical records of individual patients via ICD–10 codes. Then a predictive model is trained to foretell whether patients are at risk to develop the disease of interest. The testing was done using outpatient records from a nation-wide repository available for the period 2011-2016. The results show improvement of the overall performance of all tested algorithms (kNN, Naive Bayes, Tree, Logistic regression, ANN), when the clinical texts are enriched with LOD resources. + R19-1019 + 10.26615/978-954-452-056-4_019 + + + Parallel Sentence Retrieval From Comparable Corpora for Biomedical Text Simplification + RémiCardon + NataliaGrabar + 168–177 + Parallel sentences provide semantically similar information which can vary on a given dimension, such as language or register. Parallel sentences with register variation (like expert and non-expert documents) can be exploited for the automatic text simplification. The aim of automatic text simplification is to better access and understand a given information. In the biomedical field, simplification may permit patients to understand medical and health texts. Yet, there is currently no such available resources. We propose to exploit comparable corpora which are distinguished by their registers (specialized and simplified versions) to detect and align parallel sentences. These corpora are in French and are related to the biomedical area. Manually created reference data show 0.76 inter-annotator agreement. Our purpose is to state whether a given pair of specialized and simplified sentences is parallel and can be aligned or not. We treat this task as binary classification (alignment/non-alignment). We perform experiments with a controlled ratio of imbalance and on the highly unbalanced real data. Our results show that the method we present here can be used to automatically generate a corpus of parallel sentences from our comparable corpus. + R19-1020 + 10.26615/978-954-452-056-4_020 + + + Classifying Author Intention for Writer Feedback in Related Work + ArleneCasey + BonnieWebber + DorotaGlowacka + 178–187 + The ability to produce high-quality publishable material is critical to academic success but many Post-Graduate students struggle to learn to do so. While recent years have seen an increase in tools designed to provide feedback on aspects of writing, one aspect that has so far been neglected is the Related Work section of academic research papers. To address this, we have trained a supervised classifier on a corpus of 94 Related Work sections and evaluated it against a manually annotated gold standard. The classifier uses novel features pertaining to citation types and co-reference, along with patterns found from studying Related Works. We show that these novel features contribute to classifier performance with performance being favourable compared to other similar works that classify author intentions and consider feedback for academic writing. + R19-1021 + 10.26615/978-954-452-056-4_021 + + + Sparse Victory – A Large Scale Systematic Comparison of count-based and prediction-based vectorizers for text classification + RupakChakraborty + AshimaElhence + KapilArora + 188–197 + In this paper we study the performance of several text vectorization algorithms on a diverse collection of 73 publicly available datasets. Traditional sparse vectorizers like Tf-Idf and Feature Hashing have been systematically compared with the latest state of the art neural word embeddings like Word2Vec, GloVe, FastText and character embeddings like ELMo, Flair. We have carried out an extensive analysis of the performance of these vectorizers across different dimensions like classification metrics (.i.e. precision, recall, accuracy), dataset-size, and imbalanced data (in terms of the distribution of the number of class labels). Our experiments reveal that the sparse vectorizers beat the neural word and character embedding models on 61 of the 73 datasets by an average margin of 3-5% (in terms of macro f1 score) and this performance is consistent across the different dimensions of comparison. + R19-1022 + 10.26615/978-954-452-056-4_022 + + + A Fine-Grained Annotated Multi-Dialectal <fixed-case>A</fixed-case>rabic Corpus + AnisCharfi + WajdiZaghouani + Syed HassanMehdi + EsraaMohamed + 198–204 + We present ARAP-Tweet 2.0, a corpus of 5 million dialectal Arabic tweets and 50 million words of about 3000 Twitter users from 17 Arab countries. Compared to the first version, the new corpus has significant improvements in terms of the data volume and the annotation quality. It is fully balanced with respect to dialect, gender, and three age groups: under 25 years, between 25 and 34, and 35 years and above. This paper describes the process of creating the corpus starting from gathering the dialectal phrases to find the users, to annotating their accounts and retrieving their tweets. We also report on the evaluation of the annotation quality using the inter-annotator agreement measures which were applied to the whole corpus and not just a subset. The obtained results were substantial with average Cohen’s Kappa values of 0.99, 0.92, and 0.88 for the annotation of gender, dialect, and age respectively. We also discuss some challenges encountered when developing this corpus.s. + R19-1023 + 10.26615/978-954-452-056-4_023 + + + Personality-dependent Neural Text Summarization + PabloCosta + IvandréParaboni + 205–212 + In Natural Language Generation systems, personalization strategies - i.e, the use of information about a target author to generate text that (more) closely resembles human-produced language - have long been applied to improve results. The present work addresses one such strategy - namely, the use of Big Five personality information about the target author - applied to the case of abstractive text summarization using neural sequence-to-sequence models. Initial results suggest that having access to personality information does lead to more accurate (or human-like) text summaries, and paves the way for more robust systems of this kind. + R19-1024 + 10.26615/978-954-452-056-4_024 + + + Self-Adaptation for Unsupervised Domain Adaptation + XiaCui + DanushkaBollegala + 213–222 + Lack of labelled data in the target domain for training is a common problem in domain adaptation. To overcome this problem, we propose a novel unsupervised domain adaptation method that combines projection and self-training based approaches. Using the labelled data from the source domain, we first learn a projection that maximises the distance among the nearest neighbours with opposite labels in the source domain. Next, we project the source domain labelled data using the learnt projection and train a classifier for the target class prediction. We then use the trained classifier to predict pseudo labels for the target domain unlabelled data. Finally, we learn a projection for the target domain as we did for the source domain using the pseudo-labelled target domain data, where we maximise the distance between nearest neighbours having opposite pseudo labels. Experiments on a standard benchmark dataset for domain adaptation show that the proposed method consistently outperforms numerous baselines and returns competitive results comparable to that of SOTA including self-training, tri-training, and neural adaptations. + R19-1025 + 10.26615/978-954-452-056-4_025 + + + Speculation and Negation detection in <fixed-case>F</fixed-case>rench biomedical corpora + ClémentDalloux + VincentClaveau + NataliaGrabar + 223–232 + In this work, we propose to address the detection of negation and speculation, and of their scope, in French biomedical documents. It has been indeed observed that they play an important role and provide crucial clues for other NLP applications. Our methods are based on CRFs and BiLSTM. We reach up to 97.21 % and 91.30 % F-measure for the detection of negation and speculation cues, respectively, using CRFs. For the computing of scope, we reach up to 90.81 % and 86.73 % F-measure on negation and speculation, respectively, using BiLSTM-CRF fed with word embeddings. + R19-1026 + 10.26615/978-954-452-056-4_026 + + + Porting Multilingual Morphological Resources to <fixed-case>O</fixed-case>nto<fixed-case>L</fixed-case>ex-Lemon + ThierryDeclerck + StefaniaRacioppa + 233–238 + We describe work consisting in porting various morphological resources to the OntoLex-Lemon model. A main objective of this work is to offer a uniform representation of different morphological data sets in order to be able to compare and interlink multilingual resources and to cross-check and interlink or merge the content of morphological resources of one and the same language. The results of our work will be published on the Linguistic Linked Open Data cloud. + R19-1027 + 10.26615/978-954-452-056-4_027 + + + Dependency-Based Self-Attention for Transformer <fixed-case>NMT</fixed-case> + HiroyukiDeguchi + AkihiroTamura + TakashiNinomiya + 239–246 + In this paper, we propose a new Transformer neural machine translation (NMT) model that incorporates dependency relations into self-attention on both source and target sides, dependency-based self-attention. The dependency-based self-attention is trained to attend to the modifiee for each token under constraints based on the dependency relations, inspired by Linguistically-Informed Self-Attention (LISA). While LISA is originally proposed for Transformer encoder for semantic role labeling, this paper extends LISA to Transformer NMT by masking future information on words in the decoder-side dependency-based self-attention. Additionally, our dependency-based self-attention operates at sub-word units created by byte pair encoding. The experiments show that our model improves 1.0 BLEU points over the baseline model on the WAT’18 Asian Scientific Paper Excerpt Corpus Japanese-to-English translation task. + R19-1028 + 10.26615/978-954-452-056-4_028 + + + Detecting Toxicity in News Articles: Application to <fixed-case>B</fixed-case>ulgarian + YoanDinkov + IvanKoychev + PreslavNakov + 247–258 + Online media aim for reaching ever bigger audience and for attracting ever longer attention span. This competition creates an environment that rewards sensational, fake, and toxic news. To help limit their spread and impact, we propose and develop a news toxicity detector that can recognize various types of toxic content. While previous research primarily focused on English, here we target Bulgarian. We created a new dataset by crawling a website that for five years has been collecting Bulgarian news articles that were manually categorized into eight toxicity groups. Then we trained a multi-class classifier with nine categories: eight toxic and one non-toxic. We experimented with different representations based on ElMo, BERT, and XLM, as well as with a variety of domain-specific features. Due to the small size of our dataset, we created a separate model for each feature type, and we ultimately combined these models into a meta-classifier. The evaluation results show an accuracy of 59.0% and a macro-F1 score of 39.7%, which represent sizable improvements over the majority-class baseline (Acc=30.3%, macro-F1=5.2%). + R19-1029 + 10.26615/978-954-452-056-4_029 + + + De-Identification of Emails: Pseudonymizing Privacy-Sensitive Data in a <fixed-case>G</fixed-case>erman Email Corpus + ElisabethEder + UlrikeKrieg-Holz + UdoHahn + 259–269 + We deal with the pseudonymization of those stretches of text in emails that might allow to identify real individual persons. This task is decomposed into two steps. First, named entities carrying privacy-sensitive information (e.g., names of persons, locations, phone numbers or dates) are identified, and, second, these privacy-bearing entities are replaced by synthetically generated surrogates (e.g., a person originally named ‘John Doe’ is renamed as ‘Bill Powers’). We describe a system architecture for surrogate generation and evaluate our approach on CodeAlltag, a German email corpus. + R19-1030 + 10.26615/978-954-452-056-4_030 + + + Lexical Quantile-Based Text Complexity Measure + MaksimEremeev + KonstantinVorontsov + 270–275 + This paper introduces a new approach to estimating the text document complexity. Common readability indices are based on average length of sentences and words. In contrast to these methods, we propose to count the number of rare words occurring abnormally often in the document. We use the reference corpus of texts and the quantile approach in order to determine what words are rare, and what frequencies are abnormal. We construct a general text complexity model, which can be adjusted for the specific task, and introduce two special models. The experimental design is based on a set of thematically similar pairs of Wikipedia articles, labeled using crowdsourcing. The experiments demonstrate the competitiveness of the proposed approach. + R19-1031 + 10.26615/978-954-452-056-4_031 + + + Demo Application for <fixed-case>LETO</fixed-case>: Learning Engine Through Ontologies + SuilanEstevez-Velarde + AndrésMontoyo + YudivianAlmeida-Cruz + YoanGutiérrez + AlejandroPiad-Morffis + RafaelMuñoz + 276–284 + The massive amount of multi-formatted information available on the Web necessitates the design of software systems that leverage this information to obtain knowledge that is valid and useful. The main challenge is to discover relevant information and continuously update, enrich and integrate knowledge from various sources of structured and unstructured data. This paper presents the Learning Engine Through Ontologies(LETO) framework, an architecture for the continuous and incremental discovery of knowledge from multiple sources of unstructured and structured data. We justify the main design decision behind LETO’s architecture and evaluate the framework’s feasibility using the Internet Movie Data Base(IMDB) and Twitter as a practical application. + R19-1032 + 10.26615/978-954-452-056-4_032 + + + Sentence Simplification for Semantic Role Labelling and Information Extraction + RichardEvans + ConstantinOrasan + 285–294 + In this paper, we report on the extrinsic evaluation of an automatic sentence simplification method with respect to two NLP tasks: semantic role labelling (SRL) and information extraction (IE). The paper begins with our observation of challenges in the intrinsic evaluation of sentence simplification systems, which motivates the use of extrinsic evaluation of these systems with respect to other NLP tasks. We describe the two NLP systems and the test data used in the extrinsic evaluation, and present arguments and evidence motivating the integration of a sentence simplification step as a means of improving the accuracy of these systems. Our evaluation reveals that their performance is improved by the simplification step: the SRL system is better able to assign semantic roles to the majority of the arguments of verbs and the IE system is better able to identify fillers for all IE template slots. + R19-1033 + 10.26615/978-954-452-056-4_033 + + + <fixed-case>O</fixed-case>llo<fixed-case>B</fixed-case>ot - Towards A Text-Based <fixed-case>A</fixed-case>rabic Health Conversational Agent: Evaluation and Results + AhmedFadhil + AhmedAbuRa’ed + 295–303 + We introduce OlloBot, an Arabic conversational agent that assists physicians and supports patients with the care process. It doesn’t replace the physicians, instead provides health tracking and support and assists physicians with the care delivery through a conversation medium. The current model comprises healthy diet, physical activity, mental health, in addition to food logging. Not only OlloBot tracks user daily food, it also offers useful tips for healthier living. We will discuss the design, development and testing of OlloBot, and highlight the findings and limitations arose from the testing. + R19-1034 + 10.26615/978-954-452-056-4_034 + + + Developing the Old <fixed-case>T</fixed-case>ibetan Treebank + ChristianFaggionato + MariekeMeelen + 304–312 + This paper presents a full procedure for the development of a segmented, POS-tagged and chunkparsed corpus of Old Tibetan. As an extremely low-resource language, Old Tibetan poses non-trivial problems in every step towards the development of a searchable treebank. We demonstrate, however, that a carefully developed, semisupervised method of optimising and extending existing tools for Classical Tibetan, as well as creating specific ones for Old Tibetan can address these issues. We thus also present the first very Tibetan Treebank in a variety of formats to facilitate research in the fields of NLP, historical linguistics and Tibetan Studies. + R19-1035 + 10.26615/978-954-452-056-4_035 + + + Summarizing Legal Rulings: Comparative Experiments + DiegoFeijo + VivianeMoreira + 313–322 + In the context of text summarization, texts in the legal domain have peculiarities related to their length and to their specialized vocabulary. Recent neural network-based approaches can achieve high-quality scores for text summarization. However, these approaches have been used mostly for generating very short abstracts for news articles. Thus, their applicability to the legal domain remains an open issue. In this work, we experimented with ten extractive and four abstractive models in a real dataset of legal rulings. These models were compared with an extractive baseline based on heuristics to select the most relevant parts of the text. Our results show that abstractive approaches significantly outperform extractive methods in terms of ROUGE scores. + R19-1036 + 10.26615/978-954-452-056-4_036 + + + Entropy as a Proxy for Gap Complexity in Open Cloze Tests + MarianoFelice + PaulaButtery + 323–327 + This paper presents a pilot study of entropy as a measure of gap complexity in open cloze tests aimed at learners of English. Entropy is used to quantify the information content in each gap, which can be used to estimate complexity. Our study shows that average gap entropy correlates positively with proficiency levels while individual gap entropy can capture contextual complexity. To the best of our knowledge, this is the first unsupervised information-theoretical approach to evaluating the quality of cloze tests. + R19-1037 + 10.26615/978-954-452-056-4_037 + + + Song Lyrics Summarization Inspired by Audio Thumbnailing + MichaelFell + ElenaCabrio + FabienGandon + AlainGiboin + 328–337 + Given the peculiar structure of songs, applying generic text summarization methods to lyrics can lead to the generation of highly redundant and incoherent text. In this paper, we propose to enhance state-of-the-art text summarization approaches with a method inspired by audio thumbnailing. Instead of searching for the thumbnail clues in the audio of the song, we identify equivalent clues in the lyrics. We then show how these summaries that take into account the audio nature of the lyrics outperform the generic methods according to both an automatic evaluation and human judgments. + R19-1038 + 10.26615/978-954-452-056-4_038 + + + Comparing Automated Methods to Detect Explicit Content in Song Lyrics + MichaelFell + ElenaCabrio + MicheleCorazza + FabienGandon + 338–344 + The Parental Advisory Label (PAL) is a warning label that is placed on audio recordings in recognition of profanity or inappropriate references, with the intention of alerting parents of material potentially unsuitable for children. Since 2015, digital providers – such as iTunes, Spotify, Amazon Music and Deezer – also follow PAL guidelines and tag such tracks as “explicit”. Nowadays, such labelling is carried out mainly manually on voluntary basis, with the drawbacks of being time consuming and therefore costly, error prone and partly a subjective task. In this paper, we compare automated methods ranging from dictionary-based lookup to state-of-the-art deep neural networks to automatically detect explicit contents in English lyrics. We show that more complex models perform only slightly better on this task, and relying on a qualitative analysis of the data, we discuss the inherent hardness and subjectivity of the task. + R19-1039 + 10.26615/978-954-452-056-4_039 + + + Linguistic classification: dealing jointly with irrelevance and inconsistency + LauraFranzoi + AndreaSgarro + AncaDinu + Liviu P.Dinu + 345–352 + In this paper, we present new methods for language classification which put to good use both syntax and fuzzy tools, and are capable of dealing with irrelevant linguistic features (i.e. features which should not contribute to the classification) and even inconsistent features (which do not make sense for specific languages). We introduce a metric distance, based on the generalized Steinhaus transform, which allows one to deal jointly with irrelevance and inconsistency. To evaluate our methods, we test them on a syntactic data set, due to the linguist G. Longobardi and his school. We obtain phylogenetic trees which sometimes outperform the ones obtained by Atkinson and Gray. + R19-1040 + 10.26615/978-954-452-056-4_040 + + + Corpus Lexicography in a Wider Context + ChenGafni + 353–359 + This paper describes a set of tools that offers comprehensive solutions for corpus lexicography. The tools perform a range of tasks, including construction of corpus lexicon, integrating information from external dictionaries, internal analysis of the lexicon, and lexical analysis of the corpus. The set of tools is particularly useful for creating dictionaries for under-resourced languages. The tools are integrated in a general-purpose software that includes additional tools for various research tasks, such as linguistic development analysis. Equipped with a user-friendly interface, the described system can be easily incorporated in research in a variety of fields. + R19-1041 + 10.26615/978-954-452-056-4_041 + + + A Universal System for Automatic Text-to-Phonetics Conversion + ChenGafni + 360–366 + This paper describes an automatic text-to-phonetics conversion system. The system was constructed to primarily serve as a research tool. It is implemented in a general-purpose linguistic software, which allows it to be incorporated in a multifaceted linguistic research in essentially any language. The system currently relies on two mechanisms to generate phonetic transcriptions from texts: (i) importing ready-made phonetic word forms from external dictionaries, and (ii) automatic generation of phonetic word forms based on a set of deterministic linguistic rules. The current paper describes the proposed system and its potential application to linguistic research. + R19-1042 + 10.26615/978-954-452-056-4_042 + + + Two Discourse Tree - Based Approaches to Indexing Answers + BorisGalitsky + DmitryIlvovsky + 367–372 + We explore anatomy of answers with respect to which text fragments from an answer are worth matching with a question and which should not be matched. We apply the Rhetorical Structure Theory to build a discourse tree of an answer and select elementary discourse units that are suitable for indexing. Manual rules for selection of these discourse units as well as automated classification based on web search engine mining are evaluated con-cerning improving search accuracy. We form two sets of question-answer pairs for FAQ and community QA search domains and use them for evaluation of the proposed indexing methodology, which delivers up to 16 percent improvement in search recall. + R19-1043 + 10.26615/978-954-452-056-4_043 + + + Discourse-Based Approach to Involvement of Background Knowledge for Question Answering + BorisGalitsky + DmitryIlvovsky + 373–381 + We introduce a concept of a virtual discourse tree to improve question answering (Q/A) recall for complex, multi-sentence questions. Augmenting the discourse tree of an answer with tree fragments obtained from text corpora playing the role of ontology, we obtain on the fly a canonical discourse representation of this answer that is independent of the thought structure of a given author. This mechanism is critical for finding an answer that is not only relevant in terms of questions entities but also in terms of inter-relations between these entities in an answer and its style. We evaluate the Q/A system enabled with virtual discourse trees and observe a substantial increase of performance answering complex questions such as Yahoo! Answers and www.2carpros.com. + R19-1044 + 10.26615/978-954-452-056-4_044 + + + On a Chatbot Providing Virtual Dialogues + BorisGalitsky + DmitryIlvovsky + ElizavetaGoncharova + 382–387 + We present a chatbot that delivers content in the form of virtual dialogues automatically produced from the plain texts that are extracted and selected from the documents. This virtual dialogue content is provided in the form of answers derived from the found and selected documents split into fragments, and questions that are automatically generated for these answers based on the initial text. + R19-1045 + 10.26615/978-954-452-056-4_045 + + + Assessing socioeconomic status of <fixed-case>T</fixed-case>witter users: A survey + DhouhaGHAZOUANI + LuigiLANCIERI + HabibOUNELLI + ChakerJEBARI + 388–398 + Every day, the emotion and opinion of different people across the world are reflected in the form of short messages using microblogging platforms. Despite the existence of enormous potential introduced by this data source, the Twitter community is still ambiguous and is not fully explored yet. While there are a huge number of studies examining the possibilities of inferring gender and age, there exist hardly researches on socioeconomic status (SES) inference of Twitter users. As socioeconomic status is essential to treating diverse questions linked to human behavior in several fields (sociology, demography, public health, etc.), we conducted a comprehensive literature review of SES studies, inference methods, and metrics. With reference to the research on literature’s results, we came to outline the most critical challenges for researchers. To the best of our knowledge, this paper is the first review that introduces the different aspects of SES inference. Indeed, this article provides the benefits for practitioners who aim to process and explore Twitter SES inference. + R19-1046 + 10.26615/978-954-452-056-4_046 + + + Divide and Extract – Disentangling Clause Splitting and Proposition Extraction + DarinaGold + TorstenZesch + 399–408 + Proposition extraction from sentences is an important task for information extraction systems Evaluation of such systems usually conflates two aspects: splitting complex sentences into clauses and the extraction of propositions. It is thus difficult to independently determine the quality of the proposition extraction step. We create a manually annotated proposition dataset from sentences taken from restaurant reviews that distinguishes between clauses that need to be split and those that do not. The resulting proposition evaluation dataset allows us to independently compare the performance of proposition extraction systems on simple and complex clauses. Although performance drastically drops on more complex sentences, we show that the same systems perform best on both simple and complex clauses. Furthermore, we show that specific kinds of subordinate clauses pose difficulties to most systems. + R19-1047 + 10.26615/978-954-452-056-4_047 + + + Sparse Coding in Authorship Attribution for Polish Tweets + PiotrGrzybowski + EwaJuralewicz + MaciejPiasecki + 409–417 + The study explores application of a simple Convolutional Neural Network for the problem of authorship attribution of tweets written in Polish. In our solution we use two-step compression of tweets using Byte Pair Encoding algorithm and vectorisation as an input to the distributional model generated for the large corpus of Polish tweets by word2vec algorithm. Our method achieves results comparable to the state-of-the-art approaches for the similar task on English tweets and expresses a very good performance in the classification of Polish tweets. We tested the proposed method in relation to the number of authors and tweets per author. We also juxtaposed results for authors with different topic backgrounds against each other. + R19-1048 + 10.26615/978-954-452-056-4_048 + + + Automatic Question Answering for Medical <fixed-case>MCQ</fixed-case>s: Can It go Further than Information Retrieval? + Le AnHa + VictoriaYaneva + 418–422 + We present a novel approach to automatic question answering that does not depend on the performance of an information retrieval (IR) system and does not require that the training data come from the same source as the questions. We evaluate the system performance on a challenging set of university-level medical science multiple-choice questions. Best performance is achieved when combining a neural approach with an IR approach, both of which work independently. Unlike previous approaches, the system achieves statistically significant improvement over the random guess baseline even for questions that are labeled as challenging based on the performance of baseline solvers. + R19-1049 + 10.26615/978-954-452-056-4_049 + + + Self-Knowledge Distillation in Natural Language Processing + SangchulHahn + HeeyoulChoi + 423–430 + Since deep learning became a key player in natural language processing (NLP), many deep learning models have been showing remarkable performances in a variety of NLP tasks. Such high performance can be explained by efficient knowledge representation of deep learning models. Knowledge distillation from pretrained deep networks suggests that we can use more information from the soft target probability to train other neural networks. In this paper, we propose a self-knowledge distillation method, based on the soft target probabilities of the training model itself, where multimode information is distilled from the word embedding space right below the softmax layer. Due to the time complexity, our method approximates the soft target probabilities. In experiments, we applied the proposed method to two different and fundamental NLP tasks: language model and neural machine translation. The experiment results show that our proposed method improves performance on the tasks. + R19-1050 + 10.26615/978-954-452-056-4_050 + + + From the Paft to the Fiiture: a Fully Automatic <fixed-case>NMT</fixed-case> and Word Embeddings Method for <fixed-case>OCR</fixed-case> Post-Correction + MikaHämäläinen + SimonHengchen + 431–436 + A great deal of historical corpora suffer from errors introduced by the OCR (optical character recognition) methods used in the digitization process. Correcting these errors manually is a time-consuming process and a great part of the automatic approaches have been relying on rules or supervised machine learning. We present a fully automatic unsupervised way of extracting parallel data for training a character-based sequence-to-sequence NMT (neural machine translation) model to conduct OCR error correction. + R19-1051 + 10.26615/978-954-452-056-4_051 + + + Investigating Terminology Translation in Statistical and Neural Machine Translation: A Case Study on <fixed-case>E</fixed-case>nglish-to-<fixed-case>H</fixed-case>indi and <fixed-case>H</fixed-case>indi-to-<fixed-case>E</fixed-case>nglish + RejwanulHaque + MdHasanuzzaman + AndyWay + 437–446 + Terminology translation plays a critical role in domain-specific machine translation (MT). In this paper, we conduct a comparative qualitative evaluation on terminology translation in phrase-based statistical MT (PB-SMT) and neural MT (NMT) in two translation directions: English-to-Hindi and Hindi-to-English. For this, we select a test set from a legal domain corpus and create a gold standard for evaluating terminology translation in MT. We also propose an error typology taking the terminology translation errors into consideration. We evaluate the MT systems’ performance on terminology translation, and demonstrate our findings, unraveling strengths, weaknesses, and similarities of PB-SMT and NMT in the area of term translation. + R19-1052 + 10.26615/978-954-452-056-4_052 + + + Beyond <fixed-case>E</fixed-case>nglish-Only Reading Comprehension: Experiments in Zero-shot Multilingual Transfer for <fixed-case>B</fixed-case>ulgarian + MomchilHardalov + IvanKoychev + PreslavNakov + 447–459 + Recently, reading comprehension models achieved near-human performance on large-scale datasets such as SQuAD, CoQA, MS Macro, RACE, etc. This is largely due to the release of pre-trained contextualized representations such as BERT and ELMo, which can be fine-tuned for the target task. Despite those advances and the creation of more challenging datasets, most of the work is still done for English. Here, we study the effectiveness of multilingual BERT fine-tuned on large-scale English datasets for reading comprehension (e.g., for RACE), and we apply it to Bulgarian multiple-choice reading comprehension. We propose a new dataset containing 2,221 questions from matriculation exams for twelfth grade in various subjects —history, biology, geography and philosophy—, and 412 additional questions from online quizzes in history. While the quiz authors gave no relevant context, we incorporate knowledge from Wikipedia, retrieving documents matching the combination of question + each answer option. Moreover, we experiment with different indexing and pre-training strategies. The evaluation results show accuracy of 42.23%, which is well above the baseline of 24.89%. + R19-1053 + 10.26615/978-954-452-056-4_053 + + + Tweaks and Tricks for Word Embedding Disruptions + AmirHazem + NicolasHernandez + 460–464 + Word embeddings are established as very effective models used in several NLP applications. If they differ in their architecture and training process, they often exhibit similar properties and remain vector space models with continuously-valued dimensions describing the observed data. The complexity resides in the developed strategies for learning the values within each dimensional space. In this paper, we introduce the concept of disruption which we define as a side effect of the training process of embedding models. Disruptions are viewed as a set of embedding values that are more likely to be noise than effective descriptive features. We show that dealing with disruption phenomenon is of a great benefit to bottom-up sentence embedding representation. By contrasting several in-domain and pre-trained embedding models, we propose two simple but very effective tweaking techniques that yield strong empirical improvements on textual similarity task. + R19-1054 + 10.26615/978-954-452-056-4_054 + + + Meta-Embedding Sentence Representation for Textual Similarity + AmirHazem + NicolasHernandez + 465–473 + Word embedding models are now widely used in most NLP applications. Despite their effectiveness, there is no clear evidence about the choice of the most appropriate model. It often depends on the nature of the task and on the quality and size of the used data sets. This remains true for bottom-up sentence embedding models. However, no straightforward investigation has been conducted so far. In this paper, we propose a systematic study of the impact of the main word embedding models on sentence representation. By contrasting in-domain and pre-trained embedding models, we show under which conditions they can be jointly used for bottom-up sentence embeddings. Finally, we propose the first bottom-up meta-embedding representation at the sentence level for textual similarity. Significant improvements are observed in several tasks including question-to-question similarity, paraphrasing and next utterance ranking. + R19-1055 + 10.26615/978-954-452-056-4_055 + + + Emoji Powered Capsule Network to Detect Type and Target of Offensive Posts in Social Media + HansiHettiarachchi + TharinduRanasinghe + 474–480 + This paper describes a novel research approach to detect type and target of offensive posts in social media using a capsule network. The input to the network was character embeddings combined with emoji embeddings. The approach was evaluated on all three subtasks in Task 6 - SemEval 2019: OffensEval: Identifying and Categorizing Offensive Language in Social Media. The evaluation also showed that even though the capsule networks have not been used commonly in natural language processing tasks, they can outperform existing state of the art solutions for offensive language detection in social media. + R19-1056 + 10.26615/978-954-452-056-4_056 + + + <fixed-case>E</fixed-case>o<fixed-case>ANN</fixed-case>: Lexical Semantic Relation Classification Using an Ensemble of Artificial Neural Networks + RayeheHosseini Pour + MehrnoushShamsfard + 481–486 + Researchers use wordnets as a knowledge base in many natural language processing tasks and applications, such as question answering, textual entailment, discourse classification, and so forth. Lexico-semantic relations among words or concepts are important parts of knowledge encoded in wordnets. As the use of wordnets becomes extensively widespread, extending the existing ones gets more attention. Manually construction and extension of lexico-semantic relations for WordNets or knowledge graphs are very time-consuming. Using automatic relation extraction methods can speed up this process. In this study, we exploit an ensemble of lstm and convolutional neural networks in a supervised manner to capture lexico-semantic relations which can either be used directly in NLP applications or compose the edges of wordnets. The whole procedure of learning vector space representation of relations is language independent. We used Princeton WordNet 3.1, FarsNet 3.0 (the Persian wordnet), Root09 and EVALution as golden standards to evaluate the predictive performance of our model and the results are comparable on the two languages. Empirical results demonstrate that our model outperforms the state of the art models. + R19-1057 + 10.26615/978-954-452-056-4_057 + + + Opinions Summarization: Aspect Similarity Recognition Relaxes The Constraint of Predefined Aspects + NguyenHuy Tien + LeTung Thanh + NguyenMinh Le + 487–496 + Recently research in opinions summarization focuses on rating expressions by aspects and/or sentiments they carry. To extract aspects of an expression, most studies require a predefined list of aspects or at least the number of aspects. Instead of extracting aspects, we rate expressions by aspect similarity recognition (ASR), which evaluates whether two expressions share at least one aspect. This subtask relaxes the limitation of predefining aspects and makes our opinions summarization applicable in domain adaptation. For the ASR subtask, we propose an attention-cell LSTM model, which integrates attention signals into the LSTM gates. According to the experimental results, the attention-cell LSTM works efficiently for learning latent aspects between two sentences in both settings of in-domain and cross-domain. In addition, the proposed extractive summarization method using ASR shows significant improvements over baselines on the Opinosis corpus. + R19-1058 + 10.26615/978-954-452-056-4_058 + + + Discourse-Aware Hierarchical Attention Network for Extractive Single-Document Summarization + TatsuyaIshigaki + HidetakaKamigaito + HiroyaTakamura + ManabuOkumura + 497–506 + Discourse relations between sentences are often represented as a tree, and the tree structure provides important information for summarizers to create a short and coherent summary. However, current neural network-based summarizers treat the source document as just a sequence of sentences and ignore the tree-like discourse structure inherent in the document. To incorporate the information of a discourse tree structure into the neural network-based summarizers, we propose a discourse-aware neural extractive summarizer which can explicitly take into account the discourse dependency tree structure of the source document. Our discourse-aware summarizer can jointly learn the discourse structure and the salience score of a sentence by using novel hierarchical attention modules, which can be trained on automatically parsed discourse dependency trees. Experimental results showed that our model achieved competitive or better performances against state-of-the-art models in terms of ROUGE scores on the DailyMail dataset. We further conducted manual evaluations. The results showed that our approach also gained the coherence of the output summaries. + R19-1059 + 10.26615/978-954-452-056-4_059 + + + Semi-Supervised Induction of <fixed-case>POS</fixed-case>-Tag Lexicons with Tree Models + MaciejJanicki + 507–515 + We approach the problem of POS tagging of morphologically rich languages in a setting where only a small amount of labeled training data is available. We show that a bigram HMM tagger benefits from re-training on a larger untagged text using Baum-Welch estimation. Most importantly, this estimation can be significantly improved by pre-guessing tags for OOV words based on morphological criteria. We consider two models for this task: a character-based recurrent neural network, which guesses the tag from the string form of the word, and a recently proposed graph-based model of morphological transformations. In the latter, the unknown POS tags can be modeled as latent variables in a way very similar to Hidden Markov Tree models and an analogue of the Forward-Backward algorithm can be formulated, which enables us to compute expected values over unknown taggings. We evaluate both the quality of the induced tag lexicon and its impact on the HMM’s tagging accuracy. In both tasks, the graph-based morphology model performs significantly better than the RNN predictor. This confirms the intuition that morphologically related words provide useful information about an unknown word’s POS tag. + R19-1060 + 10.26615/978-954-452-056-4_060 + + + Word Sense Disambiguation based on Constrained Random Walks in Linked Semantic Networks + ArkadiuszJanz + MaciejPiasecki + 516–525 + Word Sense Disambiguation remains a challenging NLP task. Due to the lack of annotated training data, especially for rare senses, the supervised approaches are usually designed for specific subdomains limited to a narrow subset of identified senses. Recent advances in this area have shown that knowledge-based approaches are more scalable and obtain more promising results in all-words WSD scenarios. In this work we present a faster WSD algorithm based on the Monte Carlo approximation of sense probabilities given a context using constrained random walks over linked semantic networks. We show that the local semantic relatedness is mostly sufficient to successfully identify correct senses when an extensive knowledge base and a proper weighting scheme are used. The proposed methods are evaluated on English (SenseEval, SemEval) and Polish (Składnica, KPWr) datasets. + R19-1061 + 10.26615/978-954-452-056-4_061 + + + Classification of Micro-Texts Using Sub-Word Embeddings + MihirJoshi + NurZincir-Heywood + 526–533 + Extracting features and writing styles from short text messages is always a challenge. Short messages, like tweets, do not have enough data to perform statistical authorship attribution. Besides, the vocabulary used in these texts is sometimes improvised or misspelled. Therefore, in this paper, we propose combining four feature extraction techniques namely character n-grams, word n-grams, Flexible Patterns and a new sub-word embedding using the skip-gram model. Our system uses a Multi-Layer Perceptron to utilize these features from tweets to analyze short text messages. This proposed system achieves 85% accuracy, which is a considerable improvement over previous systems. + R19-1062 + 10.26615/978-954-452-056-4_062 + + + Using Syntax to Resolve <fixed-case>NPE</fixed-case> in <fixed-case>E</fixed-case>nglish + PayalKhullar + AllenAntony + ManishShrivastava + 534–540 + This paper describes a novel, syntax-based system for automatic detection and resolution of Noun Phrase Ellipsis (NPE) in English. The system takes in free input English text, detects the site of nominal elision, and if present, selects potential antecedent candidates. The rules are built using the syntactic information on ellipsis and its antecedent discussed in previous theoretical linguistics literature on NPE. Additionally, we prepare a curated dataset of 337 sentences from well-known, reliable sources, containing positive and negative samples of NPE. We split this dataset into two parts, and use one part to refine our rules and the other to test the performance of our final system. We get an F1-score of 76.47% for detection and 70.27% for NPE resolution on the testset. To the best of our knowledge, ours is the first system that detects and resolves NPE in English. The curated dataset used for this task, albeit small, covers a wide variety of NPE cases and will be made public for future work. + R19-1063 + 10.26615/978-954-452-056-4_063 + + + Is Similarity Visually Grounded? Computational Model of Similarity for the <fixed-case>E</fixed-case>stonian language + ClaudiaKittask + EduardBarbu + 541–549 + Researchers in Computational Linguistics build models of similarity and test them against human judgments. Although there are many empirical studies of the computational models of similarity for the English language, the similarity for other languages is less explored. In this study we are chiefly interested in two aspects. In the first place we want to know how much of the human similarity is grounded in the visual perception. To answer this question two neural computer vision models are used and their correlation with the human derived similarity scores is computed. In the second place we investigate if language influences the similarity computation. To this purpose diverse computational models trained on Estonian resources are evaluated against human judgments + R19-1064 + 10.26615/978-954-452-056-4_064 + + + Language-Agnostic <fixed-case>T</fixed-case>witter-Bot Detection + JürgenKnauth + 550–558 + In this paper we address the problem of detecting Twitter bots. We analyze a dataset of 8385 Twitter accounts and their tweets consisting of both humans and different kinds of bots. We use this data to train machine learning classifiers that distinguish between real and bot accounts. We identify features that are easy to extract while still providing good results. We analyze different feature groups based on account specific, tweet specific and behavioral specific features and measure their performance compared to other state of the art bot detection methods. For easy future portability of our work we focus on language-agnostic features. With AdaBoost, the best performing classifier, we achieve an accuracy of 0.988 and an AUC of 0.995. As the creation of good training data in machine learning is often difficult - especially in the domain of Twitter bot detection - we additionally analyze to what extent smaller amounts of training data lead to useful results by reviewing cross-validated learning curves. Our results indicate that using few but expressive features already has a good practical benefit for bot detection, especially if only a small amount of training data is available. + R19-1065 + 10.26615/978-954-452-056-4_065 + + + Multi-level analysis and recognition of the text sentiment on the example of consumer opinions + JanKocoń + MonikaZaśko-Zielińska + PiotrMiłkowski + 559–567 + In this article, we present a novel multi-domain dataset of Polish text reviews, annotated with sentiment on different levels: sentences and the whole documents. The annotation was made by linguists in a 2+1 scheme (with inter-annotator agreement analysis). We present a preliminary approach to the classification of labelled data using logistic regression, bidirectional long short-term memory recurrent neural networks (BiLSTM) and bidirectional encoder representations from transformers (BERT). + R19-1066 + 10.26615/978-954-452-056-4_066 + + + A Qualitative Evaluation Framework for Paraphrase Identification + VenelinKovatchev + M. AntoniaMarti + MariaSalamo + JavierBeltran + 568–577 + In this paper, we present a new approach for the evaluation, error analysis, and interpretation of supervised and unsupervised Paraphrase Identification (PI) systems. Our evaluation framework makes use of a PI corpus annotated with linguistic phenomena to provide a better understanding and interpretation of the performance of various PI systems. Our approach allows for a qualitative evaluation and comparison of the PI models using human interpretable categories. It does not require modification of the training objective of the systems and does not place additional burden on the developers. We replicate several popular supervised and unsupervised PI systems. Using our evaluation framework we show that: 1) Each system performs differently with respect to a set of linguistic phenomena and makes qualitatively different kinds of errors; 2) Some linguistic phenomena are more challenging than others across all systems. + R19-1067 + 10.26615/978-954-452-056-4_067 + + + Study on Unsupervised Statistical Machine Translation for Backtranslation + AnushKumar + Nihal V.Nayak + AdityaChandra + Mydhili K.Nair + 578–582 + Machine Translation systems have drastically improved over the years for several language pairs. Monolingual data is often used to generate synthetic sentences to augment the training data which has shown to improve the performance of machine translation models. In our paper, we make use of an Unsupervised Statistical Machine Translation (USMT) to generate synthetic sentences. Our study compares the performance improvements in Neural Machine Translation model when using synthetic sentences from supervised and unsupervised Machine Translation models. Our approach of using USMT for backtranslation shows promise in low resource conditions and achieves an improvement of 3.2 BLEU score over the Neural Machine Translation model. + R19-1068 + 10.26615/978-954-452-056-4_068 + + + Towards Functionally Similar Corpus Resources for Translation + MariaKunilovskaya + SergeSharoff + 583–592 + The paper describes a computational approach to produce functionally comparable monolingual corpus resources for translation studies and contrastive analysis. We exploit a text-external approach, based on a set of Functional Text Dimensions to model text functions, so that each text can be represented as a vector in a multidimensional space of text functions. These vectors can be used to find reasonably homogeneous subsets of functionally similar texts across different corpora. Our models for predicting text functions are based on recurrent neural networks and traditional feature-based machine learning approaches. In addition to using the categories of the British National Corpus as our test case, we investigated the functional comparability of the English parts from the two parallel corpora: CroCo (English-German) and RusLTC (English-Russian) and applied our models to define functionally similar clusters in them. Our results show that the Functional Text Dimensions provide a useful description for text categories, while allowing a more flexible representation for texts with hybrid functions. + R19-1069 + 10.26615/978-954-452-056-4_069 + + + Question Similarity in Community Question Answering: A Systematic Exploration of Preprocessing Methods and Models + FlorianKunneman + Thiago CastroFerreira + EmielKrahmer + Antalvan den Bosch + 593–601 + Community Question Answering forums are popular among Internet users, and a basic problem they encounter is trying to find out if their question has already been posed before. To address this issue, NLP researchers have developed methods to automatically detect question-similarity, which was one of the shared tasks in SemEval. The best performing systems for this task made use of Syntactic Tree Kernels or the SoftCosine metric. However, it remains unclear why these methods seem to work, whether their performance can be improved by better preprocessing methods and what kinds of errors they (and other methods) make. In this paper, we therefore systematically combine and compare these two approaches with the more traditional BM25 and translation-based models. Moreover, we analyze the impact of preprocessing steps (lowercasing, suppression of punctuation and stop words removal) and word meaning similarity based on different distributions (word translation probability, Word2Vec, fastText and ELMo) on the performance of the task. We conduct an error analysis to gain insight into the differences in performance between the system set-ups. The implementation is made publicly available from https://github.com/fkunneman/DiscoSumo/tree/master/ranlp. + R19-1070 + 10.26615/978-954-452-056-4_070 + + + A Classification-Based Approach to Cognate Detection Combining Orthographic and Semantic Similarity Information + SofieLabat + ElsLefever + 602–610 + This paper presents proof-of-concept experiments for combining orthographic and semantic information to distinguish cognates from non-cognates. To this end, a context-independent gold standard is developed by manually labelling English-Dutch pairs of cognates and false friends in bilingual term lists. These annotated cognate pairs are then used to train and evaluate a supervised binary classification system for the automatic detection of cognates. Two types of information sources are incorporated in the classifier: fifteen string similarity metrics capture form similarity between source and target words, while word embeddings model semantic similarity between the words. The experimental results show that even though the system already achieves good results by only incorporating orthographic information, the performance further improves by including semantic information in the form of embeddings. + R19-1071 + 10.26615/978-954-452-056-4_071 + + + Resolving Pronouns for a Resource-Poor Language, <fixed-case>M</fixed-case>alayalam Using Resource-Rich Language, <fixed-case>T</fixed-case>amil. + SobhaLalitha Devi + 611–618 + In this paper we give in detail how a resource rich language can be used for resolving pronouns for a less resource language. The source language, which is resource rich language in this study, is Tamil and the resource poor language is Malayalam, both belonging to the same language family, Dravidian. The Pronominal resolution developed for Tamil uses CRFs. Our approach is to leverage the Tamil language model to test Malayalam data and the processing required for Malayalam data is detailed. The similarity at the syntactic level between the languages is exploited in identifying the features for developing the Tamil language model. The word form or the lexical item is not considered as a feature for training the CRFs. Evaluation on Malayalam Wikipedia data shows that our approach is correct and the results, though not as good as Tamil, but comparable. + R19-1072 + 10.26615/978-954-452-056-4_072 + + + Semantic Role Labeling with Pretrained Language Models for Known and Unknown Predicates + DaniilLarionov + ArtemShelmanov + ElenaChistova + IvanSmirnov + 619–628 + We build the first full pipeline for semantic role labelling of Russian texts. The pipeline implements predicate identification, argument extraction, argument classification (labeling), and global scoring via integer linear programming. We train supervised neural network models for argument classification using Russian semantically annotated corpus – FrameBank. However, we note that this resource provides annotations only to a very limited set of predicates. We combat the problem of annotation scarcity by introducing two models that rely on different sets of features: one for “known” predicates that are present in the training set and one for “unknown” predicates that are not. We show that the model for “unknown” predicates can alleviate the lack of annotation by using pretrained embeddings. We perform experiments with various types of embeddings including the ones generated by deep pretrained language models: word2vec, FastText, ELMo, BERT, and show that embeddings generated by deep pretrained language models are superior to classical shallow embeddings for argument classification of both “known” and “unknown” predicates. + R19-1073 + 10.26615/978-954-452-056-4_073 + + + Structural Approach to Enhancing <fixed-case>W</fixed-case>ord<fixed-case>N</fixed-case>et with Conceptual Frame Semantics + SvetlozaraLeseva + IvelinaStoyanova + 629–637 + This paper outlines procedures for enhancing WordNet with conceptual information from FrameNet. The mapping of the two resources is non-trivial. We define a number of techniques for the validation of the consistency of the mapping and the extension of its coverage which make use of the structure of both resources and the systematic relations between synsets in WordNet and between frames in FrameNet, as well as between synsets and frames). We present a case study on causativity, a relation which provides enhancement complementary to the one using hierarchical relations, by means of linking in a systematic way large parts of the lexicon. We show how consistency checks and denser relations may be implemented on the basis of this relation. We, then, propose new frames based on causative-inchoative correspondences and in conclusion touch on the possibilities for defining new frames based on the types of specialisation that takes place from parent to child synset. + R19-1074 + 10.26615/978-954-452-056-4_074 + + + Compositional Hyponymy with Positive Operators + MarthaLewis + 638–647 + Language is used to describe concepts, and many of these concepts are hierarchical. Moreover, this hierarchy should be compatible with forming phrases and sentences. We use linear-algebraic methods that allow us to encode words as collections of vectors. The representations we use have an ordering, related to subspace inclusion, which we interpret as modelling hierarchical information. The word representations built can be understood within a compositional distributional semantic framework, providing methods for composing words to form phrase and sentence level representations. We show that the resulting representations give competitive results on both word-level hyponymy and sentence-level entailment datasets. + R19-1075 + 10.26615/978-954-452-056-4_075 + + + The Impact of Semantic Linguistic Features in Relation Extraction: A Logical Relational Learning Approach + RinaldoLima + BernardEspinasse + FredericoFreitas + 648–654 + Relation Extraction (RE) consists in detecting and classifying semantic relations between entities in a sentence. The vast majority of the state-of-the-art RE systems relies on morphosyntactic features and supervised machine learning algorithms. This paper tries to answer important questions concerning both the impact of semantic based features, and the integration of external linguistic knowledge resources on RE performance. For that, a RE system based on a logical and relational learning algorithm was used and evaluated on three reference datasets from two distinct domains. The yielded results confirm that the classifiers induced using the proposed richer feature set outperformed the classifiers built with morphosyntactic features in average 4% (F1-measure). + R19-1076 + 10.26615/978-954-452-056-4_076 + + + Detecting Anorexia in <fixed-case>S</fixed-case>panish Tweets + PilarLópez Úbeda + Flor MiriamPlaza del Arco + Manuel CarlosDíaz Galiano + L. AlfonsoUrena Lopez + MaiteMartin + 655–663 + Mental health is one of the main concerns of today’s society. Early detection of symptoms can greatly help people with mental disorders. People are using social networks more and more to express emotions, sentiments and mental states. Thus, the treatment of this information using NLP technologies can be applied to the automatic detection of mental problems such as eating disorders. However, the first step to solving the problem should be to provide a corpus in order to evaluate our systems. In this paper, we specifically focus on detecting anorexia messages on Twitter. Firstly, we have generated a new corpus of tweets extracted from different accounts including anorexia and non-anorexia messages in Spanish. The corpus is called SAD: Spanish Anorexia Detection corpus. In order to validate the effectiveness of the SAD corpus, we also propose several machine learning approaches for automatically detecting anorexia symptoms in the corpus. The good results obtained show that the application of textual classification methods is a promising option for developing this kind of system demonstrating that these tools could be used by professionals to help in the early detection of mental problems. + R19-1077 + 10.26615/978-954-452-056-4_077 + + + A type-theoretical reduction of morphological, syntactic and semantic compositionality to a single level of description + ErkkiLuuk + 664–673 + The paper presents NLC, a new formalism for modeling natural language (NL) compositionality. NLC is a functional type system (i.e. one based on mathematical functions and their types). Its main features include a close correspondence with NL and an integrated modeling of morphological, syntactic and semantic compositionality. The integration is effected with a subclass of compound types (types which are syntactic compounds of multiple types or their terms), while the correspondence is sought with function types and polymorphism. The paper also presents an implementation of NLC in Coq. The implementation formalizes a diverse fragment of NL, with NLC expressions type checking and failing to type check in exactly the same ways that NL expressions pass and fail their acceptability tests. Among other things, this demonstrates the possibility of reducing morphological, syntactic and semantic compositionality to a single level of description. The level is tentatively identified with semantic compositionality — an interpretation which, besides being supported by results from language processing, has interesting implications on NL structure and modeling. + R19-1078 + 10.26615/978-954-452-056-4_078 + + + v-trel: Vocabulary Trainer for Tracing Word Relations - An Implicit Crowdsourcing Approach + VerenaLyding + ChristosRodosthenous + FedericoSangati + Umairul Hassan + LionelNicolas + AlexanderKönig + JolitaHorbacauskiene + AnisiaKatinskaia + 674–683 + In this paper, we present our work on developing a vocabulary trainer that uses exercises generated from language resources such as ConceptNet and crowdsources the responses of the learners to enrich the language resource. We performed an empirical evaluation of our approach with 60 non-native speakers over two days, which shows that new entries to expand Concept-Net can efficiently be gathered through vocabulary exercises on word relations. We also report on the feedback gathered from the users and an expert from language teaching, and discuss the potential of the vocabulary trainer application from the user and language learner perspective. The feedback suggests that v-trel has educational potential, while in its current state some shortcomings could be identified. + R19-1079 + 10.26615/978-954-452-056-4_079 + + + Jointly Learning Author and Annotated Character N-gram Embeddings: A Case Study in Literary Text + SurajMaharjan + DeepthiMave + PrashaShrestha + ManuelMontes + Fabio A.González + ThamarSolorio + 684–692 + An author’s way of presenting a story through his/her writing style has a great impact on whether the story will be liked by readers or not. In this paper, we learn representations for authors of literary texts together with representations for character n-grams annotated with their functional roles. We train a neural character n-gram based language model using an external corpus of literary texts and transfer learned representations for use in downstream tasks. We show that augmenting the knowledge from external works of authors produces results competitive with other style-based methods for book likability prediction, genre classification, and authorship attribution. + R19-1080 + 10.26615/978-954-452-056-4_080 + + + Generating Challenge Datasets for Task-Oriented Conversational Agents through Self-Play + SourabhMajumdar + Serra SinemTekiroglu + MarcoGuerini + 693–702 + End-to-end neural approaches are becoming increasingly common in conversational scenarios due to their promising performances when provided with sufficient amount of data. In this paper, we present a novel methodology to address the interpretability of neural approaches in such scenarios by creating challenge datasets using dialogue self-play over multiple tasks/intents. Dialogue self-play allows generating large amount of synthetic data; by taking advantage of the complete control over the generation process, we show how neural approaches can be evaluated in terms of unseen dialogue patterns. We propose several out-of-pattern test cases each of which introduces a natural and unexpected user utterance phenomenon. As a proof of concept, we built a single and a multiple memory network, and show that these two architectures have diverse performances depending on the peculiar dialogue patterns. + R19-1081 + 10.26615/978-954-452-056-4_081 + + + Sentiment Polarity Detection in <fixed-case>A</fixed-case>zerbaijani Social News Articles + SevdaMammadli + ShamsaddinHuseynov + HuseynAlkaramov + UlviyyaJafarli + UmidSuleymanov + SamirRustamov + 703–710 + Text classification field of natural language processing has been experiencing remarkable growth in recent years. Especially, sentiment analysis has received a considerable attention from both industry and research community. However, only a few research examples exist for Azerbaijani language. The main objective of this research is to apply various machine learning algorithms for determining the sentiment of news articles in Azerbaijani language. Approximately, 30.000 social news articles have been collected from online news sites and labeled manually as negative or positive according to their sentiment categories. Initially, text preprocessing was implemented to data in order to eliminate the noise. Secondly, to convert text to a more machine-readable form, BOW (bag of words) model has been applied. More specifically, two methodologies of BOW model, which are tf-idf and frequency based model have been used as vectorization methods. Additionally, SVM, Random Forest, and Naive Bayes algorithms have been applied as the classification algorithms, and their combinations with two vectorization approaches have been tested and analyzed. Experimental results indicate that SVM outperforms other classification algorithms. + R19-1082 + 10.26615/978-954-452-056-4_082 + + + <fixed-case>I</fixed-case>nforex — a Collaborative Systemfor Text Corpora Annotation and Analysis Goes Open + MichałMarcińczuk + MarcinOleksy + 711–719 + In the paper we present the latest changes introduce to Inforex — a web-based system for qualitative and collaborative text corpora annotation and analysis. One of the most important news is the release of source codes. Now the system is available on the GitHub repository (https://github.com/CLARIN-PL/Inforex) as an open source project. The system can be easily setup and run in a Docker container what simplifies the installation process. The major improvements include: semi-automatic text annotation, multilingual text preprocessing using CLARIN-PL web services, morphological tagging of XML documents, improved editor for annotation attribute, batch annotation attribute editor, morphological disambiguation, extended word sense annotation. This paper contains a brief description of the mentioned improvements. We also present two use cases in which various Inforex features were used and tested in real-life projects. + R19-1083 + 10.26615/978-954-452-056-4_083 + + + Semantic Language Model for <fixed-case>T</fixed-case>unisian Dialect + AbirMASMOUDI + RimLaatar + Mariemellouze + lamiahadrich belguith + 720–729 + In this paper, we describe the process of creating a statistical Language Model (LM) for the Tunisian Dialect. Indeed, this work is part of the realization of Automatic Speech Recognition (ASR) system for the Tunisian Railway Transport Network. Since our eld of work has been limited, there are several words with similar behaviors (semantic for example) but they do not have the same appearance probability; their class groupings will therefore be possible. For these reasons, we propose to build an n-class LM that is based mainly on the integration of purely semantic data. Indeed, each class represents an abstraction of similar labels. In order to improve the sequence labeling task, we proposed to use a discriminative algorithm based on the Conditional Random Field (CRF) model. To better judge our choice of creating an n-class word model, we compared the created model with the 3-gram type model on the same test corpus of evaluation. Additionally, to assess the impact of using the CRF model to perform the semantic labelling task in order to construct semantic classes, we compared the n-class created model with using the CRF in the semantic labelling task and the n- class model without using the CRF in the semantic labelling task. The drawn comparison of the predictive power of the n-class model obtained by applying the CRF model in the semantic labelling is that it is better than the other two models presenting the highest value of its perplexity. + R19-1084 + 10.26615/978-954-452-056-4_084 + + + Automatic diacritization of <fixed-case>T</fixed-case>unisian dialect text using Recurrent Neural Network + AbirMasmoudi + MariemEllouze + LamiaHadrich belguith + 730–739 + The absence of diacritical marks in the Arabic texts generally leads to morphological, syntactic and semantic ambiguities. This can be more blatant when one deals with under-resourced languages, such as the Tunisian dialect, which suffers from unavailability of basic tools and linguistic resources, like sufficient amount of corpora, multilingual dictionaries, morphological and syntactic analyzers. Thus, this language processing faces greater challenges due to the lack of these resources. The automatic diacritization of MSA text is one of the various complex problems that can be solved by deep neural networks today. Since the Tunisian dialect is an under-resourced language of MSA and as there are a lot of resemblance between both languages, we suggest to investigate a recurrent neural network (RNN) for this dialect diacritization problem. This model will be compared to our previous models models CRF and SMT (CITATION) based on the same dialect corpus. We can experimentally show that our model can achieve better outcomes (DER of 10.72%), as compared to the two models CRF (DER of 20.25%) and SMT (DER of 33.15%). + R19-1085 + 10.26615/978-954-452-056-4_085 + + + Comparing <fixed-case>MT</fixed-case> Approaches for Text Normalization + ClaudiaMatos Veliz + OrpheeDe Clercq + VeroniqueHoste + 740–749 + One of the main characteristics of social media data is the use of non-standard language. Since NLP tools have been trained on traditional text material their performance drops when applied to social media data. One way to overcome this is to first perform text normalization. In this work, we apply text normalization to noisy English and Dutch text coming from different social media genres: text messages, message board posts and tweets. We consider the normalization task as a Machine Translation problem and test the two leading paradigms: statistical and neural machine translation. For SMT we explore the added value of varying background corpora for training the language model. For NMT we have a look at data augmentation since the parallel datasets we are working with are limited in size. Our results reveal that when relying on SMT to perform the normalization it is beneficial to use a background corpus that is close to the genre you are normalizing. Regarding NMT, we find that the translations - or normalizations - coming out of this model are far from perfect and that for a low-resource language like Dutch adding additional training data works better than artificially augmenting the data. + R19-1086 + 10.26615/978-954-452-056-4_086 + + + Sentiment and Emotion Based Representations for Fake Reviews Detection + AlimuddinMelleng + AnnaJurek-Loughrey + DeepakP + 750–757 + Fake reviews are increasingly prevalent across the Internet. They can be unethical as well as harmful. They can affect businesses and mislead individual customers. As the opinions on the Web are increasingly used the detection of fake reviews has become more and more critical. In this study, we explore the effectiveness of sentiment and emotions based representations for the task of building machine learning models for fake review detection. We perform empirical studies over three real world datasets and demonstrate that improved data representation can be achieved by combining sentiment and emotion extraction methods, as well as by performing sentiment and emotion analysis on a part-by-part basis by segmenting the reviews. + R19-1087 + 10.26615/978-954-452-056-4_087 + + + Turning silver into gold: error-focused corpus reannotation with active learning + Pierre AndréMénard + AntoineMougeot + 758–767 + While high quality gold standard annotated corpora are crucial for most tasks in natural language processing, many annotated corpora published in recent years, created by annotators or tools, contains noisy annotations. These corpora can be viewed as more silver than gold standards, even if they are used in evaluation campaigns or to compare systems’ performances. As upgrading a silver corpus to gold level is still a challenge, we explore the application of active learning techniques to detect errors using four datasets designed for document classification and part-of-speech tagging. Our results show that the proposed method for the seeding step improves the chance of finding incorrect annotations by a factor of 2.73 when compared to random selection, a 14.71% increase from the baseline methods. Our query method provides an increase in the error detection precision on average by a factor of 1.78 against random selection, an increase of 61.82% compared to other query approaches. + R19-1088 + 10.26615/978-954-452-056-4_088 + + + Community Perspective on Replicability in Natural Language Processing + MargotMieskes + KarënFort + AurélieNévéol + CyrilGrouin + KevinCohen + 768–775 + With recent efforts in drawing attention to the task of replicating and/or reproducing results, for example in the context of COLING 2018 and various LREC workshops, the question arises how the NLP community views the topic of replicability in general. Using a survey, in which we involve members of the NLP community, we investigate how our community perceives this topic, its relevance and options for improvement. Based on over two hundred participants, the survey results confirm earlier observations, that successful reproducibility requires more than having access to code and data. Additionally, the results show that the topic has to be tackled from the authors’, reviewers’ and community’s side. + R19-1089 + 10.26615/978-954-452-056-4_089 + + + Unsupervised Data Augmentation for Less-Resourced Languages with no Standardized Spelling + AliceMillour + KarënFort + 776–784 + Building representative linguistic resources and NLP tools for non-standardized languages is challenging: when spelling is not determined by a norm, multiple written forms can be encountered for a given word, inducing a large proportion of out-of-vocabulary words. To embrace this diversity, we propose a methodology based on crowdsourced alternative spellings we use to extract rules applied to match OOV words with one of their spelling variants. This virtuous process enables the unsupervised augmentation of multi-variant lexicons without expert rule definition. We apply this multilingual methodology on Alsatian, a French regional language and provide an intrinsic evaluation of the correctness of the variants pairs, and an extrinsic evaluation on a downstream task. We show that in a low-resource scenario, 145 inital pairs can lead to the generation of 876 additional variant pairs, and a diminution of OOV words improving the part-of-speech tagging performance by 1 to 4%. + R19-1090 + 10.26615/978-954-452-056-4_090 + + + Neural Feature Extraction for Contextual Emotion Detection + ElhamMohammadi + HessamAmini + LeilaKosseim + 785–794 + This paper describes a new approach for the task of contextual emotion detection. The approach is based on a neural feature extractor, composed of a recurrent neural network with an attention mechanism, followed by a classifier, that can be neural or SVM-based. We evaluated the model with the dataset of the task 3 of SemEval 2019 (EmoContext), which includes short 3-turn conversations, tagged with 4 emotion classes. The best performing setup was achieved using ELMo word embeddings and POS tags as input, bidirectional GRU as hidden units, and an SVM as the final classifier. This configuration reached 69.93% in terms of micro-average F1 score on the main 3 emotion classes, a score that outperformed the baseline system by 11.25%. + R19-1091 + 10.26615/978-954-452-056-4_091 + + + Empirical Study of Diachronic Word Embeddings for Scarce Data + SyrielleMontariol + AlexandreAllauzen + 795–803 + Word meaning change can be inferred from drifts of time-varying word embeddings. However, temporal data may be too sparse to build robust word embeddings and to discriminate significant drifts from noise. In this paper, we compare three models to learn diachronic word embeddings on scarce data: incremental updating of a Skip-Gram from Kim et al. (2014), dynamic filtering from Bamler & Mandt (2017), and dynamic Bernoulli embeddings from Rudolph & Blei (2018). In particular, we study the performance of different initialisation schemes and emphasise what characteristics of each model are more suitable to data scarcity, relying on the distribution of detected drifts. Finally, we regularise the loss of these models to better adapt to scarce data. + R19-1092 + 10.26615/978-954-452-056-4_092 + + + A Fast and Accurate Partially Deterministic Morphological Analysis + HajimeMorita + TomoyaIwakura + 804–809 + This paper proposes a partially deterministic morphological analysis method for improved processing speed. Maximum matching is a fast deterministic method for morphological analysis. However, the method tends to decrease performance due to lack of consideration of contextual information. In order to use maximum matching safely, we propose the use of Context Independent Strings (CISs), which are strings that do not have ambiguity in terms of morphological analysis. Our method first identifies CISs in a sentence using maximum matching without contextual information, then analyzes the unprocessed part of the sentence using a bi-gram-based morphological analysis model. We evaluate the method on a Japanese morphological analysis task. The experimental results show a 30% reduction of running time while maintaining improved accuracy. + R19-1093 + 10.26615/978-954-452-056-4_093 + + + incom.py - A Toolbox for Calculating Linguistic Distances and Asymmetries between Related Languages + MariusMosbach + IrinaStenger + TaniaAvgustinova + DietrichKlakow + 810–818 + Languages may be differently distant from each other and their mutual intelligibility may be asymmetric. In this paper we introduce incom.py, a toolbox for calculating linguistic distances and asymmetries between related languages. incom.py allows linguist experts to quickly and easily perform statistical analyses and compare those with experimental results. We demonstrate the efficacy of incom.py in an incomprehension experiment on two Slavic languages: Bulgarian and Russian. Using incom.py we were able to validate three methods to measure linguistic distances and asymmetries: Levenshtein distance, word adaptation surprisal, and conditional entropy as predictors of success in a reading intercomprehension experiment. + R19-1094 + 10.26615/978-954-452-056-4_094 + + + A Holistic Natural Language Generation Framework for the Semantic Web + Axel-CyrilleNgonga Ngomo + DiegoMoussallem + LorenzBühmann + 819–828 + With the ever-growing generation of data for the Semantic Web comes an increasing demand for this data to be made available to non-semantic Web experts. One way of achieving this goal is to translate the languages of the Semantic Web into natural language. We present LD2NL, a framework that allows verbalizing the three key languages of the Semantic Web, i.e., RDF, OWL, and SPARQL. Our framework is based on a bottom-up approach to verbalization. We evaluated LD2NL in an open survey with 86 persons. Our results suggest that our framework can generate verbalizations that are close to natural languages and that can be easily understood by non-experts. Therewith, it enables non-domain experts to interpret Semantic Web data with more than 91% of the accuracy of domain experts. + R19-1095 + 10.26615/978-954-452-056-4_095 + + + Building a Comprehensive <fixed-case>R</fixed-case>omanian Knowledge Base for Drug Administration + BogdanNicula + MihaiDascalu + Maria-DorinelaSîrbu + ȘtefanTrăușan-Matu + AlexandruNuță + 829–836 + Information on drug administration is obtained traditionally from doctors and pharmacists, as well as leaflets which provide in most cases cumbersome and hard-to-follow details. Thus, the need for medical knowledge bases emerges to provide access to concrete and well-structured information which can play an important role in informing patients. This paper introduces a Romanian medical knowledge base focused on drug-drug interactions, on representing relevant drug information, and on symptom-disease relations. The knowledge base was created by extracting and transforming information using Natural Language Processing techniques from both structured and unstructured sources, together with manual annotations. The resulting Romanian ontologies are aligned with larger English medical ontologies. Our knowledge base supports queries regarding drugs (e.g., active ingredients, concentration, expiration date), drug-drug interaction, symptom-disease relations, as well as drug-symptom relations. + R19-1096 + 10.26615/978-954-452-056-4_096 + + + Summary Refinement through Denoising + NikolaNikolov + AlessandroCalmanovici + RichardHahnloser + 837–843 + We propose a simple method for post-processing the outputs of a text summarization system in order to refine its overall quality. Our approach is to train text-to-text rewriting models to correct information redundancy errors that may arise during summarization. We train on synthetically generated noisy summaries, testing three different types of noise that introduce out-of-context information within each summary. When applied on top of extractive and abstractive summarization baselines, our summary denoising models yield metric improvements while reducing redundancy. + R19-1097 + 10.26615/978-954-452-056-4_097 + + + Large-Scale Hierarchical Alignment for Data-driven Text Rewriting + NikolaNikolov + RichardHahnloser + 844–853 + We propose a simple unsupervised method for extracting pseudo-parallel monolingual sentence pairs from comparable corpora representative of two different text styles, such as news articles and scientific papers. Our approach does not require a seed parallel corpus, but instead relies solely on hierarchical search over pre-trained embeddings of documents and sentences. We demonstrate the effectiveness of our method through automatic and extrinsic evaluation on text simplification from the normal to the Simple Wikipedia. We show that pseudo-parallel sentences extracted with our method not only supplement existing parallel data, but can even lead to competitive performance on their own. + R19-1098 + 10.26615/978-954-452-056-4_098 + + + Dependency-Based Relative Positional Encoding for Transformer <fixed-case>NMT</fixed-case> + YutaroOmote + AkihiroTamura + TakashiNinomiya + 854–861 + This paper proposes a new Transformer neural machine translation model that incorporates syntactic distances between two source words into the relative position representations of the self-attention mechanism. In particular, the proposed model encodes pair-wise relative depths on a source dependency tree, which are differences between the depths of the two source words, in the encoder’s self-attention. The experiments show that our proposed model achieves 0.5 point gain in BLEU on the Asian Scientific Paper Excerpt Corpus Japanese-to-English translation task. + R19-1099 + 10.26615/978-954-452-056-4_099 + + + From Image to Text in Sentiment Analysis via Regression and Deep Learning + DanielaOnita + Liviu P.Dinu + AdrianaBirlutiu + 862–868 + Images and text represent types of content which are used together for conveying user emotions in online social networks. These contents are usually associated with a sentiment category. In this paper, we investigate an approach for mapping images to text for three types of sentiment categories: positive, neutral and negative. The mapping from images to text is performed using a Kernel Ridge Regression model. We considered two types of image features: i) RGB pixel-values features, and ii) features extracted with a deep learning approach. The experimental evaluation was performed on a Twitter data set containing both text and images and the sentiment associated with these. The experimental results show a difference in performance for different sentiment categories, in particular the mapping that we propose performs better for the positive sentiment category in comparison with the neutral and negative ones. Furthermore, the experimental results show that the more complex deep learning features perform better than the RGB pixel-value features for all sentiment categories and for larger training sets. + R19-1100 + 10.26615/978-954-452-056-4_100 + + + Building a Morphological Analyser for Laz + EsraOnal + FrancisTyers + 869–877 + This study is an attempt to contribute to documentation and revitalization efforts of endangered Laz language, a member of South Caucasian language family mainly spoken on northeastern coastline of Turkey. It constitutes the first steps to create a general computational model for word form recognition and production for Laz by building a rule-based morphological analyser using Helsinki Finite-State Toolkit (HFST). The evaluation results show that the analyser has a 64.9% coverage over a corpus collected for this study with 111,365 tokens. We have also performed an error analysis on randomly selected 100 tokens from the corpus which are not covered by the analyser, and these results show that the errors mostly result from Turkish words in the corpus and missing stems in our lexicon. + R19-1101 + 10.26615/978-954-452-056-4_101 + + + Term Based Semantic Clusters for Very Short Text Classification + JasperPaalman + ShantanuMullick + KalliopiZervanou + YingqianZhang + 878–887 + Very short texts, such as tweets and invoices, present challenges in classification. Although term occurrences are strong indicators of content, in very short texts, the sparsity of these texts makes it difficult to capture important semantic relationships. A solution calls for a method that not only considers term occurrence, but also handles sparseness well. In this work, we introduce such an approach, the Term Based Semantic Clusters (TBSeC) that employs terms to create distinctive semantic concept clusters. These clusters are ranked using a semantic similarity function which in turn defines a semantic feature space that can be used for text classification. Our method is evaluated in an invoice classification task. Compared to well-known content representation methods the proposed method performs competitively. + R19-1102 + 10.26615/978-954-452-056-4_102 + + + Quotation Detection and Classification with a Corpus-Agnostic Model + SeanPapay + SebastianPadó + 888–894 + The detection of quotations (i.e., reported speech, thought, and writing) has established itself as an NLP analysis task. However, state-of-the-art models have been developed on the basis of specific corpora and incorpo- rate a high degree of corpus-specific assumptions and knowledge, which leads to fragmentation. In the spirit of task-agnostic modeling, we present a corpus-agnostic neural model for quotation detection and evaluate it on three corpora that vary in language, text genre, and structural assumptions. The model (a) approaches the state-of-the-art on the corpora when using established feature sets and (b) shows reasonable performance even when us- ing solely word forms, which makes it applicable for non-standard (i.e., historical) corpora. + R19-1103 + 10.26615/978-954-452-056-4_103 + + + Validation of Facts Against Textual Sources + Vamsi KrishnaPendyala + SimranSinha + SatyaPrakash + ShriyaReddy + AnupamJamatia + 895–903 + In today’s digital world of information, a fact verification system to disprove assertions made in speech, print media or online content is the need of the hour. We propose a system which would verify a claim against a source and classify the claim to be true, false, out-of-context or an inappropriate claim with respect to the textual source provided to the system. A true label is used if the claim is true, false if it is false, if the claim has no relation with the source then it is classified as out-of-context and if the claim cannot be verified at all then it is classified as inappropriate. This would help us to verify a claim or a fact as well as know about the source or our knowledge base against which we are trying to verify our facts. We used a two-step approach to achieve our goal. At first, we retrieved evidence related to the claims from the textual source using the Term Frequency-Inverse Document Frequency(TF-IDF) vectors. Later we classified the claim-evidence pairs as true, false, inappropriate and out of context using a modified version of textual entailment module. Textual entailment module calculates the probability of each sentence supporting the claim, contradicting the claim or not providing any relevant information using Bi-LSTM network to assess the veracity of the claim. The accuracy of the best performing system is 64.49% + R19-1104 + 10.26615/978-954-452-056-4_104 + + + A Neural Network Component for Knowledge-Based Semantic Representations of Text + AlejandroPiad-Morffis + RafaelMuñoz + YoanGutiérrez + YudivianAlmeida-Cruz + SuilanEstevez-Velarde + AndrésMontoyo + 904–911 + This paper presents Semantic Neural Networks (SNNs), a knowledge-aware component based on deep learning. SNNs can be trained to encode explicit semantic knowledge from an arbitrary knowledge base, and can subsequently be combined with other deep learning architectures. At prediction time, SNNs provide a semantic encoding extracted from the input data, which can be exploited by other neural network components to build extended representation models that can face alternative problems. The SNN architecture is defined in terms of the concepts and relations present in a knowledge base. Based on this architecture, a training procedure is developed. Finally, an experimental setup is presented to illustrate the behaviour and performance of a SNN for a specific NLP problem, in this case, opinion mining for the classification of movie reviews. + R19-1105 + 10.26615/978-954-452-056-4_105 + + + Toponym Detection in the Bio-Medical Domain: A Hybrid Approach with Deep Learning + AlistairPlum + TharinduRanasinghe + ConstantinOrasan + 912–921 + This paper compares how different machine learning classifiers can be used together with simple string matching and named entity recognition to detect locations in texts. We compare five different state-of-the-art machine learning classifiers in order to predict whether a sentence contains a location or not. Following this classification task, we use a string matching algorithm with a gazetteer to identify the exact index of a toponym within the sentence. We evaluate different approaches in terms of machine learning classifiers, text pre-processing and location extraction on the SemEval-2019 Task 12 dataset, compiled for toponym resolution in the bio-medical domain. Finally, we compare the results with our system that was previously submitted to the SemEval-2019 task evaluation. + R19-1106 + 10.26615/978-954-452-056-4_106 + + + Combining <fixed-case>PBSMT</fixed-case> and <fixed-case>NMT</fixed-case> Back-translated Data for Efficient <fixed-case>NMT</fixed-case> + AlbertoPoncelas + MajaPopović + DimitarShterionov + GideonMaillette de Buy Wenniger + AndyWay + 922–931 + Neural Machine Translation (NMT) models achieve their best performance when large sets of parallel data are used for training. Consequently, techniques for augmenting the training set have become popular recently. One of these methods is back-translation, which consists on generating synthetic sentences by translating a set of monolingual, target-language sentences using a Machine Translation (MT) model. Generally, NMT models are used for back-translation. In this work, we analyze the performance of models when the training data is extended with synthetic data using different MT approaches. In particular we investigate back-translated data generated not only by NMT but also by Statistical Machine Translation (SMT) models and combinations of both. The results reveal that the models achieve the best performances when the training set is augmented with back-translated data created by merging different MT approaches. + R19-1107 + 10.26615/978-954-452-056-4_107 + + + Unsupervised dialogue intent detection via hierarchical topic model + ArtemPopov + VictorBulatov + DaryaPolyudova + EugeniaVeselova + 932–938 + One of the challenges during a task-oriented chatbot development is the scarce availability of the labeled training data. The best way of getting one is to ask the assessors to tag each dialogue according to its intent. Unfortunately, performing labeling without any provisional collection structure is difficult since the very notion of the intent is ill-defined. In this paper, we propose a hierarchical multimodal regularized topic model to obtain a first approximation of the intent set. Our rationale for hierarchical models usage is their ability to take into account several degrees of the dialogues relevancy. We attempt to build a model that can distinguish between subject-based (e.g. medicine and transport topics) and action-based (e.g. filing of an application and tracking application status) similarities. In order to achieve this, we divide set of all features into several groups according to part-of-speech analysis. Various feature groups are treated differently on different hierarchy levels. + R19-1108 + 10.26615/978-954-452-056-4_108 + + + Graph Embeddings for Frame Identification + AlexanderPopov + JenniferSikos + 939–948 + Lexical resources such as WordNet (Miller, 1995) and FrameNet (Baker et al., 1998) are organized as graphs, where relationships between words are made explicit via the structure of the resource. This work explores how structural information from these lexical resources can lead to gains in a downstream task, namely frame identification. While much of the current work in frame identification uses various neural architectures to predict frames, those neural architectures only use representations of frames based on annotated corpus data. We demonstrate how incorporating knowledge directly from the FrameNet graph structure improves the performance of a neural network-based frame identification system. Specifically, we construct a bidirectional LSTM with a loss function that incorporates various graph- and corpus-based frame embeddings for learning and ultimately achieves strong performance gains with the graph-based embeddings over corpus-based embeddings alone. + R19-1109 + 10.26615/978-954-452-056-4_109 + + + Know Your Graph. State-of-the-Art Knowledge-Based <fixed-case>WSD</fixed-case> + AlexanderPopov + KirilSimov + PetyaOsenova + 949–958 + This paper introduces several improvements over the current state of the art in knowledge-based word sense disambiguation. Those innovations are the result of modifying and enriching a knowledge base created originally on the basis of WordNet. They reflect several separate but connected strategies: manipulating the shape and the content of the knowledge base, assigning weights over the relations in the knowledge base, and the addition of new relations to it. The main contribution of the paper is to demonstrate that the previously proposed knowledge bases organize linguistic and world knowledge suboptimally for the task of word sense disambiguation. In doing so, the paper also establishes a new state of the art for knowledge-based approaches. Its best models are competitive in the broader context of supervised systems as well. + R19-1110 + 10.26615/978-954-452-056-4_110 + + + Are ambiguous conjunctions problematic for machine translation? + MajaPopović + SheilaCastilho + 959–966 + The translation of ambiguous words still poses challenges for machine translation. In this work, we carry out a systematic quantitative analysis regarding the ability of different machine translation systems to disambiguate the source language conjunctions “but” and “and”. We evaluate specialised test sets focused on the translation of these two conjunctions. The test sets contain source languages that do not distinguish different variants of the given conjunction, whereas the target languages do. In total, we evaluate the conjunction “but” on 20 translation outputs, and the conjunction “and” on 10. All machine translation systems almost perfectly recognise one variant of the target conjunction, especially for the source conjunction “but”. The other target variant, however, represents a challenge for machine translation systems, with accuracy varying from 50% to 95% for “but” and from 20% to 57% for “and”. The major error for all systems is replacing the correct target variant with the opposite one. + R19-1111 + 10.26615/978-954-452-056-4_111 + + + <fixed-case>ULSA</fixed-case>na: Universal Language Semantic Analyzer + OndřejPražák + MiloslavKonopik + 967–972 + We present a live cross-lingual system capable of producing shallow semantic annotations of natural language sentences for 51 languages at this time. The domain of the input sentences is in principle unconstrained. The system uses single training data (in English) for all the languages. The resulting semantic annotations are therefore consistent across different languages. We use CoNLL Semantic Role Labeling training data and Universal dependencies as the basis for the system. The system is publicly available and supports processing data in batches; therefore, it can be easily used by the community for the following research tasks. + R19-1112 + 10.26615/978-954-452-056-4_112 + + + Machine Learning Approach to Fact-Checking in West <fixed-case>S</fixed-case>lavic Languages + PavelPřibáň + TomášHercig + JosefSteinberger + 973–979 + Fake news detection and closely-related fact-checking have recently attracted a lot of attention. Automatization of these tasks has been already studied for English. For other languages, only a few studies can be found (e.g. (Baly et al., 2018)), and to the best of our knowledge, no research has been conducted for West Slavic languages. In this paper, we present datasets for Czech, Polish, and Slovak. We also ran initial experiments which set a baseline for further research into this area. + R19-1113 + 10.26615/978-954-452-056-4_113 + + + <fixed-case>NE</fixed-case>-Table: A Neural key-value table for Named Entities + JanarthananRajendran + JatinGanhotra + XiaoxiaoGuo + MoYu + SatinderSingh + LazarosPolymenakos + 980–993 + Many Natural Language Processing (NLP) tasks depend on using Named Entities (NEs) that are contained in texts and in external knowledge sources. While this is easy for humans, the present neural methods that rely on learned word embeddings may not perform well for these NLP tasks, especially in the presence of Out-Of-Vocabulary (OOV) or rare NEs. In this paper, we propose a solution for this problem, and present empirical evaluations on: a) a structured Question-Answering task, b) three related Goal-Oriented dialog tasks, and c) a Reading-Comprehension task, which show that the proposed method can be effective in dealing with both in-vocabulary and OOV NEs. We create extended versions of dialog bAbI tasks 1,2 and 4 and OOV versions of the CBT test set which are available at - https://github.com/IBM/ne-table-datasets/ + R19-1114 + 10.26615/978-954-452-056-4_114 + + + Enhancing Unsupervised Sentence Similarity Methods with Deep Contextualised Word Representations + TharinduRanasinghe + ConstantinOrasan + RuslanMitkov + 994–1003 + Calculating Semantic Textual Similarity (STS) plays a significant role in many applications such as question answering, document summarisation, information retrieval and information extraction. All modern state of the art STS methods rely on word embeddings one way or another. The recently introduced contextualised word embeddings have proved more effective than standard word embeddings in many natural language processing tasks. This paper evaluates the impact of several contextualised word embeddings on unsupervised STS methods and compares it with the existing supervised/unsupervised STS methods for different datasets in different languages and different domains + R19-1115 + 10.26615/978-954-452-056-4_115 + + + Semantic Textual Similarity with <fixed-case>S</fixed-case>iamese Neural Networks + TharinduRanasinghe + ConstantinOrasan + RuslanMitkov + 1004–1011 + Calculating the Semantic Textual Similarity (STS) is an important research area in natural language processing which plays a significant role in many applications such as question answering, document summarisation, information retrieval and information extraction. This paper evaluates Siamese recurrent architectures, a special type of neural networks, which are used here to measure STS. Several variants of the architecture are compared with existing methods + R19-1116 + 10.26615/978-954-452-056-4_116 + + + Analysing the Impact of Supervised Machine Learning on Automatic Term Extraction: <fixed-case>HAMLET</fixed-case> vs <fixed-case>T</fixed-case>ermo<fixed-case>S</fixed-case>tat + AylaRigouts Terryn + PatrickDrouin + VeroniqueHoste + ElsLefever + 1012–1021 + Traditional approaches to automatic term extraction do not rely on machine learning (ML) and select the top n ranked candidate terms or candidate terms above a certain predefined cut-off point, based on a limited number of linguistic and statistical clues. However, supervised ML approaches are gaining interest. Relatively little is known about the impact of these supervised methodologies; evaluations are often limited to precision, and sometimes recall and f1-scores, without information about the nature of the extracted candidate terms. Therefore, the current paper presents a detailed and elaborate analysis and comparison of a traditional, state-of-the-art system (TermoStat) and a new, supervised ML approach (HAMLET), using the results obtained for the same, manually annotated, Dutch corpus about dressage. + R19-1117 + 10.26615/978-954-452-056-4_117 + + + Distant Supervision for Sentiment Attitude Extraction + NicolayRusnachenko + NataliaLoukachevitch + ElenaTutubalina + 1022–1030 + News articles often convey attitudes between the mentioned subjects, which is essential for understanding the described situation. In this paper, we describe a new approach to distant supervision for extracting sentiment attitudes between named entities mentioned in texts. Two factors (pair-based and frame-based) were used to automatically label an extensive news collection, dubbed as RuAttitudes. The latter became a basis for adaptation and training convolutional architectures, including piecewise max pooling and full use of information across different sentences. The results show that models, trained with RuAttitudes, outperform ones that were trained with only supervised learning approach and achieve 13.4% increase in F1-score on RuSentRel collection. + R19-1118 + 10.26615/978-954-452-056-4_118 + + + Self-Attentional Models Application in Task-Oriented Dialogue Generation Systems + MansourSaffar Mehrjardi + AmineTrabelsi + OsmarR. Zaiane + 1031–1040 + Self-attentional models are a new paradigm for sequence modelling tasks which differ from common sequence modelling methods, such as recurrence-based and convolution-based sequence learning, in the way that their architecture is only based on the attention mechanism. Self-attentional models have been used in the creation of the state-of-the-art models in many NLP task such as neural machine translation, but their usage has not been explored for the task of training end-to-end task-oriented dialogue generation systems yet. In this study, we apply these models on the DSTC2 dataset for training task-oriented chatbots. Our finding shows that self-attentional models can be exploited to create end-to-end task-oriented chatbots which not only achieve higher evaluation scores compared to recurrence-based models, but also do so more efficiently. + R19-1119 + 10.26615/978-954-452-056-4_119 + + + Whom to Learn From? Graph- vs. Text-based Word Embeddings + MałgorzataSalawa + AntónioBranco + RubenBranco + JoãoAntónio Rodrigues + ChakavehSaedi + 1041–1051 + Vectorial representations of meaning can be supported by empirical data from diverse sources and obtained with diverse embedding approaches. This paper aims at screening this experimental space and reports on an assessment of word embeddings supported (i) by data in raw texts vs. in lexical graphs, (ii) by lexical information encoded in association- vs. inference-based graphs, and obtained (iii) by edge reconstruction- vs. matrix factorisation vs. random walk-based graph embedding methods. The results observed with these experiments indicate that the best solutions with graph-based word embeddings are very competitive, consistently outperforming mainstream text-based ones. + R19-1120 + 10.26615/978-954-452-056-4_120 + + + Persistence pays off: Paying Attention to What the <fixed-case>LSTM</fixed-case> Gating Mechanism Persists + GiancarloSalton + JohnKelleher + 1052–1059 + Recurrent Neural Network Language Models composed of LSTM units, especially those augmented with an external memory, have achieved state-of-the-art results in Language Modeling. However, these models still struggle to process long sequences which are more likely to contain long-distance dependencies because of information fading. In this paper we demonstrate an effective mechanism for retrieving information in a memory augmented LSTM LM based on attending to information in memory in proportion to the number of timesteps the LSTM gating mechanism persisted the information. + R19-1121 + 10.26615/978-954-452-056-4_121 + + + Development and Evaluation of Three Named Entity Recognition Systems for <fixed-case>S</fixed-case>erbian - The Case of Personal Names + BranislavaŠandrih + CvetanaKrstev + RankaStankovic + 1060–1068 + In this paper we present a rule- and lexicon-based system for the recognition of Named Entities (NE) in Serbian newspaper texts that was used to prepare a gold standard annotated with personal names. It was further used to prepare training sets for four different levels of annotation, which were further used to train two Named Entity Recognition (NER) systems: Stanford and spaCy. All obtained models, together with a rule- and lexicon-based system were evaluated on two sample texts: a part of the gold standard and an independent newspaper text of approximately the same size. The results show that rule- and lexicon-based system outperforms trained models in all four scenarios (measured by F1), while Stanford models has the highest precision. All systems obtain best results in recognizing full names, while the recognition of first names only is rather poor. The produced models are incorporated into a Web platform NER&Beyond that provides various NE-related functions. + R19-1122 + 10.26615/978-954-452-056-4_122 + + + Moral Stance Recognition and Polarity Classification from <fixed-case>T</fixed-case>witter and Elicited Text + WesleySantos + IvandréParaboni + 1069–1075 + We introduce a labelled corpus of stances about moral issues for the Brazilian Portuguese language, and present reference results for both the stance recognition and polarity classification tasks. The corpus is built from Twitter and further expanded with data elicited through crowd sourcing and labelled by their own authors. Put together, the corpus and reference results are expected to be taken as a baseline for further studies in the field of stance recognition and polarity classification from text. + R19-1123 + 10.26615/978-954-452-056-4_123 + + + The “Jump and Stay” Method to Discover Proper Verb Centered Constructions in Corpus Lattices + BálintSass + 1076–1084 + The research presented here is based on the theoretical model of corpus lattices. We implemented this as an effective data structure, and developed an algorithm based on this structure to discover essential verbal expressions from corpus data. The idea behind the algorithm is the “jump and stay” principle, which tells us that our target expressions will be found at such places in the lattice where the value of a suitable function (defined on the vertex set of the corpus lattice) significantly increases (jumps) and then remains the same (stays). We evaluated our method on Hungarian data. Evaluation shows that about 75% of the obtained expressions are correct, actual errors are rare. Thus, this paper is 1. a proof of concept concerning the corpus lattice model, opening the way to investigate this structure further through our implementation; and 2. a proof of concept of the “jump and stay” idea and the algorithm itself, opening the way to apply it further, e.g. for other languages. + R19-1124 + 10.26615/978-954-452-056-4_124 + + + Offence in Dialogues: A Corpus-Based Study + JohannesSchäfer + BenBurtenshaw + 1085–1093 + In recent years an increasing number of analyses of offensive language has been published, however, dealing mainly with the automatic detection and classification of isolated instances. In this paper we aim to understand the impact of offensive messages in online conversations diachronically, and in particular the change in offensiveness of dialogue turns. In turn, we aim to measure the progression of offence level as well as its direction - For example, whether a conversation is escalating or declining in offence. We present our method of extracting linear dialogues from tree-structured conversations in social media data and make our code publicly available. Furthermore, we discuss methods to analyse this dataset through changes in discourse offensiveness. Our paper includes two main contributions; first, using a neural network to measure the level of offensiveness in conversations; and second, the analysis of conversations around offensive comments using decoupling functions. + R19-1125 + 10.26615/978-954-452-056-4_125 + + + <fixed-case>E</fixed-case>mo<fixed-case>T</fixed-case>ag – Towards an Emotion-Based Analysis of Emojis + Abu Awal MdShoeb + ShahabRaji + Gerardde Melo + 1094–1103 + Despite being a fairly recent phenomenon, emojis have quickly become ubiquitous. Besides their extensive use in social media, they are now also invoked in customer surveys and feedback forms. Hence, there is a need for techniques to understand their sentiment and emotion. In this work, we provide a method to quantify the emotional association of basic emotions such as anger, fear, joy, and sadness for a set of emojis. We collect and process a unique corpus of 20 million emoji-centric tweets, such that we can capture rich emoji semantics using a comparably small dataset. We evaluate the induced emotion profiles of emojis with regard to their ability to predict word affect intensities as well as sentiment scores. + R19-1126 + 10.26615/978-954-452-056-4_126 + + + A Morpho-Syntactically Informed <fixed-case>LSTM</fixed-case>-<fixed-case>CRF</fixed-case> Model for Named Entity Recognition + LiliaSimeonova + KirilSimov + PetyaOsenova + PreslavNakov + 1104–1113 + We propose a morphologically informed model for named entity recognition, which is based on LSTM-CRF architecture and combines word embeddings, Bi-LSTM character embeddings, part-of-speech (POS) tags, and morphological information. While previous work has focused on learning from raw word input, using word and character embeddings only, we show that for morphologically rich languages, such as Bulgarian, access to POS information contributes more to the performance gains than the detailed morphological information. Thus, we show that named entity recognition needs only coarse-grained POS tags, but at the same time it can benefit from simultaneously using some POS information of different granularity. Our evaluation results over a standard dataset show sizeable improvements over the state-of-the-art for Bulgarian NER. + R19-1127 + 10.26615/978-954-452-056-4_127 + + + Named Entity Recognition in Information Security Domain for <fixed-case>R</fixed-case>ussian + AnastasiiaSirotina + NataliaLoukachevitch + 1114–1120 + In this paper we discuss the named entity recognition task for Russian texts related to cybersecurity. First of all, we describe the problems that arise in course of labeling unstructured texts from information security domain. We introduce guidelines for human annotators, according to which a corpus has been marked up. Then, a CRF-based system and different neural architectures have been implemented and applied to the corpus. The named entity recognition systems have been evaluated and compared to determine the most efficient one. + R19-1128 + 10.26615/978-954-452-056-4_128 + + + Cross-Family Similarity Learning for Cognate Identification in Low-Resource Languages + ElielSoisalon-Soininen + MarkGranroth-Wilding + 1121–1130 + We address the problem of cognate identification across vocabulary pairs of any set of languages. In particular, we focus on the case where the examined pair of languages are low-resource to the extent that no training data whatsoever in these languages, or even closely related ones, are available for the task. We investigate the extent to which training data from another, unrelated language family can be used instead. Our approach consists of learning a similarity metric from example cognates in Indo-European languages and applying it to low-resource Sami languages of the Uralic family. We apply two models following previous work: a Siamese convolutional neural network (S-CNN) and a support vector machine (SVM), and compare them with a Levenshtein-distance baseline. We test performance on three Sami languages and find that the S-CNN outperforms the other approaches, suggesting that it is better able to learn such general characteristics of cognateness that carry over across language families. We also experiment with fine-tuning the S-CNN model with data from within the language family in order to quantify how well this model can make use of a small amount of target-domain data to adapt. + R19-1129 + 10.26615/978-954-452-056-4_129 + + + Automatic Detection of Translation Direction + IliaSominsky + ShulyWintner + 1131–1140 + Parallel corpora are crucial resources for NLP applications, most notably for machine translation. The direction of the (human) translation of parallel corpora has been shown to have significant implications for the quality of statistical machine translation systems that are trained with such corpora. We describe a method for determining the direction of the (manual) translation of parallel corpora at the sentence-pair level. Using several linguistically-motivated features, coupled with a neural network model, we obtain high accuracy on several language pairs. Furthermore, we demonstrate that the accuracy is correlated with the (typological) distance between the two languages. + R19-1130 + 10.26615/978-954-452-056-4_130 + + + Automated Text Simplification as a Preprocessing Step for Machine Translation into an Under-resourced Language + SanjaŠtajner + MajaPopović + 1141–1150 + In this work, we investigate the possibility of using fully automatic text simplification system on the English source in machine translation (MT) for improving its translation into an under-resourced language. We use the state-of-the-art automatic text simplification (ATS) system for lexically and syntactically simplifying source sentences, which are then translated with two state-of-the-art English-to-Serbian MT systems, the phrase-based MT (PBMT) and the neural MT (NMT). We explore three different scenarios for using the ATS in MT: (1) using the raw output of the ATS; (2) automatically filtering out the sentences with low grammaticality and meaning preservation scores; and (3) performing a minimal manual correction of the ATS output. Our results show improvement in fluency of the translation regardless of the chosen scenario, and difference in success of the three scenarios depending on the MT approach used (PBMT or NMT) with regards to improving translation fluency and post-editing effort. + R19-1131 + 10.26615/978-954-452-056-4_131 + + + Investigating Multilingual Abusive Language Detection: A Cautionary Tale + KennethSteimel + DanielDakota + YueChen + SandraKübler + 1151–1160 + Abusive language detection has received much attention in the last years, and recent approaches perform the task in a number of different languages. We investigate which factors have an effect on multilingual settings, focusing on the compatibility of data and annotations. In the current paper, we focus on English and German. Our findings show large differences in performance between the two languages. We find that the best performance is achieved by different classification algorithms. Sampling to address class imbalance issues is detrimental for German and beneficial for English. The only similarity that we find is that neither data set shows clear topics when we compare the results of topic modeling to the gold standard. Based on our findings, we can conclude that a multilingual optimization of classifiers is not possible even in settings where comparable data sets are used. + R19-1132 + 10.26615/978-954-452-056-4_132 + + + Augmenting a <fixed-case>B</fixed-case>i<fixed-case>LSTM</fixed-case> Tagger with a Morphological Lexicon and a Lexical Category Identification Step + SteinþórSteingrímsson + ÖrvarKárason + HrafnLoftsson + 1161–1168 + Previous work on using BiLSTM models for PoS tagging has primarily focused on small tagsets. We evaluate BiLSTM models for tagging Icelandic, a morphologically rich language, using a relatively large tagset. Our baseline BiLSTM model achieves higher accuracy than any other previously published tagger, when not taking advantage of a morphological lexicon. When we extend the model by incorporating such data, we outperform the earlier state-of-the-art results by a significant margin. We also report on work in progress that attempts to address the problem of data sparsity inherent to morphologically detailed, fine-grained tagsets. We experiment with training a separate model on only the lexical category and using the coarse-grained output tag as an input into to the main model. This method further increases the accuracy and reduces the tagging errors by 21.3% compared to previous state-of-the-art results. Finally, we train and test our tagger on a new gold standard for Icelandic. + R19-1133 + 10.26615/978-954-452-056-4_133 + + + Comparison of Machine Learning Approaches for Industry Classification Based on Textual Descriptions of Companies + AndreyTagarev + NikolaTulechki + SvetlaBoytcheva + 1169–1175 + This paper addresses the task of categorizing companies within industry classification schemes. The datasets consists of encyclopedic articles about companies and their economic activities. The target classification schema is build by mapping linked open data in a semi-supervised manner. Target classes are build bottom-up from DBpedia. We apply several state of the art text classification techniques, based both on deep-learning and classical vector-space models. + R19-1134 + 10.26615/978-954-452-056-4_134 + + + A Quantum-Like Approach to Word Sense Disambiguation + FabioTamburini + 1176–1185 + This paper presents a novel algorithm for Word Sense Disambiguation (WSD) based on Quantum Probability Theory. The Quantum WSD algorithm requires concepts representations as vectors in the complex domain and thus we have developed a technique for computing complex word and sentence embeddings based on the Paragraph Vectors algorithm. Despite the proposed method is quite simple and that it does not require long training phases, when it is evaluated on a standardized benchmark for this task it exhibits state-of-the-art (SOTA) performances. + R19-1135 + 10.26615/978-954-452-056-4_135 + + + Understanding Neural Machine Translation by Simplification: The Case of Encoder-free Models + GongboTang + RicoSennrich + JoakimNivre + 1186–1193 + In this paper, we try to understand neural machine translation (NMT) via simplifying NMT architectures and training encoder-free NMT models. In an encoder-free model, the sums of word embeddings and positional embeddings represent the source. The decoder is a standard Transformer or recurrent neural network that directly attends to embeddings via attention mechanisms. Experimental results show (1) that the attention mechanism in encoder-free models acts as a strong feature extractor, (2) that the word embeddings in encoder-free models are competitive to those in conventional models, (3) that non-contextualized source representations lead to a big performance drop, and (4) that encoder-free models have different effects on alignment quality for German-English and Chinese-English. + R19-1136 + 10.26615/978-954-452-056-4_136 + + + Text-Based Joint Prediction of Numeric and Categorical Attributes of Entities in Knowledge Bases + VThejas + AbhijeetGupta + SebastianPadó + 1194–1202 + Collaboratively constructed knowledge bases play an important role in information systems, but are essentially always incomplete. Thus, a large number of models has been developed for Knowledge Base Completion, the task of predicting new attributes of entities given partial descriptions of these entities. Virtually all of these models either concentrate on numeric attributes (<Italy,GDP,2T$>) or they concentrate on categorical attributes (<Tim Cook,chairman,Apple>). In this paper, we propose a simple feed-forward neural architecture to jointly predict numeric and categorical attributes based on embeddings learned from textual occurrences of the entities in question. Following insights from multi-task learning, our hypothesis is that due to the correlations among attributes of different kinds, joint prediction improves over separate prediction. Our experiments on seven FreeBase domains show that this hypothesis is true of the two attribute types: we find substantial improvements for numeric attributes in the joint model, while performance remains largely unchanged for categorical attributes. Our analysis indicates that this is the case because categorical attributes, many of which describe membership in various classes, provide useful ‘background knowledge’ for numeric prediction, while this is true to a lesser degree in the inverse direction. + R19-1137 + 10.26615/978-954-452-056-4_137 + + + <fixed-case>S</fixed-case>en<fixed-case>Z</fixed-case>i: A Sentiment Analysis Lexicon for the Latinised <fixed-case>A</fixed-case>rabic (<fixed-case>A</fixed-case>rabizi) + TahaTobaili + MiriamFernandez + HarithAlani + SanaaSharafeddine + HazemHajj + GoranGlavaš + 1203–1211 + Arabizi is an informal written form of dialectal Arabic transcribed in Latin alphanumeric characters. It has a proven popularity on chat platforms and social media, yet it suffers from a severe lack of natural language processing (NLP) resources. As such, texts written in Arabizi are often disregarded in sentiment analysis tasks for Arabic. In this paper we describe the creation of a sentiment lexicon for Arabizi that was enriched with word embeddings. The result is a new Arabizi lexicon consisting of 11.3K positive and 13.3K negative words. We evaluated this lexicon by classifying the sentiment of Arabizi tweets achieving an F1-score of 0.72. We provide a detailed error analysis to present the challenges that impact the sentiment analysis of Arabizi. + R19-1138 + 10.26615/978-954-452-056-4_138 + + + Mining the <fixed-case>UK</fixed-case> Web Archive for Semantic Change Detection + AdamTsakalidis + MaryaBazzi + MihaiCucuringu + PierpaoloBasile + BarbaraMcGillivray + 1212–1221 + Semantic change detection (i.e., identifying words whose meaning has changed over time) started emerging as a growing area of research over the past decade, with important downstream applications in natural language processing, historical linguistics and computational social science. However, several obstacles make progress in the domain slow and difficult. These pertain primarily to the lack of well-established gold standard datasets, resources to study the problem at a fine-grained temporal resolution, and quantitative evaluation approaches. In this work, we aim to mitigate these issues by (a) releasing a new labelled dataset of more than 47K word vectors trained on the UK Web Archive over a short time-frame (2000-2013); (b) proposing a variant of Procrustes alignment to detect words that have undergone semantic shift; and (c) introducing a rank-based approach for evaluation purposes. Through extensive numerical experiments and validation, we illustrate the effectiveness of our approach against competitive baselines. Finally, we also make our resources publicly available to further enable research in the domain. + R19-1139 + 10.26615/978-954-452-056-4_139 + + + Cross-Lingual Word Embeddings for Morphologically Rich Languages + AhmetÜstün + GosseBouma + Gertjanvan Noord + 1222–1228 + Cross-lingual word embedding models learn a shared vector space for two or more languages so that words with similar meaning are represented by similar vectors regardless of their language. Although the existing models achieve high performance on pairs of morphologically simple languages, they perform very poorly on morphologically rich languages such as Turkish and Finnish. In this paper, we propose a morpheme-based model in order to increase the performance of cross-lingual word embeddings on morphologically rich languages. Our model includes a simple extension which enables us to exploit morphemes for cross-lingual mapping. We applied our model for the Turkish-Finnish language pair on the bilingual word translation task. Results show that our model outperforms the baseline models by 2% in the nearest neighbour ranking. + R19-1140 + 10.26615/978-954-452-056-4_140 + + + It Takes Nine to Smell a Rat: Neural Multi-Task Learning for Check-Worthiness Prediction + SlavenaVasileva + PepaAtanasova + LluísMàrquez + AlbertoBarrón-Cedeño + PreslavNakov + 1229–1239 + We propose a multi-task deep-learning approach for estimating the check-worthiness of claims in political debates. Given a political debate, such as the 2016 US Presidential and Vice-Presidential ones, the task is to predict which statements in the debate should be prioritized for fact-checking. While different fact-checking organizations would naturally make different choices when analyzing the same debate, we show that it pays to learn from multiple sources simultaneously (PolitiFact, FactCheck, ABC, CNN, NPR, NYT, Chicago Tribune, The Guardian, and Washington Post) in a multi-task learning setup, even when a particular source is chosen as a target to imitate. Our evaluation shows state-of-the-art results on a standard dataset for the task of check-worthiness prediction. + R19-1141 + 10.26615/978-954-452-056-4_141 + + + Deep learning contextual models for prediction of sport event outcome from sportsman’s interviews + BorisVelichkov + IvanKoychev + SvetlaBoytcheva + 1240–1246 + This paper presents an approach for prediction of results for sport events. Usually the sport forecasting approaches are based on structured data. We test the hypothesis that the sports results can be predicted by using natural language processing and machine learning techniques applied over interviews with the players shortly before the sport events. The proposed method uses deep learning contextual models, applied over unstructured textual documents. Several experiments were performed for interviews with players in individual sports like boxing, martial arts, and tennis. The results from the conducted experiment confirmed our initial assumption that an interview from a sportsman before a match contains information that can be used for prediction the outcome from it. Furthermore, the results provide strong evidence in support of our research hypothesis, that is, we can predict the outcome from a sport match analyzing an interview, given before it. + R19-1142 + 10.26615/978-954-452-056-4_142 + + + Exploiting Frame-Semantics and Frame-Semantic Parsing for Automatic Extraction of Typological Information from Descriptive Grammars of Natural Languages + Shafqat MumtazVirk + AzamSheikh Muhammad + LarsBorin + Muhammad IrfanAslam + SaaniaIqbal + NaziaKhurram + 1247–1256 + We describe a novel system for automatic extraction of typological linguistic information from descriptive grammars of natural languages, applying the theory of frame semantics in the form of frame-semantic parsing. The current proof-of-concept system covers a few selected linguistic features, but the methodology is general and can be extended not only to other typological features but also to descriptive grammars written in languages other than English. Such a system is expected to be a useful assistance for automatic curation of typological databases which otherwise are built manually, a very labor and time consuming as well as cognitively taxing enterprise. + R19-1143 + 10.26615/978-954-452-056-4_143 + + + Exploiting Open <fixed-case>IE</fixed-case> for Deriving Multiple Premises Entailment Corpus + MartinVíta + JakubKlímek + 1257–1264 + Natural language inference (NLI) is a key part of natural language understanding. The NLI task is defined as a decision problem whether a given sentence – hypothesis – can be inferred from a given text. Typically, we deal with a text consisting of just a single premise/single sentence, which is called a single premise entailment (SPE) task. Recently, a derived task of NLI from multiple premises (MPE) was introduced together with the first annotated corpus and corresponding several strong baselines. Nevertheless, the further development in MPE field requires accessibility of huge amounts of annotated data. In this paper we introduce a novel method for rapid deriving of MPE corpora from an existing NLI (SPE) annotated data that does not require any additional annotation work. This proposed approach is based on using an open information extraction system. We demonstrate the application of the method on a well known SNLI corpus. Over the obtained corpus, we provide the first evaluations as well as we state a strong baseline. + R19-1144 + 10.26615/978-954-452-056-4_144 + + + Towards Adaptive Text Summarization: How Does Compression Rate Affect Summary Readability of <fixed-case>L</fixed-case>2 Texts? + TatianaVodolazova + ElenaLloret + 1265–1274 + This paper addresses the problem of readability of automatically generated summaries in the context of second language learning. For this we experimented with a new corpus of level-annotated simplified English texts. The texts were summarized using a total of 7 extractive and abstractive summarization systems with compression rates of 20%, 40%, 60% and 80%. We analyzed the generated summaries in terms of lexical, syntactic and length-based features of readability, and concluded that summary complexity depends on the compression rate, summarization technique and the nature of the summarized corpus. Our experiments demonstrate the importance of choosing appropriate summarization techniques that align with user’s needs and language proficiency. + R19-1145 + 10.26615/978-954-452-056-4_145 + + + The Impact of Rule-Based Text Generation on the Quality of Abstractive Summaries + TatianaVodolazova + ElenaLloret + 1275–1284 + In this paper we describe how an abstractive text summarization method improved the informativeness of automatic summaries by integrating syntactic text simplification, subject-verb-object concept frequency scoring and a set of rules that transform text into its semantic representation. We analyzed the impact of each component of our approach on the quality of generated summaries and tested it on DUC 2002 dataset. Our experiments showed that our approach outperformed other state-of-the-art abstractive methods while maintaining acceptable linguistic quality and redundancy rate. + R19-1146 + 10.26615/978-954-452-056-4_146 + + + <fixed-case>ETNLP</fixed-case>: A Visual-Aided Systematic Approach to Select Pre-Trained Embeddings for a Downstream Task + SonVu Xuan + ThanhVu + SonTran + LiliJiang + 1285–1294 + Given many recent advanced embedding models, selecting pre-trained word representation (i.e., word embedding) models best fit for a specific downstream NLP task is non-trivial. In this paper, we propose a systematic approach to extracting, evaluating, and visualizing multiple sets of pre-trained word embed- dings to determine which embeddings should be used in a downstream task. First, for extraction, we provide a method to extract a subset of the embeddings to be used in the downstream NLP tasks. Second, for evaluation, we analyse the quality of pre-trained embeddings using an input word analogy list. Finally, we visualize the embedding space to explore the embedded words interactively. We demonstrate the effectiveness of the proposed approach on our pre-trained word embedding models in Vietnamese to select which models are suitable for a named entity recogni- tion (NER) task. Specifically, we create a large Vietnamese word analogy list to evaluate and select the pre-trained embedding models for the task. We then utilize the selected embed- dings for the NER task and achieve the new state-of-the-art results on the task benchmark dataset. We also apply the approach to another downstream task of privacy-guaranteed embedding selection, and show that it helps users quickly select the most suitable embeddings. In addition, we create an open-source system using the proposed systematic approach to facilitate similar studies on other NLP tasks. The source code and data are available at https: //github.com/vietnlp/etnlp. + R19-1147 + 10.26615/978-954-452-056-4_147 + + + Tagger for Polish Computer Mediated Communication Texts + WiktorWalentynowicz + MaciejPiasecki + MarcinOleksy + 1295–1303 + In this paper we present a morpho-syntactic tagger dedicated to Computer-mediated Communication texts in Polish. Its construction is based on an expanded RNN-based neural network adapted to the work on noisy texts. Among several techniques, the tagger utilises fastText embedding vectors, sequential character embedding vectors, and Brown clustering for the coarse-grained representation of sentence structures. In addition a set of manually written rules was proposed for post-processing. The system was trained to disambiguate descriptions of words in relation to Parts of Speech tags together with the full morphological information in terms of values for the different grammatical categories. We present also evaluation of several model variants on the gold standard annotated CMC data, comparison to the state-of-the-art taggers for Polish and error analysis. The proposed tagger shows significantly better results in this domain and demonstrates the viability of adaptation. + R19-1148 + 10.26615/978-954-452-056-4_148 + + + Evaluation of vector embedding models in clustering of text documents + TomaszWalkowiak + MateuszGniewkowski + 1304–1311 + The paper presents an evaluation of word embedding models in clustering of texts in the Polish language. Authors verified six different embedding models, starting from widely used word2vec, across fastText with character n-grams embedding, to deep learning-based ELMo and BERT. Moreover, four standardisation methods, three distance measures and four clustering methods were evaluated. The analysis was performed on two corpora of texts in Polish classified into subjects. The Adjusted Mutual Information (AMI) metric was used to verify the quality of clustering results. The performed experiments show that Skipgram models with n-grams character embedding, built on KGR10 corpus and provided by Clarin-PL, outperforms other publicly available models for Polish. Moreover, presented results suggest that Yeo–Johnson transformation for document vectors standardisation and Agglomerative Clustering with a cosine distance should be used for grouping of text documents. + R19-1149 + 10.26615/978-954-452-056-4_149 + + + Bigger versus Similar: Selecting a Background Corpus for First Story Detection Based on Distributional Similarity + FeiWang + Robert J.Ross + John D.Kelleher + 1312–1320 + The current state of the art for First Story Detection (FSD) are nearest neighbour-based models with traditional term vector representations; however, one challenge faced by FSD models is that the document representation is usually defined by the vocabulary and term frequency from a background corpus. Consequently, the ideal background corpus should arguably be both large-scale to ensure adequate term coverage, and similar to the target domain in terms of the language distribution. However, given these two factors cannot always be mutually satisfied, in this paper we examine whether the distributional similarity of common terms is more important than the scale of common terms for FSD. As a basis for our analysis we propose a set of metrics to quantitatively measure the scale of common terms and the distributional similarity between corpora. Using these metrics we rank different background corpora relative to a target corpus. We also apply models based on different background corpora to the FSD task. Our results show that term distributional similarity is more predictive of good FSD performance than the scale of common terms; and, thus we demonstrate that a smaller recent domain-related corpus will be more suitable than a very large-scale general corpus for FSD. + R19-1150 + 10.26615/978-954-452-056-4_150 + + + Predicting Sentiment of Polish Language Short Texts + AleksanderWawer + JulitaSobiczewska + 1321–1327 + The goal of this paper is to use all available Polish language data sets to seek the best possible performance in supervised sentiment analysis of short texts. We use text collections with labelled sentiment such as tweets, movie reviews and a sentiment treebank, in three comparison modes. In the first, we examine the performance of models trained and tested on the same text collection using standard cross-validation (in-domain). In the second we train models on all available data except the given test collection, which we use for testing (one vs rest cross-domain). In the third, we train a model on one data set and apply it to another one (one vs one cross-domain). We compare wide range of methods including machine learning on bag-of-words representation, bidirectional recurrent neural networks as well as the most recent pre-trained architectures ELMO and BERT. We formulate conclusions as to cross-domain and in-domain performance of each method. Unsurprisingly, BERT turned out to be a strong performer, especially in the cross-domain setting. What is surprising however, is solid performance of the relatively simple multinomial Naive Bayes classifier, which performed equally well as BERT on several data sets. + R19-1151 + 10.26615/978-954-452-056-4_151 + + + Improving Named Entity Linking Corpora Quality + AlbertWeichselbraun + Adrian M.P.Brasoveanu + PhilippKuntschik + Lyndon J.B.Nixon + 1328–1337 + Gold standard corpora and competitive evaluations play a key role in benchmarking named entity linking (NEL) performance and driving the development of more sophisticated NEL systems. The quality of the used corpora and the used evaluation metrics are crucial in this process. We, therefore, assess the quality of three popular evaluation corpora, identifying four major issues which affect these gold standards: (i) the use of different annotation styles, (ii) incorrect and missing annotations, (iii) Knowledge Base evolution, (iv) and differences in annotating co-occurrences. This paper addresses these issues by formalizing NEL annotations and corpus versioning which allows standardizing corpus creation, supports corpus evolution, and paves the way for the use of lenses to automatically transform between different corpus configurations. In addition, the use of clearly defined scoring rules and evaluation metrics ensures a better comparability of evaluation results. + R19-1152 + 10.26615/978-954-452-056-4_152 + + + Sequential Graph Dependency Parser + SeanWelleck + KyunghyunCho + 1338–1345 + We propose a method for non-projective dependency parsing by incrementally predicting a set of edges. Since the edges do not have a pre-specified order, we propose a set-based learning method. Our method blends graph, transition, and easy-first parsing, including a prior state of the parser as a special case. The proposed transition-based method successfully parses near the state of the art on both projective and non-projective languages, without assuming a certain parsing order. + R19-1153 + 10.26615/978-954-452-056-4_153 + + + Term-Based Extraction of Medical Information: Pre-Operative Patient Education Use Case + MartinWolf + VolhaPetukhova + DietrichKlakow + 1346–1355 + The processing of medical information is not a trivial task for medical non-experts. The paper presents an artificial assistant designed to facilitate a reliable access to medical online contents. Interactions are modelled as doctor-patient Question Answering sessions within a pre-operative patient education scenario where the system addresses patient’s information needs explaining medical events and procedures. This implies an accurate medical information extraction from and reasoning with available medical knowledge and large amounts of unstructured multilingual online data. Bridging the gap between medical knowledge and data, we explore a language-agnostic approach to medical concepts mining from the standard terminologies, and the data-driven collection of the corresponding seed terms in a distant supervision setting for German. Experimenting with different terminologies, features and term matching strategies, we achieved a promising F-score of 0.91 on the medical term extraction task. The concepts and terms are used to search and retrieve definitions from the verified online free resources. The proof-of-concept definition retrieval system is designed and evaluated showing promising results, acceptable by humans in 92% of cases. + R19-1154 + 10.26615/978-954-452-056-4_154 + + + A Survey of the Perceived Text Adaptation Needs of Adults with Autism + VictoriaYaneva + ConstantinOrasan + Le AnHa + NataliaPonomareva + 1356–1363 + NLP approaches to automatic text adaptation often rely on user-need guidelines which are generic and do not account for the differences between various types of target groups. One such group are adults with high-functioning autism, who are usually able to read long sentences and comprehend difficult words but whose comprehension may be impeded by other linguistic constructions. This is especially challenging for real-world user-generated texts such as product reviews, which cannot be controlled editorially and are thus a particularly good applcation for automatic text adaptation systems. In this paper we present a mixed-methods survey conducted with 24 adult web-users diagnosed with autism and an age-matched control group of 33 neurotypical participants. The aim of the survey was to identify whether the group with autism experienced any barriers when reading online reviews, what these potential barriers were, and what NLP methods would be best suited to improve the accessibility of online reviews for people with autism. The group with autism consistently reported significantly greater difficulties with understanding online product reviews compared to the control group and identified issues related to text length, poor topic organisation, and the use of irony and sarcasm. + R19-1155 + 10.26615/978-954-452-056-4_155 + + + An Open, Extendible, and Fast <fixed-case>T</fixed-case>urkish Morphological Analyzer + Olcay TanerYıldız + BegümAvar + GökhanErcan + 1364–1372 + In this paper, we present a two-level morphological analyzer for Turkish. The morphological analyzer consists of five main components: finite state transducer, rule engine for suffixation, lexicon, trie data structure, and LRU cache. We use Java language to implement finite state machine logic and rule engine, Xml language to describe the finite state transducer rules of the Turkish language, which makes the morphological analyzer both easily extendible and easily applicable to other languages. Empowered with the comprehensiveness of a lexicon of 54,000 bare-forms including 19,000 proper nouns, our morphological analyzer presents one of the most reliable analyzers produced so far. The analyzer is compared with Turkish morphological analyzers in the literature. By using LRU cache and a trie data structure, the system can analyze 100,000 words per second, which enables users to analyze huge corpora in a few hours. + R19-1156 + 10.26615/978-954-452-056-4_156 + + + Self-Attention Networks for Intent Detection + SevinjYolchuyeva + GézaNémeth + BálintGyires-Tóth + 1373–1379 + Self-attention networks (SAN) have shown promising performance in various Natural Language Processing (NLP) scenarios, especially in machine translation. One of the main points of SANs is the strength of capturing long-range and multi-scale dependencies from the data. In this paper, we present a novel intent detection system which is based on a self-attention network and a Bi-LSTM. Our approach shows improvement by using a transformer model and deep averaging network-based universal sentence encoder compared to previous solutions. We evaluate the system on Snips, Smart Speaker, Smart Lights, and ATIS datasets by different evaluation metrics. The performance of the proposed model is compared with LSTM with the same datasets. + R19-1157 + 10.26615/978-954-452-056-4_157 + + + <fixed-case>T</fixed-case>urkish Tweet Classification with Transformer Encoder + Atıf EmreYüksel + Yaşar AlimTürkmen + ArzucanÖzgür + BernaAltınel + 1380–1387 + Short-text classification is a challenging task, due to the sparsity and high dimensionality of the feature space. In this study, we aim to analyze and classify Turkish tweets based on their topics. Social media jargon and the agglutinative structure of the Turkish language makes this classification task even harder. As far as we know, this is the first study that uses a Transformer Encoder for short text classification in Turkish. The model is trained in a weakly supervised manner, where the training data set has been labeled automatically. Our results on the test set, which has been manually labeled, show that performing morphological analysis improves the classification performance of the traditional machine learning algorithms Random Forest, Naive Bayes, and Support Vector Machines. Still, the proposed approach achieves an F-score of 89.3 % outperforming those algorithms by at least 5 points. + R19-1158 + 10.26615/978-954-452-056-4_158 + + + Multilingual Dynamic Topic Model + ElaineZosa + MarkGranroth-Wilding + 1388–1396 + Dynamic topic models (DTMs) capture the evolution of topics and trends in time series data.Current DTMs are applicable only to monolingual datasets. In this paper we present the multilingual dynamic topic model (ML-DTM), a novel topic model that combines DTM with an existing multilingual topic modeling method to capture cross-lingual topics that evolve across time. We present results of this model on a parallel German-English corpus of news articles and a comparable corpus of Finnish and Swedish news articles. We demonstrate the capability of ML-DTM to track significant events related to a topic and show that it finds distinct topics and performs as well as existing multilingual topic models in aligning cross-lingual topics. + R19-1159 + 10.26615/978-954-452-056-4_159 + + + A Wide-Coverage Context-Free Grammar for <fixed-case>I</fixed-case>celandic and an Accompanying Parsing System + VilhjálmurÞorsteinsson + HuldaÓladóttir + HrafnLoftsson + 1397–1404 + We present an open-source, wide-coverage context-free grammar (CFG) for Icelandic, and an accompanying parsing system. The grammar has over 5,600 nonterminals, 4,600 terminals and 19,000 productions in fully expanded form, with feature agreement constraints for case, gender, number and person. The parsing system consists of an enhanced Earley-based parser and a mechanism to select best-scoring parse trees from shared packed parse forests. Our parsing system is able to parse about 90% of all sentences in articles published on the main Icelandic news websites. Preliminary evaluation with evalb shows an F-measure of 70.72% on parsed sentences. Our system demonstrates that parsing a morphologically rich language using a wide-coverage CFG can be practical. + R19-1160 + 10.26615/978-954-452-056-4_160 + +
+ + + Proceedings of the Student Research Workshop Associated with RANLP 2019 + R19-2 + University of BarcelonaVenelin Kovatchev + IrinaTemnikova + Belgrade UniversityBranislava Šandrih + Bulgarian Academy of SciencesIvelina Nikolova + OntotextAD + INCOMA Ltd. +
Varna, Bulgaria
+ September + 2019 + + + R19-2000 + + + Normalization of <fixed-case>K</fixed-case>azakh Texts + AssinaAbdussaitova + AlinaAmangeldiyeva + 1–6 + Kazakh language, like other agglutinative languages, has specific difficulties on both recognition of wrong words and generation the corrections for misspelt words. The main goal of this work is to develop a better algorithm for the normalization of Kazakh texts based on traditional and Machine Learning methods, as well as the new approach which is also considered in this paper. The procedure of election among methods of normalization has been conducted in a manner of comparative analysis. The results of the comparative analysis turned up successful and are shown in details. + R19-2001 + 10.26615/issn.2603-2821.2019_001 + + + Classification Approaches to Identify Informative Tweets + PiushAggarwal + 7–15 + Social media platforms have become prime forums for reporting news, with users sharing what they saw, heard or read on social media. News from social media is potentially useful for various stakeholders including aid organizations, news agencies, and individuals. However, social media also contains a vast amount of non-news content. For users to be able to draw on benefits from news reported on social media it is necessary to reliably identify news content and differentiate it from non-news. In this paper, we tackle the challenge of classifying a social post as news or not. To this end, we provide a new manually annotated dataset containing 2,992 tweets from 5 different topical categories. Unlike earlier datasets, it includes postings posted by personal users who do not promote a business or a product and are not affiliated with any organization. We also investigate various baseline systems and evaluate their performance on the newly generated dataset. Our results show that the best classifiers are the SVM and BERT models. + R19-2002 + 10.26615/issn.2603-2821.2019_002 + + + Dialect-Specific Models for Automatic Speech Recognition of <fixed-case>A</fixed-case>frican <fixed-case>A</fixed-case>merican Vernacular <fixed-case>E</fixed-case>nglish + RachelDorn + 16–20 + African American Vernacular English (AAVE) is a widely-spoken dialect of English, yet it is under-represented in major speech corpora. As a result, speakers of this dialect are often misunderstood by NLP applications. This study explores the effect on transcription accuracy of an automatic voice recognition system when AAVE data is used. Models trained on AAVE data and on Standard American English data were compared to a baseline model trained on a combination of the two dialects. The accuracy for both dialect-specific models was significantly higher than the baseline model, with the AAVE model showing over 18% improvement. By isolating the effect of having AAVE speakers in the training data, this study highlights the importance of increasing diversity in the field of natural language processing. + R19-2003 + 10.26615/issn.2603-2821.2019_003 + + + Multilingual Language Models for Named Entity Recognition in <fixed-case>G</fixed-case>erman and <fixed-case>E</fixed-case>nglish + AntoniaBaumann + 21–27 + We assess the language specificity of recent language models by exploring the potential of a multilingual language model. In particular, we evaluate Google’s multilingual BERT (mBERT) model on Named Entity Recognition (NER) in German and English. We expand the work on language model fine-tuning by Howard and Ruder (2018), applying it to the BERT architecture. We successfully reproduce the NER results published by Devlin et al. (2019).Our results show that the multilingual language model generalises well for NER in the chosen languages, matching the native model in English and comparing well with recent approaches for German. However, it does not benefit from the added fine-tuning methods. + R19-2004 + 10.26615/issn.2603-2821.2019_004 + + + Parts of Speech Tagging for <fixed-case>K</fixed-case>annada + SwaroopL R + RakshithGowda G S + SourabhU + ShriramHegde + 28–31 + Parts of speech (POS) tagging is the process of assigning the part of speech tag to each and every word in a sentence. In this paper, we have presented POS tagger for Kannada, a low resource south Asian language, using Condition Random Fields. POS tagger developed in the work uses novel features native to Kannada language. The novel features include Sandhi splitting, where a compound word is broken down into two or more meaningful constituent words. The proposed model is trained and tested on the tagged dataset which contains 21 thousand sentences and achieves a highest accuracy of 94.56%. + R19-2005 + 10.26615/issn.2603-2821.2019_005 + + + Cross-Lingual Coreference: The Case of <fixed-case>B</fixed-case>ulgarian and <fixed-case>E</fixed-case>nglish + ZaraKancheva + 32–38 + The paper presents several common approaches towards cross- and multi-lingual coreference resolution in a search of the most effective practices to be applied within the work on Bulgarian-English manual coreference annotation of a short story. The work aims at outlining the typology of the differences in the annotated parallel texts. The results of the research prove to be comparable with the tendencies observed in similar works on other Slavic languages and show surprising differences between the types of markables and their frequency in Bulgarian and English. + R19-2006 + 10.26615/issn.2603-2821.2019_006 + + + Towards Accurate Text Verbalization for <fixed-case>ASR</fixed-case> Based on Audio Alignment + DianaGeneva + GeorgiShopov + 39–47 + Verbalization of non-lexical linguistic units plays an important role in language modeling for automatic speech recognition systems. Most verbalization methods require valuable resources such as ground truth, large training corpus and expert knowledge which are often unavailable. On the other hand a considerable amount of audio data along with its transcribed text are freely available on the Internet and could be utilized for the task of verbalization. This paper presents a methodology for accurate verbalization of audio transcriptions based on phone-level alignment between the transcriptions and their corresponding audio recordings. Comparing this approach to a more general rule-based verbalization method shows a significant improvement in ASR recognition of non-lexical units. In the process of evaluating this approach we also expose the indirect influence of verbalization accuracy on the quality of acoustic models trained on automatically derived speech corpora. + R19-2007 + 10.26615/issn.2603-2821.2019_007 + + + Evaluation of Stacked Embeddings for <fixed-case>B</fixed-case>ulgarian on the Downstream Tasks <fixed-case>POS</fixed-case> and <fixed-case>NERC</fixed-case> + IvaMarinova + 48–54 + This paper reports on experiments with different stacks of word embeddings and evaluation of their usefulness for Bulgarian downstream tasks such as Named Entity Recognition and Classification (NERC) and Part-of-speech (POS) Tagging. Word embeddings stay in the core of the development of NLP, with several key language models being created over the last two years like FastText (CITATION), ElMo (CITATION), BERT (CITATION) and Flair (CITATION). Stacking or combining different word embeddings is another technique used in this paper and still not reported for Bulgarian NERC. Well-established architecture is used for the sequence tagging task such as BI-LSTM-CRF, and different pre-trained language models are combined in the embedding layer to decide which combination of them scores better. + R19-2008 + 10.26615/issn.2603-2821.2019_008 + + + Overview on <fixed-case>NLP</fixed-case> Techniques for Content-based Recommender Systems for Books + MelaniaBerbatova + 55–61 + Recommender systems are an essential part of today’s largest websites. Without them, it would be hard for users to find the right products and content. One of the most popular methods for recommendations is content-based filtering. It relies on analysing product metadata, a great part of which is textual data. Despite their frequent use, there is still no standard procedure for developing and evaluating content-based recommenders. In this paper, we will first examine current approaches for designing, training and evaluating recommender systems based on textual data for books recommendations for GoodReads’ website. We will give critiques on existing methods and suggest how natural language techniques can be employed for the improvement of content-based recommenders. + R19-2009 + 10.26615/issn.2603-2821.2019_009 + + + Corpora and Processing Tools for Non-standard Contemporary and Diachronic Balkan <fixed-case>S</fixed-case>lavic + TeodoraVukovic + NoraMuheim + OlivierWinistörfer + IvanŠimko + AnastasiaMakarova + SanjaBradjan + 62–68 + The paper describes three corpora of different varieties of BS that are currently being developed with the goal of providing data for the analysis of the diatopic and diachronic variation in non-standard Balkan Slavic. The corpora includes spoken materials from Torlak, Macedonian dialects, as well as the manuscripts of pre-standardized Bulgarian. Apart from the texts, tools for PoS annotation and lemmatization for all varieties are being created, as well as syntactic parsing for Torlak and Bulgarian varieties. The corpora are built using a unified methodology, relying on the pest practices and state-of-the-art methods from the field. The uniform methodology allows the contrastive analysis of the data from different varieties. The corpora under construction can be considered a crucial contribution to the linguistic research on the languages in the Balkans as they provide the lacking data needed for the studies of linguistic variation in the Balkan Slavic, and enable the comparison of the said varieties with other neighbouring languages. + R19-2010 + 10.26615/issn.2603-2821.2019_010 + + + Question Answering Systems Approaches and Challenges + ReemAlqifari + 69–75 + Question answering (QA) systems permit the user to ask a question using natural language, and the system provides a concise and correct answer. QA systems can be implemented for different types of datasets, structured or unstructured. In this paper, some of the recent studies will be reviewed and the limitations will be discussed. Consequently, the current issues are analyzed with the proposed solutions. + R19-2011 + 10.26615/issn.2603-2821.2019_011 + + + Adding Linguistic Knowledge to <fixed-case>NLP</fixed-case> Tasks for <fixed-case>B</fixed-case>ulgarian: The Verb Paradigm Patterns + IvayloRadev + 76–82 + This paper discusses some possible usages of one unexplored lexical language resource containing Bulgarian verb paradigms and their English translations. This type of data can be used for machine translation, generation of pseudo corpora/language exercises, and evaluation of parsers. Upon completion, the resource will be linked with other existing resources such as the morphological lexicon, valency lexicon, as well as BTB-WordNet. + R19-2012 + 10.26615/issn.2603-2821.2019_012 + + + Multilingual Complex Word Identification: Convolutional Neural Networks with Morphological and Linguistic Features + Kim ChengSHEANG + 83–89 + The paper is about our experiments with Complex Word Identification system using deep learning approach with word embeddings and engineered features. + R19-2013 + 10.26615/issn.2603-2821.2019_013 + + + Neural Network-based Models with Commonsense Knowledge for Machine Reading Comprehension + DenisSmirnov + 90–94 + State-of-the-art machine reading comprehension models are capable of producing answers for factual questions about a given piece of text. However, some type of questions requires commonsense knowledge which cannot be inferred from the given text passage. Thus, external semantic information could enhance the performance of these models. This PhD research proposal provides a brief overview of some existing machine reading comprehension datasets and models and outlines possible ways of their improvement. + R19-2014 + 10.26615/issn.2603-2821.2019_014 + +
+
diff --git a/data/xml/W17.xml b/data/xml/W17.xml index 63b56dd26a..c959c8dd17 100644 --- a/data/xml/W17.xml +++ b/data/xml/W17.xml @@ -21,206 +21,206 @@ A Morphological Parser for Odawa - Dustin Bowers - Antti Arppe - Jordan Lachler - Sjur Moshagen - Trond Trosterud + DustinBowers + AnttiArppe + JordanLachler + SjurMoshagen + TrondTrosterud 1–9 W17-0101 10.18653/v1/W17-0101 Creating lexical resources for polysynthetic languages—the case of <fixed-case>A</fixed-case>rapaho - Ghazaleh Kazeminejad - Andrew Cowell - Mans Hulden + GhazalehKazeminejad + AndrewCowell + MansHulden 10–18 W17-0102 10.18653/v1/W17-0102 From Small to Big Data: paper manuscripts to <fixed-case>RDF</fixed-case> triples of <fixed-case>A</fixed-case>ustralian Indigenous Vocabularies - Nick Thieberger - Conal Tuohy + NickThieberger + ConalTuohy 19–23 W17-0103 10.18653/v1/W17-0103 Issues in digital text representation, on-line dissemination, sharing and re-use for <fixed-case>A</fixed-case>frican minority languages - Emmanuel Ngué Um + Emmanuel NguéUm 24–32 W17-0104 10.18653/v1/W17-0104 Developing collection management tools to create more robust and reliable linguistic data - Gary Holton - Kavon Hooshiar - Nick Thieberger + GaryHolton + KavonHooshiar + NickThieberger 33–38 W17-0105 10.18653/v1/W17-0105 <fixed-case>STREAMLI</fixed-case>n<fixed-case>ED</fixed-case> Challenges: Aligning Research Interests with Shared Tasks - Gina-Anne Levow - Emily M. Bender - Patrick Littell - Kristen Howell - Shobhana Chelliah - Joshua Crowgey - Dan Garrette - Jeff Good - Sharon Hargus - David Inman - Michael Maxwell - Michael Tjalve - Fei Xia + Gina-AnneLevow + Emily M.Bender + PatrickLittell + KristenHowell + ShobhanaChelliah + JoshuaCrowgey + DanGarrette + JeffGood + SharonHargus + DavidInman + MichaelMaxwell + MichaelTjalve + FeiXia 39–47 W17-0106 10.18653/v1/W17-0106 Work With What You’ve Got - Lucy Bell - Lawrence Bell + LucyBell + LawrenceBell 48–51 W17-0107 10.18653/v1/W17-0107 Converting a comprehensive lexical database into a computational model: The case of East Cree verb inflection - Antti Arppe - Marie-Odile Junker - Delasie Torkornoo + AnttiArppe + Marie-OdileJunker + DelasieTorkornoo 52–56 W17-0108 10.18653/v1/W17-0108 Instant annotations in <fixed-case>ELAN</fixed-case> corpora of spoken and written <fixed-case>K</fixed-case>omi, an endangered language of the Barents Sea region - Ciprian Gerstenberger - Niko Partanen - Michael Rießler + CiprianGerstenberger + NikoPartanen + MichaelRießler 57–66 W17-0109 10.18653/v1/W17-0109 Inferring Case Systems from <fixed-case>IGT</fixed-case>: Enriching the Enrichment - Kristen Howell - Emily M. Bender - Michel Lockwood - Fei Xia - Olga Zamaraeva + KristenHowell + Emily M.Bender + MichelLockwood + FeiXia + OlgaZamaraeva 67–75 W17-0110 10.18653/v1/W17-0110 Case Studies in the Automatic Characterization of Grammars from Small Wordlists - Jordan Kodner - Spencer Kaplan - Hongzhi Xu - Mitchell P. Marcus - Charles Yang + JordanKodner + SpencerCaplan + HongzhiXu + Mitchell P.Marcus + CharlesYang 76–84 W17-0111 10.18653/v1/W17-0111 Endangered Data for Endangered Languages: Digitizing Print dictionaries - Michael Maxwell - Aric Bills + MichaelMaxwell + AricBills 85–91 W17-0112 10.18653/v1/W17-0112 A computationally-assisted procedure for discovering poetic organization within oral tradition - David Meyer + DavidMeyer 92–100 W17-0113 10.18653/v1/W17-0113 Improving Coverage of an <fixed-case>I</fixed-case>nuktitut Morphological Analyzer Using a Segmental Recurrent Neural Network - Jeffrey Micher + JeffreyMicher 101–106 W17-0114 10.18653/v1/W17-0114 Click reduction in fluent speech: a semi-automated analysis of <fixed-case>M</fixed-case>angetti Dune !<fixed-case>X</fixed-case>ung - Amanda Miller - Micha Elsner + AmandaMiller + MichaElsner 107–115 W17-0115 10.18653/v1/W17-0115 <fixed-case>DECCA</fixed-case> Repurposed: Detecting transcription inconsistencies without an orthographic standard - C. Anton Rytting - Julie Yelle + C. AntonRytting + JulieYelle 116–121 W17-0116 10.18653/v1/W17-0116 Jejueo talking dictionary: A collaborative online database for language revitalization - Moira Saltzman + MoiraSaltzman 122–129 W17-0117 10.18653/v1/W17-0117 Computational Support for Finding Word Classes: A Case Study of Abui - Olga Zamaraeva - František Kratochvíl - Emily M. Bender - Fei Xia - Kristen Howell + OlgaZamaraeva + FrantišekKratochvíl + Emily M.Bender + FeiXia + KristenHowell 130–140 W17-0118 10.18653/v1/W17-0118 Waldayu and Waldayu Mobile: Modern digital dictionary interfaces for endangered languages - Patrick Littell - Aidan Pine - Henry Davis + PatrickLittell + AidanPine + HenryDavis 141–150 W17-0119 10.18653/v1/W17-0119 Connecting Documentation and Revitalization: A New Approach to Language Apps - Alexa N. Little + Alexa N.Little 151–155 W17-0120 10.18653/v1/W17-0120 Developing a Suite of Mobile Applications for Collaborative Language Documentation - Mat Bettinson - Steven Bird + MatBettinson + StevenBird 156–164 W17-0121 10.18653/v1/W17-0121 Cross-language forced alignment to assist community-based linguistics for low resource languages - Timothy Kempton + TimothyKempton 165–169 W17-0122 10.18653/v1/W17-0122 A case study on using speech-to-translation alignments for language documentation - Antonios Anastasopoulos - David Chiang + AntoniosAnastasopoulos + DavidChiang 170–178 W17-0123 10.18653/v1/W17-0123 @@ -242,384 +242,384 @@ Joint <fixed-case>UD</fixed-case> Parsing of <fixed-case>N</fixed-case>orwegian <fixed-case>B</fixed-case>okmål and Nynorsk - Erik Velldal - Lilja Øvrelid - Petter Hohle + ErikVelldal + LiljaØvrelid + PetterHohle 1–10 W17-0201 Replacing <fixed-case>OOV</fixed-case> Words For Dependency Parsing With Distributional Semantics - Prasanth Kolachina - Martin Riedl - Chris Biemann + PrasanthKolachina + MartinRiedl + ChrisBiemann 11–19 W17-0202 Real-valued Syntactic Word Vectors (<fixed-case>RSV</fixed-case>) for Greedy Neural Dependency Parsing - Ali Basirat - Joakim Nivre + AliBasirat + JoakimNivre 20–28 W17-0203 Tagging Named Entities in 19th Century and Modern <fixed-case>F</fixed-case>innish Newspaper Material with a <fixed-case>F</fixed-case>innish Semantic Tagger - Kimmo Kettunen - Laura Löfberg + KimmoKettunen + LauraLöfberg 29–36 W17-0204 Machine Learning for Rhetorical Figure Detection: More Chiasmus with Less Annotation - Marie Dubremetz - Joakim Nivre + MarieDubremetz + JoakimNivre 37–45 W17-0205 Coreference Resolution for <fixed-case>S</fixed-case>wedish and <fixed-case>G</fixed-case>erman using Distant Supervision - Alexander Wallin - Pierre Nugues + AlexanderWallin + PierreNugues 46–55 W17-0206 Aligning phonemes using finte-state methods - Kimmo Koskenniemi + KimmoKoskenniemi 56–64 W17-0207 Acoustic Model Compression with <fixed-case>MAP</fixed-case> adaptation - Katri Leino - Mikko Kurimo + KatriLeino + MikkoKurimo 65–69 W17-0208 <fixed-case>OCR</fixed-case> and post-correction of historical <fixed-case>F</fixed-case>innish texts - Senka Drobac - Pekka Kauppinen - Krister Lindén + SenkaDrobac + PekkaKauppinen + KristerLindén 70–76 W17-0209 Twitter Topic Modeling by Tweet Aggregation - Asbjørn Steinskog - Jonas Therkelsen - Björn Gambäck + AsbjørnSteinskog + JonasTherkelsen + BjörnGambäck 77–86 W17-0210 A Multilingual Entity Linker Using <fixed-case>P</fixed-case>age<fixed-case>R</fixed-case>ank and Semantic Graphs - Anton Södergren - Pierre Nugues + AntonSödergren + PierreNugues 87–95 W17-0211 Linear Ensembles of Word Embedding Models - Avo Muromägi - Kairit Sirts - Sven Laur + AvoMuromägi + KairitSirts + SvenLaur 96–104 W17-0212 Using Pseudowords for Algorithm Comparison: An Evaluation Framework for Graph-based Word Sense Induction - Flavio Massimiliano Cecchini - Chris Biemann - Martin Riedl + FlavioMassimiliano Cecchini + ChrisBiemann + MartinRiedl 105–114 W17-0213 North-Sámi to <fixed-case>F</fixed-case>innish rule-based machine translation system - Tommi Pirinen - Francis M. Tyers - Trond Trosterud - Ryan Johnson - Kevin Unhammer - Tiina Puolakainen + TommiPirinen + Francis M.Tyers + TrondTrosterud + RyanJohnson + KevinUnhammer + TiinaPuolakainen 115–122 W17-0214 Machine translation with North Saami as a pivot language - Lene Antonsen - Ciprian Gerstenberger - Maja Kappfjell - Sandra Nystø Rahka - Marja-Liisa Olthuis - Trond Trosterud - Francis M. Tyers + LeneAntonsen + CiprianGerstenberger + MajaKappfjell + SandraNystø Rahka + Marja-LiisaOlthuis + TrondTrosterud + Francis M.Tyers 123–131 W17-0215 <fixed-case>SWEGRAM</fixed-case> – A Web-Based Tool for Automatic Annotation and Analysis of <fixed-case>S</fixed-case>wedish Texts - Jesper Näsman - Beáta Megyesi - Anne Palmér + JesperNäsman + BeátaMegyesi + AnnePalmér 132–141 W17-0216 Optimizing a <fixed-case>P</fixed-case>o<fixed-case>S</fixed-case> Tagset for <fixed-case>N</fixed-case>orwegian Dependency Parsing - Petter Hohle - Lilja Øvrelid - Erik Velldal + PetterHohle + LiljaØvrelid + ErikVelldal 142–151 W17-0217 Creating register sub-corpora for the <fixed-case>F</fixed-case>innish <fixed-case>I</fixed-case>nternet Parsebank - Veronika Laippala - Juhani Luotolahti - Aki-Juhani Kyröläinen - Tapio Salakoski - Filip Ginter + VeronikaLaippala + JuhaniLuotolahti + Aki-JuhaniKyröläinen + TapioSalakoski + FilipGinter 152–161 W17-0218 <fixed-case>KILLE</fixed-case>: a Framework for Situated Agents for Learning Language Through Interaction - Simon Dobnik - Erik de Graaf + SimonDobnik + Erikde Graaf 162–171 W17-0219 Data Collection from Persons with Mild Forms of Cognitive Impairment and Healthy Controls - Infrastructure for Classification and Prediction of Dementia - Dimitrios Kokkinakis - Kristina Lundholm Fors - Eva Björkner - Arto Nordlund + DimitriosKokkinakis + KristinaLundholm Fors + EvaBjörkner + ArtoNordlund 172–182 W17-0220 Evaluation of language identification methods using 285 languages - Tommi Jauhiainen - Krister Lindén - Heidi Jauhiainen + TommiJauhiainen + KristerLindén + HeidiJauhiainen 183–191 W17-0221 Can We Create a Tool for General Domain Event Analysis? - Siim Orasmaa - Heiki-Jaan Kaalep + SiimOrasmaa + Heiki-JaanKaalep 192–201 W17-0222 From Treebank to <fixed-case>P</fixed-case>ropbank: A Semantic-Role and <fixed-case>V</fixed-case>erb<fixed-case>N</fixed-case>et Corpus for <fixed-case>D</fixed-case>anish - Eckhard Bick + EckhardBick 202–210 W17-0223 Cross-lingual Learning of Semantic Textual Similarity with Multilingual Word Representations - Johannes Bjerva - Robert Östling + JohannesBjerva + RobertÖstling 211–215 W17-0224 Will my auxiliary tagging task help? Estimating Auxiliary Tasks Effectivity in Multi-Task Learning - Johannes Bjerva + JohannesBjerva 216–220 W17-0225 Iconic Locations in <fixed-case>S</fixed-case>wedish Sign Language: Mapping Form to Meaning with Lexical Databases - Carl Börstell - Robert Östling + CarlBörstell + RobertÖstling 221–225 W17-0226 <fixed-case>D</fixed-case>ocforia: A Multilayer Document Model - Marcus Klang - Pierre Nugues + MarcusKlang + PierreNugues 226–230 W17-0227 <fixed-case>F</fixed-case>innish resources for evaluating language model semantics - Viljami Venekoski - Jouko Vankka + ViljamiVenekoski + JoukoVankka 231–236 W17-0228 <fixed-case>M</fixed-case>álrómur: A Manually Verified Corpus of Recorded <fixed-case>I</fixed-case>celandic Speech - Steinþór Steingrímsson - Jón Guðnason - Sigrún Helgadóttir - Eiríkur Rögnvaldsson + SteinþórSteingrímsson + JónGuðnason + SigrúnHelgadóttir + EiríkurRögnvaldsson 237–240 W17-0229 The Effect of Translationese on Tuning for Statistical Machine Translation - Sara Stymne + SaraStymne 241–246 W17-0230 Multilingwis² – Explore Your Parallel Corpus - Johannes Graën - Dominique Sandoz - Martin Volk + JohannesGraën + DominiqueSandoz + MartinVolk 247–250 W17-0231 A modernised version of the Glossa corpus search system - Anders Nøklestad - Kristin Hagen - Janne Bondi Johannessen - Michał Kosek - Joel Priestley + AndersNøklestad + KristinHagen + JanneBondi Johannessen + MichałKosek + JoelPriestley 251–254 W17-0232 <fixed-case>D</fixed-case>ep_search: Efficient Search Tool for Large Dependency Parsebanks - Juhani Luotolahti - Jenna Kanerva - Filip Ginter + JuhaniLuotolahti + JennaKanerva + FilipGinter 255–258 W17-0233 Proto-<fixed-case>I</fixed-case>ndo-<fixed-case>E</fixed-case>uropean Lexicon: The Generative Etymological Dictionary of <fixed-case>I</fixed-case>ndo-<fixed-case>E</fixed-case>uropean Languages - Jouna Pyysalo + JounaPyysalo 259–262 W17-0234 Tilde <fixed-case>MODEL</fixed-case> - Multilingual Open Data for <fixed-case>EU</fixed-case> Languages - Roberts Rozis - Raivis Skadiņš + RobertsRozis + RaivisSkadiņš 263–265 W17-0235 Mainstreaming August Strindberg with Text Normalization - Adam Ek - Sofia Knuutinen + AdamEk + SofiaKnuutinen 266–270 W17-0236 Word vectors, reuse, and replicability: Towards a community repository of large-text resources - Murhaf Fares - Andrey Kutuzov - Stephan Oepen - Erik Velldal + MurhafFares + AndreyKutuzov + StephanOepen + ErikVelldal 271–276 W17-0237 Improving Optical Character Recognition of <fixed-case>F</fixed-case>innish Historical Newspapers with a Combination of Fraktur & Antiqua Models and Image Preprocessing - Mika Koistinen - Kimmo Kettunen - Tuula Pääkkönen + MikaKoistinen + KimmoKettunen + TuulaPääkkönen 277–283 W17-0238 Redefining Context Windows for Word Embedding Models: An Experimental Study - Pierre Lison - Andrey Kutuzov + PierreLison + AndreyKutuzov 284–288 W17-0239 The Effect of Excluding Out of Domain Training Data from Supervised Named-Entity Recognition - Adam Persson + AdamPersson 289–292 W17-0240 Quote Extraction and Attribution from <fixed-case>N</fixed-case>orwegian Newspapers - Andrew Salway - Paul Meurer - Knut Hofland - Øystein Reigem + AndrewSalway + PaulMeurer + KnutHofland + ØysteinReigem 293–297 W17-0241 <fixed-case>W</fixed-case>ordnet extension via word embeddings: Experiments on the <fixed-case>N</fixed-case>orwegian <fixed-case>W</fixed-case>ordnet - Heidi Sand - Erik Velldal - Lilja Øvrelid + HeidiSand + ErikVelldal + LiljaØvrelid 298–302 W17-0242 Universal Dependencies for <fixed-case>S</fixed-case>wedish Sign Language - Robert Östling - Carl Börstell - Moa Gärdenfors - Mats Wirén + RobertÖstling + CarlBörstell + MoaGärdenfors + MatsWirén 303–308 W17-0243 Services for text simplification and analysis - Johan Falkenjack - Evelina Rennes - Daniel Fahlborg - Vida Johansson - Arne Jönsson + JohanFalkenjack + EvelinaRennes + DanielFahlborg + VidaJohansson + ArneJönsson 309–313 W17-0244 Exploring Properties of Intralingual and Interlingual Association Measures Visually - Johannes Graën - Christof Bless + JohannesGraën + ChristofBless 314–317 W17-0245 <fixed-case>TALERUM</fixed-case> - Learning <fixed-case>D</fixed-case>anish by Doing <fixed-case>D</fixed-case>anish - Peter Juel Henrichsen + PeterJuel Henrichsen 318–321 W17-0246 Cross-Lingual Syntax: Relating Grammatical Framework with Universal Dependencies - Aarne Ranta - Prasanth Kolachina - Thomas Hallgren + AarneRanta + PrasanthKolachina + ThomasHallgren 322–325 W17-0247 Exploring Treebanks with <fixed-case>INESS</fixed-case> Search - Victoria Rosén - Helge Dyvik - Paul Meurer - Koenraad De Smedt + VictoriaRosén + HelgeDyvik + PaulMeurer + KoenraadDe Smedt 326–329 W17-0248 A System for Identifying and Exploring Text Repetition in Large Historical Document Corpora - Aleksi Vesanto - Filip Ginter - Hannu Salmi - Asko Nivala - Tapio Salakoski + AleksiVesanto + FilipGinter + HannuSalmi + AskoNivala + TapioSalakoski 330–333 W17-0249 @@ -642,56 +642,56 @@ Learning with learner corpora: Using the <fixed-case>TLE</fixed-case> for native language identification - Allison Adams - Sara Stymne + AllisonAdams + SaraStymne 1-7 W17-0301 Challenging learners in their individual zone of proximal development using pedagogic developmental benchmarks of syntactic complexity - Xiaobin Chen - Detmar Meurers + XiaobinChen + DetmarMeurers 8-17 W17-0302 Crossing the border twice: Reimporting prepositions to alleviate <fixed-case>L</fixed-case>1-specific transfer errors - Johannes Graën - Gerold Schneider + JohannesGraën + GeroldSchneider 18-26 W17-0303 <fixed-case>R</fixed-case>evita: a system for language learning and supporting endangered languages - Anisia Katinskaia - Javad Nouri - Roman Yangarber + AnisiaKatinskaia + JavadNouri + RomanYangarber 27-35 W17-0304 Developing a web-based workbook for <fixed-case>E</fixed-case>nglish supporting the interaction of students and teachers - Björn Rudzewitz - Ramon Ziai - Kordula De Kuthy - Detmar Meurers + BjörnRudzewitz + RamonZiai + KordulaDe Kuthy + DetmarMeurers 36-46 W17-0305 Annotating errors in student texts: First experiences and experiments - Sara Stymne - Eva Pettersson - Beáta Megyesi - Anne Palmér + SaraStymne + EvaPettersson + BeátaMegyesi + AnnePalmér 47-60 W17-0306 Building and using language resources and infrastructure to develop e-learning programs for a minority language - Heli Uibo - Jack Rueter - Sulev Iva + HeliUibo + JackRueter + SulevIva 61-67 W17-0307 @@ -713,138 +713,138 @@ Cross-Lingual Parser Selection for Low-Resource Languages - Željko Agić + ŽeljkoAgić 1–10 W17-0401 <fixed-case>S</fixed-case>wedish Prepositions are not Pure Function Words - Lars Ahrenberg + LarsAhrenberg 11–18 W17-0402 Increasing Return on Annotation Investment: The Automatic Construction of a Universal Dependency Treebank for <fixed-case>D</fixed-case>utch - Gosse Bouma - Gertjan van Noord + GosseBouma + Gertjanvan Noord 19–26 W17-0403 Converting the <fixed-case>T</fixed-case>ü<fixed-case>B</fixed-case>a-D/Z Treebank of <fixed-case>G</fixed-case>erman to Universal Dependencies - Çağrı Çöltekin - Ben Campbell - Erhard Hinrichs - Heike Telljohann + ÇağrıÇöltekin + BenCampbell + ErhardHinrichs + HeikeTelljohann 27–37 W17-0404 Universal Dependencies for <fixed-case>A</fixed-case>frikaans - Peter Dirix - Liesbeth Augustinus - Daniel van Niekerk - Frank Van Eynde + PeterDirix + LiesbethAugustinus + Danielvan Niekerk + FrankVan Eynde 38–47 W17-0405 Elliptic Constructions: Spotting Patterns in <fixed-case>UD</fixed-case> Treebanks - Kira Droganova - Daniel Zeman + KiraDroganova + DanielZeman 48–57 W17-0406 Dependency Tree Transformation with Tree Transducers - Felix Hennig - Arne Köhn + FelixHennig + ArneKöhn 58–66 W17-0407 Towards Universal Dependencies for Learner <fixed-case>C</fixed-case>hinese - John Lee - Herman Leung - Keying Li + JohnLee + HermanLeung + KeyingLi 67–71 W17-0408 Does Syntactic Informativity Predict Word Length? A Cross-Linguistic Study Based on the Universal Dependencies Corpora - Natalia Levshina + NataliaLevshina 72–78 W17-0409 <fixed-case>E</fixed-case>stonian Copular and Existential Constructions as an <fixed-case>UD</fixed-case> Annotation Problem - Kadri Muischnek - Kaili Müürisep + KadriMuischnek + KailiMüürisep 79–85 W17-0410 Universal Dependency Evaluation - Joakim Nivre - Chiao-Ting Fang + JoakimNivre + Chiao-TingFang 86–95 W17-0411 <fixed-case>U</fixed-case>dapi: Universal <fixed-case>API</fixed-case> for Universal Dependencies - Martin Popel - Zdeněk Žabokrtský - Martin Vojtek + MartinPopel + ZdeněkŽabokrtský + MartinVojtek 96–101 W17-0412 Universal Dependencies for <fixed-case>G</fixed-case>reek - Prokopis Prokopidis - Haris Papageorgiou + ProkopisProkopidis + HarisPapageorgiou 102–106 W17-0413 From Universal Dependencies to Abstract Syntax - Aarne Ranta - Prasanth Kolachina + AarneRanta + PrasanthKolachina 107–116 W17-0414 Empirically Sampling Universal Dependencies - Natalie Schluter - Željko Agić + NatalieSchluter + ŽeljkoAgić 117–122 W17-0415 Gapping Constructions in Universal Dependencies v2 - Sebastian Schuster - Matthew Lamm - Christopher D. Manning + SebastianSchuster + MatthewLamm + Christopher D.Manning 123–132 W17-0416 Toward Universal Dependencies for <fixed-case>A</fixed-case>inu - Hajime Senuma - Akiko Aizawa + HajimeSenuma + AkikoAizawa 133–139 W17-0417 Automatic Morpheme Segmentation and Labeling in Universal Dependencies Resources - Miikka Silfverberg - Mans Hulden + MiikkaSilfverberg + MansHulden 140–145 W17-0418 A Systematic Comparison of Syntactic Representations of Dependency Parsing - Guillaume Wisniewski - Ophélie Lacroix + GuillaumeWisniewski + OphélieLacroix 146–152 W17-0419 @@ -865,84 +865,84 @@ Variance in Historical Data: How bad is it and how can we profit from it for historical linguistics? - Stefanie Dipper + StefanieDipper 1–1 W17-0501 Improving <fixed-case>POS</fixed-case> Tagging in Old <fixed-case>S</fixed-case>panish Using <fixed-case>TEITOK</fixed-case> - Maarten Janssen - Josep Ausensi - Josep Fontana + MaartenJanssen + JosepAusensi + JosepFontana 2–6 W17-0502 The Making of the Royal Society Corpus - Jörg Knappen - Stefan Fischer - Hannah Kermes - Elke Teich - Peter Fankhauser + JörgKnappen + StefanFischer + HannahKermes + ElkeTeich + PeterFankhauser 7–11 W17-0503 Normalizing Medieval <fixed-case>G</fixed-case>erman Texts: from rules to deep learning - Natalia Korchagina + NataliaKorchagina 12–17 W17-0504 Ambiguity in Semantically Related Word Substitutions: an investigation in historical <fixed-case>B</fixed-case>ible translations - Maria Moritz - Marco Büchler + MariaMoritz + MarcoBüchler 18–23 W17-0505 The Lemlat 3.0 Package for Morphological Analysis of <fixed-case>L</fixed-case>atin - Marco Passarotti - Marco Budassi - Eleonora Litta - Paolo Ruffolo + MarcoPassarotti + MarcoBudassi + EleonoraLitta + PaoloRuffolo 24–31 W17-0506 <fixed-case>H</fixed-case>isto<fixed-case>B</fixed-case>ank<fixed-case>V</fixed-case>is: Detecting Language Change via Data Visualization - Christin Schätzle - Michael Hund - Frederik Dennig - Miriam Butt - Daniel Keim + ChristinSchätzle + MichaelHund + FrederikDennig + MiriamButt + DanielKeim 32–39 W17-0507 Comparing Rule-based and <fixed-case>SMT</fixed-case>-based Spelling Normalisation for <fixed-case>E</fixed-case>nglish Historical Texts - Gerold Schneider - Eva Pettersson - Michael Percillier + GeroldSchneider + EvaPettersson + MichaelPercillier 40–46 W17-0508 Data-driven Morphology and Sociolinguistics for Early Modern <fixed-case>D</fixed-case>utch - Marijn Schraagen - Marjo van Koppen - Feike Dietz + MarijnSchraagen + Marjovan Koppen + FeikeDietz 47–53 W17-0509 Applying <fixed-case>BLAST</fixed-case> to Text Reuse Detection in <fixed-case>F</fixed-case>innish Newspapers and Journals, 1771-1910 - Aleksi Vesanto - Asko Nivala - Heli Rantala - Tapio Salakoski - Hannu Salmi - Filip Ginter + AleksiVesanto + AskoNivala + HeliRantala + TapioSalakoski + HannuSalmi + FilipGinter 54–58 W17-0510 @@ -966,71 +966,71 @@ Synchronized Mediawiki based analyzer dictionary development - Jack Rueter - Mika Hämäläinen + JackRueter + MikaHämäläinen 1–7 W17-0601 10.18653/v1/W17-0601 <fixed-case>DEMO</fixed-case>: Giellatekno Open-source click-in-text dictionaries for bringing closely related languages into contact. - Jack Rueter + JackRueter 8–9 W17-0602 10.18653/v1/W17-0602 Languages under the influence: Building a database of Uralic languages - Eszter Simon - Nikolett Mus + EszterSimon + NikolettMus 10–24 W17-0603 10.18653/v1/W17-0603 Instant Annotations – Applying <fixed-case>NLP</fixed-case> Methods to the Annotation of Spoken Language Documentation Corpora - Ciprian Gerstenberger - Niko Partanen - Michael Rießler - Joshua Wilbur + CiprianGerstenberger + NikoPartanen + MichaelRießler + JoshuaWilbur 25–36 W17-0604 10.18653/v1/W17-0604 Preliminary Experiments concerning Verbal Predicative Structure Extraction from a Large <fixed-case>F</fixed-case>innish Corpus - Guersande Chaminade - Thierry Poibeau + GuersandeChaminade + ThierryPoibeau 37–55 W17-0605 10.18653/v1/W17-0605 Language technology resources and tools for Mansi: an overview - Csilla Horváth - Norbert Szilágyi - Veronika Vincze - Ágoston Nagy + CsillaHorváth + NorbertSzilágyi + VeronikaVincze + ÁgostonNagy 56–65 W17-0606 10.18653/v1/W17-0606 Annotation schemes in North Sámi dependency parsing - Francis M. Tyers - Mariya Sheyanova + Francis M.Tyers + MariyaSheyanova 66–75 W17-0607 10.18653/v1/W17-0607 A morphological analyser for Kven - Sindre Reino Trosterud - Trond Trosterud - Anna-Kaisa Räisänen - Leena Niiranen - Mervi Haavisto - Kaisa Maliniemi + SindreReino Trosterud + TrondTrosterud + Anna-KaisaRäisänen + LeenaNiiranen + MerviHaavisto + KaisaMaliniemi 76–88 W17-0608 10.18653/v1/W17-0608 @@ -1055,10 +1055,10 @@ Entropy Reduction correlates with temporal lobe activity - Matthew Nelson - Stanislas Dehaene - Christophe Pallier - John Hale + MatthewNelson + StanislasDehaene + ChristophePallier + JohnHale 1–10 W17-0701 10.18653/v1/W17-0701 @@ -1066,9 +1066,9 @@ Learning an Input Filter for Argument Structure Acquisition - Laurel Perkins - Naomi Feldman - Jeffrey Lidz + LaurelPerkins + NaomiFeldman + JeffreyLidz 11–19 W17-0702 10.18653/v1/W17-0702 @@ -1076,8 +1076,8 @@ Grounding sound change in ideal observer models of perception - Zachary Burchill - T. Florian Jaeger + ZacharyBurchill + T. FlorianJaeger 20–28 W17-0703 10.18653/v1/W17-0703 @@ -1085,7 +1085,7 @@ “Oh, <fixed-case>I</fixed-case>’ve Heard That Before”: Modelling Own-Dialect Bias After Perceptual Learning by Weighting Training Data - Rachael Tatman + RachaelTatman 29–34 W17-0704 10.18653/v1/W17-0704 @@ -1093,7 +1093,7 @@ Inherent Biases of Recurrent Neural Networks for Phonological Assimilation and Dissimilation - Amanda Doucette + AmandaDoucette 35–40 W17-0705 10.18653/v1/W17-0705 @@ -1101,7 +1101,7 @@ Predicting <fixed-case>J</fixed-case>apanese scrambling in the wild - Naho Orita + NahoOrita 41–45 W17-0706 10.18653/v1/W17-0706 @@ -1125,8 +1125,8 @@ Readers vs. Writers vs. Texts: Coping with Different Perspectives of Text Understanding in Emotion Annotation - Sven Buechel - Udo Hahn + SvenBuechel + UdoHahn 1–12 W17-0801 10.18653/v1/W17-0801 @@ -1134,11 +1134,11 @@ Finding Good Conversations Online: The Yahoo News Annotated Comments Corpus - Courtney Napoles - Joel Tetreault - Aasish Pappu - Enrica Rosato - Brian Provenzale + CourtneyNapoles + JoelTetreault + AasishPappu + EnricaRosato + BrianProvenzale 13–23 W17-0802 10.18653/v1/W17-0802 @@ -1146,8 +1146,8 @@ Crowdsourcing discourse interpretations: On the influence of context and the reliability of a connective insertion task - Merel Scholman - Vera Demberg + MerelScholman + VeraDemberg 24–33 W17-0803 10.18653/v1/W17-0803 @@ -1155,7 +1155,7 @@ A Code-Switching Corpus of <fixed-case>T</fixed-case>urkish-<fixed-case>G</fixed-case>erman Conversations - Özlem Çetinoğlu + ÖzlemÇetinoğlu 34–40 W17-0804 10.18653/v1/W17-0804 @@ -1163,9 +1163,9 @@ Annotating omission in statement pairs - Héctor Martínez Alonso - Amaury Delamaire - Benoît Sagot + HéctorMartínez Alonso + AmauryDelamaire + BenoîtSagot 41–45 W17-0805 10.18653/v1/W17-0805 @@ -1173,11 +1173,11 @@ Annotating Speech, Attitude and Perception Reports - Corien Bary - Leopold Hess - Kees Thijs - Peter Berck - Iris Hendrickx + CorienBary + LeopoldHess + KeesThijs + PeterBerck + IrisHendrickx 46–56 W17-0806 10.18653/v1/W17-0806 @@ -1185,12 +1185,12 @@ Consistent Classification of Translation Revisions: A Case Study of <fixed-case>E</fixed-case>nglish-<fixed-case>J</fixed-case>apanese Student Translations - Atsushi Fujita - Kikuko Tanabe - Chiho Toyoshima - Mayuka Yamamoto - Kyo Kageura - Anthony Hartley + AtsushiFujita + KikukoTanabe + ChihoToyoshima + MayukaYamamoto + KyoKageura + AnthonyHartley 57–66 W17-0807 10.18653/v1/W17-0807 @@ -1198,13 +1198,13 @@ Representation and Interchange of Linguistic Annotation. An In-Depth, Side-by-Side Comparison of Three Designs - Richard Eckart de Castilho - Nancy Ide - Emanuele Lapponi - Stephan Oepen - Keith Suderman - Erik Velldal - Marc Verhagen + RichardEckart de Castilho + NancyIde + EmanueleLapponi + StephanOepen + KeithSuderman + ErikVelldal + MarcVerhagen 67–75 W17-0808 10.18653/v1/W17-0808 @@ -1212,8 +1212,8 @@ <fixed-case>TDB</fixed-case> 1.1: Extensions on <fixed-case>T</fixed-case>urkish Discourse Bank - Deniz Zeyrek - Murathan Kurfalı + DenizZeyrek + MurathanKurfalı 76–81 W17-0809 10.18653/v1/W17-0809 @@ -1221,12 +1221,12 @@ Two Layers of Annotation for Representing Event Mentions in News Stories - Maria Pia di Buono - Martin Tutek - Jan Šnajder - Goran Glavaš - Bojana Dalbelo Bašić - Nataša Milić-Frayling + Maria Piadi Buono + MartinTutek + JanŠnajder + GoranGlavaš + BojanaDalbelo Bašić + NatašaMilić-Frayling 82–90 W17-0810 10.18653/v1/W17-0810 @@ -1234,11 +1234,11 @@ Word Similarity Datasets for <fixed-case>I</fixed-case>ndian Languages: Annotation and Baseline Systems - Syed Sarfaraz Akhtar - Arihant Gupta - Avijit Vajpayee - Arjit Srivastava - Manish Shrivastava + Syed SarfarazAkhtar + ArihantGupta + AvijitVajpayee + ArjitSrivastava + ManishShrivastava 91–94 W17-0811 10.18653/v1/W17-0811 @@ -1246,9 +1246,9 @@ The <fixed-case>BEC</fixed-case>au<fixed-case>SE</fixed-case> Corpus 2.0: Annotating Causality and Overlapping Relations - Jesse Dunietz - Lori Levin - Jaime Carbonell + JesseDunietz + LoriLevin + JaimeCarbonell 95–104 W17-0812 10.18653/v1/W17-0812 @@ -1256,8 +1256,8 @@ Catching the Common Cause: Extraction and Annotation of Causal Relations and their Participants - Ines Rehbein - Josef Ruppenhofer + InesRehbein + JosefRuppenhofer 105–114 W17-0813 10.18653/v1/W17-0813 @@ -1265,11 +1265,11 @@ Assessing <fixed-case>SRL</fixed-case> Frameworks with Automatic Training Data Expansion - Silvana Hartmann - Éva Mújdricza-Maydt - Ilia Kuznetsov - Iryna Gurevych - Anette Frank + SilvanaHartmann + ÉvaMújdricza-Maydt + IliaKuznetsov + IrynaGurevych + AnetteFrank 115–121 W17-0814 10.18653/v1/W17-0814 @@ -1295,10 +1295,10 @@ Inducing Script Structure from Crowdsourced Event Descriptions via Semi-Supervised Clustering - Lilian Wanzare - Alessandra Zarcone - Stefan Thater - Manfred Pinkal + LilianWanzare + AlessandraZarcone + StefanThater + ManfredPinkal 1–11 W17-0901 10.18653/v1/W17-0901 @@ -1306,16 +1306,16 @@ A Consolidated Open Knowledge Representation for Multiple Texts - Rachel Wities - Vered Shwartz - Gabriel Stanovsky - Meni Adler - Ori Shapira - Shyam Upadhyay - Dan Roth - Eugenio Martinez Camara - Iryna Gurevych - Ido Dagan + RachelWities + VeredShwartz + GabrielStanovsky + MeniAdler + OriShapira + ShyamUpadhyay + DanRoth + EugenioMartinez Camara + IrynaGurevych + IdoDagan 12–24 W17-0902 10.18653/v1/W17-0902 @@ -1323,8 +1323,8 @@ Event-Related Features in Feedforward Neural Networks Contribute to Identifying Causal Relations in Discourse - Edoardo Maria Ponti - Anna Korhonen + Edoardo MariaPonti + AnnaKorhonen 25–30 W17-0903 10.18653/v1/W17-0903 @@ -1332,9 +1332,9 @@ Stance Detection in <fixed-case>F</fixed-case>acebook Posts of a <fixed-case>G</fixed-case>erman Right-wing Party - Manfred Klenner - Don Tuggener - Simon Clematide + ManfredKlenner + DonTuggener + SimonClematide 31–40 W17-0904 10.18653/v1/W17-0904 @@ -1342,7 +1342,7 @@ Behind the Scenes of an Evolving Event Cloze Test - Nathanael Chambers + NathanaelChambers 41–45 W17-0905 10.18653/v1/W17-0905 @@ -1350,11 +1350,11 @@ <fixed-case>LSDS</fixed-case>em 2017 Shared Task: The Story Cloze Test - Nasrin Mostafazadeh - Michael Roth - Annie Louis - Nathanael Chambers - James Allen + NasrinMostafazadeh + MichaelRoth + AnnieLouis + NathanaelChambers + JamesAllen 46–51 W17-0906 10.18653/v1/W17-0906 @@ -1362,12 +1362,12 @@ Story Cloze Task: <fixed-case>UW</fixed-case> <fixed-case>NLP</fixed-case> System - Roy Schwartz - Maarten Sap - Ioannis Konstas - Leila Zilles - Yejin Choi - Noah A. Smith + RoySchwartz + MaartenSap + IoannisKonstas + LeilaZilles + YejinChoi + Noah A.Smith 52–55 W17-0907 10.18653/v1/W17-0907 @@ -1375,15 +1375,15 @@ <fixed-case>LSDS</fixed-case>em 2017: Exploring Data Generation Methods for the Story Cloze Test - Michael Bugert - Yevgeniy Puzikov - Andreas Rücklé - Judith Eckle-Kohler - Teresa Martin - Eugenio Martínez-Cámara - Daniil Sorokin - Maxime Peyrard - Iryna Gurevych + MichaelBugert + YevgeniyPuzikov + AndreasRücklé + JudithEckle-Kohler + TeresaMartin + EugenioMartínez-Cámara + DaniilSorokin + MaximePeyrard + IrynaGurevych 56–61 W17-0908 10.18653/v1/W17-0908 @@ -1391,8 +1391,8 @@ Sentiment Analysis and Lexical Cohesion for the Story Cloze Task - Michael Flor - Swapna Somasundaran + MichaelFlor + SwapnaSomasundaran 62–67 W17-0909 10.18653/v1/W17-0909 @@ -1400,8 +1400,8 @@ Resource-Lean Modeling of Coherence in Commonsense Stories - Niko Schenk - Christian Chiarcos + NikoSchenk + ChristianChiarcos 68–73 W17-0910 10.18653/v1/W17-0910 @@ -1409,10 +1409,10 @@ An <fixed-case>RNN</fixed-case>-based Binary Classifier for the Story Cloze Test - Melissa Roemmele - Sosuke Kobayashi - Naoya Inoue - Andrew Gordon + MelissaRoemmele + SosukeKobayashi + NaoyaInoue + AndrewGordon 74–80 W17-0911 10.18653/v1/W17-0911 @@ -1420,8 +1420,8 @@ IIT (BHU): System Description for LSDSem’17 Shared Task - Pranav Goel - Anil Kumar Singh + PranavGoel + Anil KumarSingh 81–86 W17-0912 10.18653/v1/W17-0912 @@ -1429,8 +1429,8 @@ Story Cloze Ending Selection Baselines and Data Examination - Todor Mihaylov - Anette Frank + TodorMihaylov + AnetteFrank 87–92 W17-0913 10.18653/v1/W17-0913 @@ -1459,14 +1459,14 @@ <fixed-case>M</fixed-case>ulti<fixed-case>L</fixed-case>ing 2017 Overview - George Giannakopoulos - John Conroy - Jeff Kubina - Peter A. Rankel - Elena Lloret - Josef Steinberger - Marina Litvak - Benoit Favre + GeorgeGiannakopoulos + JohnConroy + JeffKubina + Peter A.Rankel + ElenaLloret + JosefSteinberger + MarinaLitvak + BenoitFavre 1–6 W17-1001 10.18653/v1/W17-1001 @@ -1474,10 +1474,10 @@ Decoupling Encoder and Decoder Networks for Abstractive Document Summarization - Ying Xu - Jey Han Lau - Timothy Baldwin - Trevor Cohn + YingXu + Jey HanLau + TimothyBaldwin + TrevorCohn 7–11 W17-1002 10.18653/v1/W17-1002 @@ -1485,9 +1485,9 @@ Centroid-based Text Summarization through Compositionality of Word Embeddings - Gaetano Rossiello - Pierpaolo Basile - Giovanni Semeraro + GaetanoRossiello + PierpaoloBasile + GiovanniSemeraro 12–21 W17-1003 10.18653/v1/W17-1003 @@ -1495,8 +1495,8 @@ Query-based summarization using <fixed-case>MDL</fixed-case> principle - Marina Litvak - Natalia Vanetik + MarinaLitvak + NataliaVanetik 22–31 W17-1004 10.18653/v1/W17-1004 @@ -1504,9 +1504,9 @@ Word Embedding and Topic Modeling Enhanced Multiple Features for Content Linking and Argument / Sentiment Labeling in Online Forums - Lei Li - Liyuan Mao - Moye Chen + LeiLi + LiyuanMao + MoyeChen 32–36 W17-1005 10.18653/v1/W17-1005 @@ -1514,10 +1514,10 @@ Ultra-Concise Multi-genre Summarisation of Web2.0: towards Intelligent Content Generation - Elena Lloret - Ester Boldrini - Patricio Martínez-Barco - Manuel Palomar + ElenaLloret + EsterBoldrini + PatricioMartínez-Barco + ManuelPalomar 37–46 W17-1006 10.18653/v1/W17-1006 @@ -1525,9 +1525,9 @@ Machine Learning Approach to Evaluate <fixed-case>M</fixed-case>ulti<fixed-case>L</fixed-case>ingual Summaries - Samira Ellouze - Maher Jaoua - Lamia Hadrich Belguith + SamiraEllouze + MaherJaoua + LamiaHadrich Belguith 47–54 W17-1007 10.18653/v1/W17-1007 @@ -1551,8 +1551,8 @@ A Survey on Hate Speech Detection using Natural Language Processing - Anna Schmidt - Michael Wiegand + AnnaSchmidt + MichaelWiegand 1–10 W17-1101 10.18653/v1/W17-1101 @@ -1560,11 +1560,11 @@ <fixed-case>F</fixed-case>acebook sentiment: Reactions and Emojis - Ye Tian - Thiago Galery - Giulio Dulcinati - Emilia Molimpakis - Chao Sun + YeTian + ThiagoGalery + GiulioDulcinati + EmiliaMolimpakis + ChaoSun 11–16 W17-1102 10.18653/v1/W17-1102 @@ -1572,10 +1572,10 @@ Potential and Limitations of Cross-Domain Sentiment Classification - Jan Milan Deriu - Martin Weilenmann - Dirk Von Gruenigen - Mark Cieliebak + Jan MilanDeriu + MartinWeilenmann + DirkVon Gruenigen + MarkCieliebak 17–24 W17-1103 10.18653/v1/W17-1103 @@ -1583,10 +1583,10 @@ Aligning Entity Names with Online Aliases on Twitter - Kevin McKelvey - Peter Goutzounis - Stephen da Cruz - Nathanael Chambers + KevinMcKelvey + PeterGoutzounis + Stephenda Cruz + NathanaelChambers 25–35 W17-1104 10.18653/v1/W17-1104 @@ -1594,9 +1594,9 @@ Character-based Neural Embeddings for Tweet Clustering - Svitlana Vakulenko - Lyndon Nixon - Mihai Lupu + SvitlanaVakulenko + LyndonNixon + MihaiLupu 36–44 W17-1105 10.18653/v1/W17-1105 @@ -1605,10 +1605,10 @@ A Twitter Corpus and Benchmark Resources for <fixed-case>G</fixed-case>erman Sentiment Analysis - Mark Cieliebak - Jan Milan Deriu - Dominic Egger - Fatih Uzdilli + MarkCieliebak + Jan MilanDeriu + DominicEgger + FatihUzdilli 45–51 W17-1106 10.18653/v1/W17-1106 @@ -1636,14 +1636,14 @@ Findings of the <fixed-case>V</fixed-case>ar<fixed-case>D</fixed-case>ial Evaluation Campaign 2017 - Marcos Zampieri - Shervin Malmasi - Nikola Ljubešić - Preslav Nakov - Ahmed Ali - Jörg Tiedemann - Yves Scherrer - Noëmi Aepli + MarcosZampieri + ShervinMalmasi + NikolaLjubešić + PreslavNakov + AhmedAli + JörgTiedemann + YvesScherrer + NoëmiAepli 1–15 W17-1201 10.18653/v1/W17-1201 @@ -1651,8 +1651,8 @@ Dialectometric analysis of language variation in Twitter - Gonzalo Donoso - David Sánchez + GonzaloDonoso + DavidSánchez 16–25 W17-1202 10.18653/v1/W17-1202 @@ -1660,9 +1660,9 @@ Computational analysis of <fixed-case>G</fixed-case>ondi dialects - Taraka Rama - Çağrı Çöltekin - Pavel Sofroniev + TarakaRama + ÇağrıÇöltekin + PavelSofroniev 26–35 W17-1203 10.18653/v1/W17-1203 @@ -1670,8 +1670,8 @@ Investigating Diatopic Variation in a Historical Corpus - Stefanie Dipper - Sandra Waldenberger + StefanieDipper + SandraWaldenberger 36–45 W17-1204 10.18653/v1/W17-1204 @@ -1679,7 +1679,7 @@ Author Profiling at <fixed-case>PAN</fixed-case>: from Age and Gender Identification to Language Variety Identification (invited talk) - Paolo Rosso + PaoloRosso 46 W17-1205 10.18653/v1/W17-1205 @@ -1687,7 +1687,7 @@ The similarity and Mutual Intelligibility between <fixed-case>A</fixed-case>mharic and <fixed-case>T</fixed-case>igrigna Varieties - Tekabe Legesse Feleke + Tekabe LegesseFeleke 47–54 W17-1206 10.18653/v1/W17-1206 @@ -1695,7 +1695,7 @@ Why <fixed-case>C</fixed-case>atalan-<fixed-case>S</fixed-case>panish Neural Machine Translation? Analysis, comparison and combination with standard Rule and Phrase-based technologies - Marta R. Costa-jussà + Marta R.Costa-jussà 55–62 W17-1207 10.18653/v1/W17-1207 @@ -1703,7 +1703,7 @@ <fixed-case>K</fixed-case>urdish Interdialect Machine Translation - Hossein Hassani + HosseinHassani 63–72 W17-1208 10.18653/v1/W17-1208 @@ -1711,8 +1711,8 @@ Twitter Language Identification Of Similar Languages And Dialects Without Ground Truth - Jennifer Williams - Charlie Dagli + JenniferWilliams + CharlieDagli 73–83 W17-1209 10.18653/v1/W17-1209 @@ -1720,8 +1720,8 @@ Multi-source morphosyntactic tagging for spoken <fixed-case>R</fixed-case>usyn - Yves Scherrer - Achim Rabus + YvesScherrer + AchimRabus 84–92 W17-1210 10.18653/v1/W17-1210 @@ -1729,9 +1729,9 @@ Identifying dialects with textual and acoustic cues - Abualsoud Hanani - Aziz Qaroush - Stephen Taylor + AbualsoudHanani + AzizQaroush + StephenTaylor 93–101 W17-1211 10.18653/v1/W17-1211 @@ -1739,9 +1739,9 @@ Evaluating <fixed-case>H</fixed-case>e<fixed-case>LI</fixed-case> with Non-Linear Mappings - Tommi Jauhiainen - Krister Lindén - Heidi Jauhiainen + TommiJauhiainen + KristerLindén + HeidiJauhiainen 102–108 W17-1212 10.18653/v1/W17-1212 @@ -1749,9 +1749,9 @@ A Perplexity-Based Method for Similar Languages Discrimination - Pablo Gamallo - Jose Ramom Pichel - Iñaki Alegria + PabloGamallo + Jose RamomPichel + IñakiAlegria 109–114 W17-1213 10.18653/v1/W17-1213 @@ -1759,7 +1759,7 @@ Improving the Character Ngram Model for the <fixed-case>DSL</fixed-case> Task with <fixed-case>BM</fixed-case>25 Weighting and Less Frequently Used Feature Sets - Yves Bestgen + YvesBestgen 115–123 W17-1214 10.18653/v1/W17-1214 @@ -1767,8 +1767,8 @@ Discriminating between Similar Languages with Word-level Convolutional Neural Networks - Marcelo Criscuolo - Sandra Maria Aluísio + MarceloCriscuolo + Sandra MariaAluísio 124–130 W17-1215 10.18653/v1/W17-1215 @@ -1776,7 +1776,7 @@ Cross-lingual dependency parsing for closely related languages - <fixed-case>H</fixed-case>elsinki’s submission to <fixed-case>V</fixed-case>ar<fixed-case>D</fixed-case>ial 2017 - Jörg Tiedemann + JörgTiedemann 131–136 W17-1216 10.18653/v1/W17-1216 @@ -1784,11 +1784,11 @@ Discriminating between Similar Languages Using a Combination of Typed and Untyped Character N-grams and Words - Helena Gomez - Ilia Markov - Jorge Baptista - Grigori Sidorov - David Pinto + HelenaGomez + IliaMarkov + JorgeBaptista + GrigoriSidorov + DavidPinto 137–145 W17-1217 10.18653/v1/W17-1217 @@ -1796,8 +1796,8 @@ Tübingen system in <fixed-case>V</fixed-case>ar<fixed-case>D</fixed-case>ial 2017 shared task: experiments with language identification and cross-lingual parsing - Çağrı Çöltekin - Taraka Rama + ÇağrıÇöltekin + TarakaRama 146–155 W17-1218 10.18653/v1/W17-1218 @@ -1805,9 +1805,9 @@ When Sparse Traditional Models Outperform Dense Neural Networks: the Curious Case of Discriminating between Similar Languages - Maria Medvedeva - Martin Kroon - Barbara Plank + MariaMedvedeva + MartinKroon + BarbaraPlank 156–163 W17-1219 10.18653/v1/W17-1219 @@ -1815,8 +1815,8 @@ <fixed-case>G</fixed-case>erman Dialect Identification in Interview Transcriptions - Shervin Malmasi - Marcos Zampieri + ShervinMalmasi + MarcosZampieri 164–169 W17-1220 10.18653/v1/W17-1220 @@ -1824,8 +1824,8 @@ <fixed-case>CLUZH</fixed-case> at <fixed-case>V</fixed-case>ar<fixed-case>D</fixed-case>ial <fixed-case>GDI</fixed-case> 2017: Testing a Variety of Machine Learning Tools for the Classification of Swiss <fixed-case>G</fixed-case>erman Dialects - Simon Clematide - Peter Makarov + SimonClematide + PeterMakarov 170–177 W17-1221 10.18653/v1/W17-1221 @@ -1833,8 +1833,8 @@ <fixed-case>A</fixed-case>rabic Dialect Identification Using i<fixed-case>V</fixed-case>ectors and <fixed-case>ASR</fixed-case> Transcripts - Shervin Malmasi - Marcos Zampieri + ShervinMalmasi + MarcosZampieri 178–183 W17-1222 10.18653/v1/W17-1222 @@ -1842,7 +1842,7 @@ Discriminating between Similar Languages using Weighted Subword Features - Adrien Barbaresi + AdrienBarbaresi 184–189 W17-1223 10.18653/v1/W17-1223 @@ -1850,8 +1850,8 @@ Exploring Lexical and Syntactic Features for Language Variety Identification - Chris van der Lee - Antal van den Bosch + Chrisvan der Lee + Antalvan den Bosch 190–199 W17-1224 10.18653/v1/W17-1224 @@ -1859,8 +1859,8 @@ Learning to Identify <fixed-case>A</fixed-case>rabic and <fixed-case>G</fixed-case>erman Dialects using Multiple Kernels - Radu Tudor Ionescu - Andrei Butnaru + Radu TudorIonescu + AndreiButnaru 200–209 W17-1225 10.18653/v1/W17-1225 @@ -1868,10 +1868,10 @@ <fixed-case>S</fixed-case>lavic Forest, <fixed-case>N</fixed-case>orwegian Wood - Rudolf Rosa - Daniel Zeman - David Mareček - Zdeněk Žabokrtský + RudolfRosa + DanielZeman + DavidMareček + ZdeněkŽabokrtský 210–219 W17-1226 10.18653/v1/W17-1226 @@ -1901,8 +1901,8 @@ Identification of Languages in <fixed-case>A</fixed-case>lgerian <fixed-case>A</fixed-case>rabic Multilingual Documents - Wafia Adouane - Simon Dobnik + WafiaAdouane + SimonDobnik 1–8 W17-1301 10.18653/v1/W17-1301 @@ -1910,9 +1910,9 @@ <fixed-case>A</fixed-case>rabic Diacritization: Stats, Rules, and Hacks - Kareem Darwish - Hamdy Mubarak - Ahmed Abdelali + KareemDarwish + HamdyMubarak + AhmedAbdelali 9–17 W17-1302 10.18653/v1/W17-1302 @@ -1920,8 +1920,8 @@ Semantic Similarity of <fixed-case>A</fixed-case>rabic Sentences with Word Embeddings - El Moatez Billah Nagoudi - Didier Schwab + El Moatez BillahNagoudi + DidierSchwab 18–24 W17-1303 10.18653/v1/W17-1303 @@ -1929,8 +1929,8 @@ Morphological Analysis for the <fixed-case>M</fixed-case>altese Language: The challenges of a hybrid system - Claudia Borg - Albert Gatt + ClaudiaBorg + AlbertGatt 25–34 W17-1304 10.18653/v1/W17-1304 @@ -1938,9 +1938,9 @@ A Morphological Analyzer for Gulf <fixed-case>A</fixed-case>rabic Verbs - Salam Khalifa - Sara Hassan - Nizar Habash + SalamKhalifa + SaraHassan + NizarHabash 35–45 W17-1305 10.18653/v1/W17-1305 @@ -1948,13 +1948,13 @@ A Neural Architecture for Dialectal <fixed-case>A</fixed-case>rabic Segmentation - Younes Samih - Mohammed Attia - Mohamed Eldesouki - Ahmed Abdelali - Hamdy Mubarak - Laura Kallmeyer - Kareem Darwish + YounesSamih + MohammedAttia + MohamedEldesouki + AhmedAbdelali + HamdyMubarak + LauraKallmeyer + KareemDarwish 46–54 W17-1306 10.18653/v1/W17-1306 @@ -1962,10 +1962,10 @@ Sentiment Analysis of <fixed-case>T</fixed-case>unisian Dialects: Linguistic Ressources and Experiments - Salima Medhaffar - Fethi Bougares - Yannick Estève - Lamia Hadrich-Belguith + SalimaMedhaffar + FethiBougares + YannickEstève + LamiaHadrich-Belguith 55–61 W17-1307 10.18653/v1/W17-1307 @@ -1973,12 +1973,12 @@ <fixed-case>CAT</fixed-case>: Credibility Analysis of <fixed-case>A</fixed-case>rabic Content on Twitter - Rim El Ballouli - Wassim El-Hajj - Ahmad Ghandour - Shady Elbassuoni - Hazem Hajj - Khaled Shaban + RimEl Ballouli + WassimEl-Hajj + AhmadGhandour + ShadyElbassuoni + HazemHajj + KhaledShaban 62–71 W17-1308 10.18653/v1/W17-1308 @@ -1986,8 +1986,8 @@ A New Error Annotation for Dyslexic texts in <fixed-case>A</fixed-case>rabic - Maha Alamri - William J Teahan + MahaAlamri + William JTeahan 72–78 W17-1309 10.18653/v1/W17-1309 @@ -1995,12 +1995,12 @@ An Unsupervised Speaker Clustering Technique based on <fixed-case>SOM</fixed-case> and <fixed-case>I</fixed-case>-vectors for Speech Recognition Systems - Hany Ahmed - Mohamed Elaraby - Abdullah M. Mousa - Mostafa Elhosiny - Sherif Abdou - Mohsen Rashwan + HanyAhmed + MohamedElaraby + AbdullahM. Mousa + MostafaElhosiny + SherifAbdou + MohsenRashwan 79–83 W17-1310 10.18653/v1/W17-1310 @@ -2008,8 +2008,8 @@ <fixed-case>SHAKKIL</fixed-case>: An Automatic Diacritization System for Modern Standard <fixed-case>A</fixed-case>rabic Texts - Amany Fashwan - Sameh Alansary + AmanyFashwan + SamehAlansary 84–93 W17-1311 10.18653/v1/W17-1311 @@ -2017,9 +2017,9 @@ <fixed-case>A</fixed-case>rabic Tweets Treebanking and Parsing: A Bootstrapping Approach - Fahad Albogamy - Allan Ramsay - Hanady Ahmed + FahadAlbogamy + AllanRamsay + HanadyAhmed 94–99 W17-1312 10.18653/v1/W17-1312 @@ -2027,10 +2027,10 @@ Identifying Effective Translations for Cross-lingual <fixed-case>A</fixed-case>rabic-to-<fixed-case>E</fixed-case>nglish User-generated Speech Search - Ahmad Khwileh - Haithem Afli - Gareth Jones - Andy Way + AhmadKhwileh + HaithemAfli + GarethJones + AndyWay 100–109 W17-1313 10.18653/v1/W17-1313 @@ -2038,15 +2038,15 @@ A Characterization Study of <fixed-case>A</fixed-case>rabic Twitter Data with a Benchmarking for State-of-the-Art Opinion Mining Models - Ramy Baly - Gilbert Badaro - Georges El-Khoury - Rawan Moukalled - Rita Aoun - Hazem Hajj - Wassim El-Hajj - Nizar Habash - Khaled Shaban + RamyBaly + GilbertBadaro + GeorgesEl-Khoury + RawanMoukalled + RitaAoun + HazemHajj + WassimEl-Hajj + NizarHabash + KhaledShaban 110–118 W17-1314 10.18653/v1/W17-1314 @@ -2054,9 +2054,9 @@ Robust Dictionary Lookup in Multiple Noisy Orthographies - Lingliang Zhang - Nizar Habash - Godfried Toussaint + LingliangZhang + NizarHabash + GodfriedToussaint 119–129 W17-1315 10.18653/v1/W17-1315 @@ -2064,10 +2064,10 @@ <fixed-case>A</fixed-case>rabic <fixed-case>POS</fixed-case> Tagging: Don’t Abandon Feature Engineering Just Yet - Kareem Darwish - Hamdy Mubarak - Ahmed Abdelali - Mohamed Eldesouki + KareemDarwish + HamdyMubarak + AhmedAbdelali + MohamedEldesouki 130–137 W17-1316 10.18653/v1/W17-1316 @@ -2075,10 +2075,10 @@ Toward a Web-based Speech Corpus for <fixed-case>A</fixed-case>lgerian Dialectal <fixed-case>A</fixed-case>rabic Varieties - Soumia Bougrine - Aicha Chorana - Abdallah Lakhdari - Hadda Cherroun + SoumiaBougrine + AichaChorana + AbdallahLakhdari + HaddaCherroun 138–146 W17-1317 10.18653/v1/W17-1317 @@ -2086,7 +2086,7 @@ Not All Segments are Created Equal: Syntactically Motivated Sentiment Analysis in Lexical Space - Muhammad Abdul-Mageed + MuhammadAbdul-Mageed 147–156 W17-1318 10.18653/v1/W17-1318 @@ -2094,12 +2094,12 @@ An enhanced automatic speech recognition system for <fixed-case>A</fixed-case>rabic - Mohamed Amine Menacer - Odile Mella - Dominique Fohr - Denis Jouvet - David Langlois - Kamel Smaili + Mohamed AmineMenacer + OdileMella + DominiqueFohr + DenisJouvet + DavidLanglois + KamelSmaili 157–165 W17-1319 10.18653/v1/W17-1319 @@ -2107,9 +2107,9 @@ Universal Dependencies for <fixed-case>A</fixed-case>rabic - Dima Taji - Nizar Habash - Daniel Zeman + DimaTaji + NizarHabash + DanielZeman 166–176 W17-1320 10.18653/v1/W17-1320 @@ -2117,9 +2117,9 @@ A Layered Language Model based Hybrid Approach to Automatic Full Diacritization of <fixed-case>A</fixed-case>rabic - Mohamed Al-Badrashiny - Abdelati Hawwari - Mona Diab + MohamedAl-Badrashiny + AbdelatiHawwari + MonaDiab 177–184 W17-1321 10.18653/v1/W17-1321 @@ -2127,8 +2127,8 @@ <fixed-case>A</fixed-case>rabic Textual Entailment with Word Embeddings - Nada Almarwani - Mona Diab + NadaAlmarwani + MonaDiab 185–190 W17-1322 10.18653/v1/W17-1322 @@ -2156,7 +2156,7 @@ Toward Pan-<fixed-case>S</fixed-case>lavic <fixed-case>NLP</fixed-case>: Some Experiments with Language Adaptation - Serge Sharoff + SergeSharoff 1–2 W17-1401 10.18653/v1/W17-1401 @@ -2164,9 +2164,9 @@ Clustering of <fixed-case>R</fixed-case>ussian Adjective-Noun Constructions using Word Embeddings - Andrey Kutuzov - Elizaveta Kuzmenko - Lidia Pivovarova + AndreyKutuzov + ElizavetaKuzmenko + LidiaPivovarova 3–13 W17-1402 10.18653/v1/W17-1402 @@ -2174,8 +2174,8 @@ A Preliminary Study of <fixed-case>C</fixed-case>roatian Lexical Substitution - Domagoj Alagić - Jan Šnajder + DomagojAlagić + JanŠnajder 14–19 W17-1403 10.18653/v1/W17-1403 @@ -2183,8 +2183,8 @@ Projecting Multiword Expression Resources on a Polish Treebank - Agata Savary - Jakub Waszczuk + AgataSavary + JakubWaszczuk 20–26 W17-1404 10.18653/v1/W17-1404 @@ -2192,8 +2192,8 @@ Lexicon Induction for Spoken <fixed-case>R</fixed-case>usyn – Challenges and Results - Achim Rabus - Yves Scherrer + AchimRabus + YvesScherrer 27–32 W17-1405 10.18653/v1/W17-1405 @@ -2201,9 +2201,9 @@ The Universal Dependencies Treebank for <fixed-case>S</fixed-case>lovenian - Kaja Dobrovoljc - Tomaž Erjavec - Simon Krek + KajaDobrovoljc + TomažErjavec + SimonKrek 33–38 W17-1406 10.18653/v1/W17-1406 @@ -2211,10 +2211,10 @@ Universal Dependencies for <fixed-case>S</fixed-case>erbian in Comparison with <fixed-case>C</fixed-case>roatian and Other <fixed-case>S</fixed-case>lavic Languages - Tanja Samardžić - Mirjana Starović - Željko Agić - Nikola Ljubešić + TanjaSamardžić + MirjanaStarović + ŽeljkoAgić + NikolaLjubešić 39–44 W17-1407 10.18653/v1/W17-1407 @@ -2222,7 +2222,7 @@ Spelling Correction for Morphologically Rich Language: a Case Study of <fixed-case>R</fixed-case>ussian - Alexey Sorokin + AlexeySorokin 45–53 W17-1408 10.18653/v1/W17-1408 @@ -2230,10 +2230,10 @@ Debunking Sentiment Lexicons: A Case of Domain-Specific Sentiment Classification for <fixed-case>C</fixed-case>roatian - Paula Gombar - Zoran Medić - Domagoj Alagić - Jan Šnajder + PaulaGombar + ZoranMedić + DomagojAlagić + JanŠnajder 54–59 W17-1409 10.18653/v1/W17-1409 @@ -2241,9 +2241,9 @@ Adapting a State-of-the-Art Tagger for South <fixed-case>S</fixed-case>lavic Languages to Non-Standard Text - Nikola Ljubešić - Tomaž Erjavec - Darja Fišer + NikolaLjubešić + TomažErjavec + DarjaFišer 60–68 W17-1410 10.18653/v1/W17-1410 @@ -2251,8 +2251,8 @@ Comparison of Short-Text Sentiment Analysis Methods for <fixed-case>C</fixed-case>roatian - Leon Rotim - Jan Šnajder + LeonRotim + JanŠnajder 69–75 W17-1411 10.18653/v1/W17-1411 @@ -2260,11 +2260,11 @@ The First Cross-Lingual Challenge on Recognition, Normalization, and Matching of Named Entities in <fixed-case>S</fixed-case>lavic Languages - Jakub Piskorski - Lidia Pivovarova - Jan Šnajder - Josef Steinberger - Roman Yangarber + JakubPiskorski + LidiaPivovarova + JanŠnajder + JosefSteinberger + RomanYangarber 76–85 W17-1412 10.18653/v1/W17-1412 @@ -2272,9 +2272,9 @@ <fixed-case>L</fixed-case>iner2 — a Generic Framework for Named Entity Recognition - Michał Marcińczuk - Jan Kocoń - Marcin Oleksy + MichałMarcińczuk + JanKocoń + MarcinOleksy 86–91 W17-1413 10.18653/v1/W17-1413 @@ -2282,9 +2282,9 @@ Language-Independent Named Entity Analysis Using Parallel Projection and Rule-Based Disambiguation - James Mayfield - Paul McNamee - Cash Costello + JamesMayfield + PaulMcNamee + CashCostello 92–96 W17-1414 10.18653/v1/W17-1414 @@ -2292,7 +2292,7 @@ Comparison of String Similarity Measures for Obscenity Filtering - Ekaterina Chernyak + EkaterinaChernyak 97–101 W17-1415 10.18653/v1/W17-1415 @@ -2300,8 +2300,8 @@ Stylometric Analysis of Parliamentary Speeches: Gender Dimension - Justina Mandravickaitė - Tomas Krilavičius + JustinaMandravickaitė + TomasKrilavičius 102–107 W17-1416 10.18653/v1/W17-1416 @@ -2309,10 +2309,10 @@ Towards Never Ending Language Learning for Morphologically Rich Languages - Kseniya Buraya - Lidia Pivovarova - Sergey Budkov - Andrey Filchenkov + KseniyaBuraya + LidiaPivovarova + SergeyBudkov + AndreyFilchenkov 108–118 W17-1417 10.18653/v1/W17-1417 @@ -2320,9 +2320,9 @@ Gender Profiling for <fixed-case>S</fixed-case>lovene Twitter communication: the Influence of Gender Marking, Content and Style - Ben Verhoeven - Iza Škrjanec - Senja Pollak + BenVerhoeven + IzaŠkrjanec + SenjaPollak 119–125 W17-1418 10.18653/v1/W17-1418 @@ -2346,8 +2346,8 @@ Use Generalized Representations, But Do Not Forget Surface Features - Nafise Sadat Moosavi - Michael Strube + Nafise SadatMoosavi + MichaelStrube 1–7 W17-1501 10.18653/v1/W17-1501 @@ -2355,10 +2355,10 @@ Enriching Basque Coreference Resolution System using Semantic Knowledge sources - Ander Soraluze - Olatz Arregi - Xabier Arregi - Arantza Díaz de Ilarraza + AnderSoraluze + OlatzArregi + XabierArregi + ArantzaDíaz de Ilarraza 8–16 W17-1502 10.18653/v1/W17-1502 @@ -2366,8 +2366,8 @@ Improving Polish Mention Detection with Valency Dictionary - Maciej Ogrodniczuk - Bartłomiej Nitoń + MaciejOgrodniczuk + BartłomiejNitoń 17–23 W17-1503 10.18653/v1/W17-1503 @@ -2375,8 +2375,8 @@ A <fixed-case>G</fixed-case>oogle-Proof Collection of <fixed-case>F</fixed-case>rench <fixed-case>W</fixed-case>inograd Schemas - Pascal Amsili - Olga Seminck + PascalAmsili + OlgaSeminck 24–29 W17-1504 10.18653/v1/W17-1504 @@ -2384,8 +2384,8 @@ Using Coreference Links to Improve <fixed-case>S</fixed-case>panish-to-<fixed-case>E</fixed-case>nglish Machine Translation - Lesly Miculicich Werlen - Andrei Popescu-Belis + LeslyMiculicich Werlen + AndreiPopescu-Belis 30–40 W17-1505 10.18653/v1/W17-1505 @@ -2393,8 +2393,8 @@ Multi-source annotation projection of coreference chains: assessing strategies and testing opportunities - Yulia Grishina - Manfred Stede + YuliaGrishina + ManfredStede 41–50 W17-1506 10.18653/v1/W17-1506 @@ -2402,7 +2402,7 @@ <fixed-case>CORBON</fixed-case> 2017 Shared Task: Projection-Based Coreference Resolution - Yulia Grishina + YuliaGrishina 51–55 W17-1507 10.18653/v1/W17-1507 @@ -2410,9 +2410,9 @@ Projection-based Coreference Resolution Using Deep Syntax - Michal Novák - Anna Nedoluzhko - Zdeněk Žabokrtský + MichalNovák + AnnaNedoluzhko + ZdeněkŽabokrtský 56–64 W17-1508 10.18653/v1/W17-1508 @@ -2440,7 +2440,7 @@ Gender as a Variable in Natural-Language Processing: Ethical Considerations - Brian Larson + BrianLarson 1–11 W17-1601 10.18653/v1/W17-1601 @@ -2448,8 +2448,8 @@ These are not the Stereotypes You are Looking For: Bias and Fairness in Authorial Gender Attribution - Corina Koolen - Andreas van Cranenburgh + CorinaKoolen + Andreasvan Cranenburgh 12–22 W17-1602 10.18653/v1/W17-1602 @@ -2457,7 +2457,7 @@ A Quantitative Study of Data in the <fixed-case>NLP</fixed-case> community - Margot Mieskes + MargotMieskes 23–29 W17-1603 10.18653/v1/W17-1603 @@ -2465,8 +2465,8 @@ Ethical by Design: Ethics Best Practices for Natural Language Processing - Jochen L. Leidner - Vassilis Plachouras + Jochen L.Leidner + VassilisPlachouras 30–40 W17-1604 10.18653/v1/W17-1604 @@ -2474,11 +2474,11 @@ Building Better Open-Source Tools to Support Fairness in Automated Scoring - Nitin Madnani - Anastassia Loukina - Alina von Davier - Jill Burstein - Aoife Cahill + NitinMadnani + AnastassiaLoukina + Alinavon Davier + JillBurstein + AoifeCahill 41–52 W17-1605 10.18653/v1/W17-1605 @@ -2486,7 +2486,7 @@ Gender and Dialect Bias in <fixed-case>Y</fixed-case>ou<fixed-case>T</fixed-case>ube’s Automatic Captions - Rachael Tatman + RachaelTatman 53–59 W17-1606 10.18653/v1/W17-1606 @@ -2494,9 +2494,9 @@ Integrating the Management of Personal Data Protection and Open Science with Research Ethics - Dave Lewis - Joss Moorkens - Kaniz Fatema + DaveLewis + JossMoorkens + KanizFatema 60–65 W17-1607 10.18653/v1/W17-1607 @@ -2504,12 +2504,12 @@ Ethical Considerations in <fixed-case>NLP</fixed-case> Shared Tasks - Carla Parra Escartín - Wessel Reijers - Teresa Lynn - Joss Moorkens - Andy Way - Chao-Hong Liu + CarlaParra Escartín + WesselReijers + TeresaLynn + JossMoorkens + AndyWay + Chao-HongLiu 66–73 W17-1608 10.18653/v1/W17-1608 @@ -2517,9 +2517,9 @@ Social Bias in Elicited Natural Language Inferences - Rachel Rudinger - Chandler May - Benjamin Van Durme + RachelRudinger + ChandlerMay + BenjaminVan Durme 74–79 W17-1609 10.18653/v1/W17-1609 @@ -2527,9 +2527,9 @@ A Short Review of Ethical Challenges in Clinical Natural Language Processing - Simon Šuster - Stéphan Tulkens - Walter Daelemans + SimonŠuster + StéphanTulkens + WalterDaelemans 80–87 W17-1610 10.18653/v1/W17-1610 @@ -2537,7 +2537,7 @@ Goal-Oriented Design for Ethical Machine Learning and <fixed-case>NLP</fixed-case> - Tyler Schnoebelen + TylerSchnoebelen 88–93 W17-1611 10.18653/v1/W17-1611 @@ -2545,9 +2545,9 @@ Ethical Research Protocols for Social Media Health Research - Adrian Benton - Glen Coppersmith - Mark Dredze + AdrianBenton + GlenCoppersmith + MarkDredze 94–102 W17-1612 10.18653/v1/W17-1612 @@ -2555,10 +2555,10 @@ Say the Right Thing Right: Ethics Issues in Natural Language Generation Systems - Charese Smiley - Frank Schilder - Vassilis Plachouras - Jochen L. Leidner + ChareseSmiley + FrankSchilder + VassilisPlachouras + Jochen L.Leidner 103–108 W17-1613 10.18653/v1/W17-1613 @@ -2584,8 +2584,8 @@ <fixed-case>P</fixed-case>ara<fixed-case>D</fixed-case>i: Dictionary of Paraphrases of <fixed-case>C</fixed-case>zech Complex Predicates with Light Verbs - Petra Barančíková - Václava Kettnerová + PetraBarančíková + VáclavaKettnerová 1–10 W17-1701 10.18653/v1/W17-1701 @@ -2593,10 +2593,10 @@ Multi-word Entity Classification in a Highly Multilingual Environment - Sophie Chesney - Guillaume Jacquet - Ralf Steinberger - Jakub Piskorski + SophieChesney + GuillaumeJacquet + RalfSteinberger + JakubPiskorski 11–20 W17-1702 10.18653/v1/W17-1702 @@ -2604,9 +2604,9 @@ Using bilingual word-embeddings for multilingual collocation extraction - Marcos Garcia - Marcos García-Salido - Margarita Alonso-Ramos + MarcosGarcia + MarcosGarcía-Salido + MargaritaAlonso-Ramos 21–30 W17-1703 10.18653/v1/W17-1703 @@ -2614,17 +2614,17 @@ The <fixed-case>PARSEME</fixed-case> Shared Task on Automatic Identification of Verbal Multiword Expressions - Agata Savary - Carlos Ramisch - Silvio Cordeiro - Federico Sangati - Veronika Vincze - Behrang QasemiZadeh - Marie Candito - Fabienne Cap - Voula Giouli - Ivelina Stoyanova - Antoine Doucet + AgataSavary + CarlosRamisch + SilvioCordeiro + FedericoSangati + VeronikaVincze + BehrangQasemiZadeh + MarieCandito + FabienneCap + VoulaGiouli + IvelinaStoyanova + AntoineDoucet 31–47 W17-1704 10.18653/v1/W17-1704 @@ -2632,9 +2632,9 @@ <fixed-case>US</fixed-case>zeged: Identifying Verbal Multiword Expressions with <fixed-case>POS</fixed-case> Tagging and Parsing Techniques - Katalin Ilona Simkó - Viktória Kovács - Veronika Vincze + Katalin IlonaSimkó + ViktóriaKovács + VeronikaVincze 48–53 W17-1705 10.18653/v1/W17-1705 @@ -2642,9 +2642,9 @@ Parsing and <fixed-case>MWE</fixed-case> Detection: Fips at the <fixed-case>PARSEME</fixed-case> Shared Task - Luka Nerima - Vasiliki Foufi - Éric Wehrli + LukaNerima + VasilikiFoufi + ÉricWehrli 54–59 W17-1706 10.18653/v1/W17-1706 @@ -2652,9 +2652,9 @@ Neural Networks for Multi-Word Expression Detection - Natalia Klyueva - Antoine Doucet - Milan Straka + NataliaKlyueva + AntoineDoucet + MilanStraka 60–65 W17-1707 10.18653/v1/W17-1707 @@ -2662,8 +2662,8 @@ Factoring Ambiguity out of the Prediction of Compositionality for <fixed-case>G</fixed-case>erman Multi-Word Expressions - Stefan Bott - Sabine Schulte im Walde + StefanBott + SabineSchulte im Walde 66–72 W17-1708 10.18653/v1/W17-1708 @@ -2671,7 +2671,7 @@ Multiword expressions and lexicalism: the view from <fixed-case>LFG</fixed-case> - Jamie Y. Findlay + Jamie Y.Findlay 73–79 W17-1709 10.18653/v1/W17-1709 @@ -2679,9 +2679,9 @@ Understanding Idiomatic Variation - Kristina Geeraert - R. Harald Baayen - John Newman + KristinaGeeraert + R. HaraldBaayen + JohnNewman 80–90 W17-1710 10.18653/v1/W17-1710 @@ -2689,9 +2689,9 @@ Discovering Light Verb Constructions and their Translations from Parallel Corpora without Word Alignment - Natalie Vargas - Carlos Ramisch - Helena Caseli + NatalieVargas + CarlosRamisch + HelenaCaseli 91–96 W17-1711 10.18653/v1/W17-1711 @@ -2699,8 +2699,8 @@ Identification of Multiword Expressions for <fixed-case>L</fixed-case>atvian and <fixed-case>L</fixed-case>ithuanian: Hybrid Approach - Justina Mandravickaitė - Tomas Krilavičius + JustinaMandravickaitė + TomasKrilavičius 97–101 W17-1712 10.18653/v1/W17-1712 @@ -2708,7 +2708,7 @@ Show Me Your Variance and <fixed-case>I</fixed-case> Tell You Who You Are - Deriving Compound Compositionality from Word Alignments - Fabienne Cap + FabienneCap 102–107 W17-1713 10.18653/v1/W17-1713 @@ -2716,8 +2716,8 @@ Semantic annotation to characterize contextual variation in terminological noun compounds: a pilot study - Melania Cabezas-García - Antonio San Martín + MelaniaCabezas-García + AntonioSan Martín 108–113 W17-1714 10.18653/v1/W17-1714 @@ -2725,13 +2725,13 @@ Detection of Verbal Multi-Word Expressions via Conditional Random Fields with Syntactic Dependency Features and Semantic Re-Ranking - Alfredo Maldonado - Lifeng Han - Erwan Moreau - Ashjan Alsulaimani - Koel Dutta Chowdhury - Carl Vogel - Qun Liu + AlfredoMaldonado + LifengHan + ErwanMoreau + AshjanAlsulaimani + Koel DuttaChowdhury + CarlVogel + QunLiu 114–120 W17-1715 10.18653/v1/W17-1715 @@ -2739,10 +2739,10 @@ A data-driven approach to verbal multiword expression detection. <fixed-case>PARSEME</fixed-case> Shared Task system description paper - Tiberiu Boros - Sonia Pipa - Verginica Barbu Mititelu - Dan Tufis + TiberiuBoros + SoniaPipa + VerginicaBarbu Mititelu + DanTufis 121–126 W17-1716 10.18653/v1/W17-1716 @@ -2750,9 +2750,9 @@ The <fixed-case>ATILF</fixed-case>-<fixed-case>LLF</fixed-case> System for Parseme Shared Task: a Transition-based Verbal Multiword Expression Tagger - Hazem Al Saied - Matthieu Constant - Marie Candito + HazemAl Saied + MatthieuConstant + MarieCandito 127–132 W17-1717 10.18653/v1/W17-1717 @@ -2760,10 +2760,10 @@ Investigating the Opacity of Verb-Noun Multiword Expression Usages in Context - Shiva Taslimipoor - Omid Rohanian - Ruslan Mitkov - Afsaneh Fazly + ShivaTaslimipoor + OmidRohanian + RuslanMitkov + AfsanehFazly 133–138 W17-1718 10.18653/v1/W17-1718 @@ -2771,9 +2771,9 @@ Compositionality in Verb-Particle Constructions - Archna Bhatia - Choh Man Teng - James Allen + ArchnaBhatia + Choh ManTeng + JamesAllen 139–148 W17-1719 10.18653/v1/W17-1719 @@ -2781,11 +2781,11 @@ Rule-Based Translation of <fixed-case>S</fixed-case>panish Verb-Noun Combinations into Basque - Uxoa Iñurrieta - Itziar Aduriz - Arantza Díaz de Ilarraza - Gorka Labaka - Kepa Sarasola + UxoaIñurrieta + ItziarAduriz + ArantzaDíaz de Ilarraza + GorkaLabaka + KepaSarasola 149–154 W17-1720 10.18653/v1/W17-1720 @@ -2793,7 +2793,7 @@ Verb-Particle Constructions in Questions - Veronika Vincze + VeronikaVincze 155–160 W17-1721 10.18653/v1/W17-1721 @@ -2801,7 +2801,7 @@ Simple Compound Splitting for <fixed-case>G</fixed-case>erman - Marion Weller-Di Marco + MarionWeller-Di Marco 161–166 W17-1722 10.18653/v1/W17-1722 @@ -2809,8 +2809,8 @@ Identification of Ambiguous Multiword Expressions Using Sequence Models and Lexical Resources - Manon Scholivet - Carlos Ramisch + ManonScholivet + CarlosRamisch 167–175 W17-1723 10.18653/v1/W17-1723 @@ -2818,8 +2818,8 @@ Comparing Recurring Lexico-Syntactic Trees (<fixed-case>RLT</fixed-case>s) and Ngram Techniques for Extended Phraseology Extraction - Agnès Tutin - Olivier Kraif + AgnèsTutin + OlivierKraif 176–180 W17-1724 10.18653/v1/W17-1724 @@ -2827,8 +2827,8 @@ Benchmarking Joint Lexical and Syntactic Analysis on Multiword-Rich Data - Matthieu Constant - Héctor Martinez Alonso + MatthieuConstant + HéctorMartinez Alonso 181–186 W17-1725 10.18653/v1/W17-1725 @@ -2836,9 +2836,9 @@ Semi-Automated Resolution of Inconsistency for a Harmonized Multiword Expression and Dependency Parse Annotation - King Chan - Julian Brooke - Timothy Baldwin + KingChan + JulianBrooke + TimothyBaldwin 187–193 W17-1726 10.18653/v1/W17-1726 @@ -2846,8 +2846,8 @@ Combining Linguistic Features for the Detection of <fixed-case>C</fixed-case>roatian Multiword Expressions - Maja Buljan - Jan Šnajder + MajaBuljan + JanŠnajder 194–199 W17-1727 10.18653/v1/W17-1727 @@ -2855,8 +2855,8 @@ Complex Verbs are Different: Exploring the Visual Modality in Multi-Modal Models to Predict Compositionality - Maximilian Köper - Sabine Schulte im Walde + MaximilianKöper + SabineSchulte im Walde 200–206 W17-1728 10.18653/v1/W17-1728 @@ -2881,11 +2881,11 @@ Understanding the Semantics of Narratives of Interpersonal Violence through Reader Annotations and Physiological Reactions - Alexander Calderwood - Elizabeth A. Pruett - Raymond Ptucha - Christopher Homan - Cecilia Ovesdotter Alm + AlexanderCalderwood + Elizabeth A.Pruett + RaymondPtucha + ChristopherHoman + CeciliaOvesdotter Alm 1–9 W17-1801 10.18653/v1/W17-1801 @@ -2893,8 +2893,8 @@ Intension, Attitude, and Tense Annotation in a High-Fidelity Semantic Representation - Gene Kim - Lenhart Schubert + GeneKim + LenhartSchubert 10–15 W17-1802 10.18653/v1/W17-1802 @@ -2902,8 +2902,8 @@ Towards a lexicon of event-selecting predicates for a <fixed-case>F</fixed-case>rench <fixed-case>F</fixed-case>act<fixed-case>B</fixed-case>ank - Ingrid Falk - Fabienne Martin + IngridFalk + FabienneMartin 16–21 W17-1803 10.18653/v1/W17-1803 @@ -2911,10 +2911,10 @@ Universal Dependencies to Logical Form with Negation Scope - Federico Fancellu - Siva Reddy - Adam Lopez - Bonnie Webber + FedericoFancellu + SivaReddy + AdamLopez + BonnieWebber 22–32 W17-1804 10.18653/v1/W17-1804 @@ -2925,7 +2925,7 @@ is able to handle phenomena related to scope by means of an higher-order type th Meaning Banking beyond Events and Roles - Johan Bos + JohanBos 33 W17-1805 10.18653/v1/W17-1805 @@ -2933,9 +2933,9 @@ is able to handle phenomena related to scope by means of an higher-order type th The Scope and Focus of Negation: A Complete Annotation Framework for <fixed-case>I</fixed-case>talian - Begoña Altuna - Anne-Lyse Minard - Manuela Speranza + BegoñaAltuna + Anne-LyseMinard + ManuelaSperanza 34–42 W17-1806 10.18653/v1/W17-1806 @@ -2943,9 +2943,9 @@ is able to handle phenomena related to scope by means of an higher-order type th Annotation of negation in the <fixed-case>IULA</fixed-case> <fixed-case>S</fixed-case>panish Clinical Record Corpus - Montserrat Marimon - Jorge Vivaldi - Núria Bel + MontserratMarimon + JorgeVivaldi + NúriaBel 43–52 W17-1807 10.18653/v1/W17-1807 @@ -2953,11 +2953,11 @@ is able to handle phenomena related to scope by means of an higher-order type th Annotating Negation in <fixed-case>S</fixed-case>panish Clinical Texts - Noa Cruz - Roser Morante - Manuel J. Maña López - Jacinto Mata Vázquez - Carlos L. Parra Calderón + NoaCruz + RoserMorante + Manuel J.Maña López + JacintoMata Vázquez + Carlos L.Parra Calderón 53–58 W17-1808 10.18653/v1/W17-1808 @@ -2965,9 +2965,9 @@ is able to handle phenomena related to scope by means of an higher-order type th Neural Networks for Negation Cue Detection in <fixed-case>C</fixed-case>hinese - Hangfeng He - Federico Fancellu - Bonnie Webber + HangfengHe + FedericoFancellu + BonnieWebber 59–63 W17-1809 10.18653/v1/W17-1809 @@ -2975,9 +2975,9 @@ is able to handle phenomena related to scope by means of an higher-order type th An open-source tool for negation detection: a maximum-margin approach - Martine Enger - Erik Velldal - Lilja Øvrelid + MartineEnger + ErikVelldal + LiljaØvrelid 64–69 W17-1810 10.18653/v1/W17-1810 @@ -3001,8 +3001,8 @@ is able to handle phenomena related to scope by means of an higher-order type th Compositional Semantics using Feature-Based Models from <fixed-case>W</fixed-case>ord<fixed-case>N</fixed-case>et - Pablo Gamallo - Martín Pereira-Fariña + PabloGamallo + MartínPereira-Fariña 1–11 W17-1901 10.18653/v1/W17-1901 @@ -3010,10 +3010,10 @@ is able to handle phenomena related to scope by means of an higher-order type th Automated <fixed-case>W</fixed-case>ord<fixed-case>N</fixed-case>et Construction Using Word Embeddings - Mikhail Khodak - Andrej Risteski - Christiane Fellbaum - Sanjeev Arora + MikhailKhodak + AndrejRisteski + ChristianeFellbaum + SanjeevArora 12–23 W17-1902 10.18653/v1/W17-1902 @@ -3021,8 +3021,8 @@ is able to handle phenomena related to scope by means of an higher-order type th Improving Verb Metaphor Detection by Propagating Abstractness to Words, Phrases and Individual Senses - Maximilian Köper - Sabine Schulte im Walde + MaximilianKöper + SabineSchulte im Walde 24–30 W17-1903 10.18653/v1/W17-1903 @@ -3030,9 +3030,9 @@ is able to handle phenomena related to scope by means of an higher-order type th Improving Clinical Diagnosis Inference through Integration of Structured and Unstructured Knowledge - Yuan Ling - Yuan An - Sadid Hasan + YuanLing + YuanAn + SadidHasan 31–36 W17-1904 10.18653/v1/W17-1904 @@ -3040,9 +3040,9 @@ is able to handle phenomena related to scope by means of an higher-order type th Classifying Lexical-semantic Relationships by Exploiting Sense/Concept Representations - Kentaro Kanada - Tetsunori Kobayashi - Yoshihiko Hayashi + KentaroKanada + TetsunoriKobayashi + YoshihikoHayashi 37–46 W17-1905 10.18653/v1/W17-1905 @@ -3050,8 +3050,8 @@ is able to handle phenomena related to scope by means of an higher-order type th Supervised and unsupervised approaches to measuring usage similarity - Milton King - Paul Cook + MiltonKing + PaulCook 47–52 W17-1906 10.18653/v1/W17-1906 @@ -3059,9 +3059,9 @@ is able to handle phenomena related to scope by means of an higher-order type th Lexical Disambiguation of <fixed-case>I</fixed-case>gbo using Diacritic Restoration - Ignatius Ezeani - Mark Hepple - Ikechukwu Onyenwe + IgnatiusEzeani + MarkHepple + IkechukwuOnyenwe 53–60 W17-1907 10.18653/v1/W17-1907 @@ -3069,10 +3069,10 @@ is able to handle phenomena related to scope by means of an higher-order type th Creating and Validating Multilingual Semantic Representations for Six Languages: Expert versus Non-Expert Crowds - Mahmoud El-Haj - Paul Rayson - Scott Piao - Stephen Wattam + MahmoudEl-Haj + PaulRayson + ScottPiao + StephenWattam 61–71 W17-1908 10.18653/v1/W17-1908 @@ -3080,10 +3080,10 @@ is able to handle phenomena related to scope by means of an higher-order type th Using Linked Disambiguated Distributional Networks for Word Sense Disambiguation - Alexander Panchenko - Stefano Faralli - Simone Paolo Ponzetto - Chris Biemann + AlexanderPanchenko + StefanoFaralli + Simone PaoloPonzetto + ChrisBiemann 72–78 W17-1909 10.18653/v1/W17-1909 @@ -3091,11 +3091,11 @@ is able to handle phenomena related to scope by means of an higher-order type th One Representation per Word - Does it make Sense for Composition? - Thomas Kober - Julie Weeds - John Wilkie - Jeremy Reffin - David Weir + ThomasKober + JulieWeeds + JohnWilkie + JeremyReffin + DavidWeir 79–90 W17-1910 10.18653/v1/W17-1910 @@ -3103,8 +3103,8 @@ is able to handle phenomena related to scope by means of an higher-order type th Elucidating Conceptual Properties from Word Embeddings - Kyoung-Rok Jang - Sung-Hyon Myaeng + Kyoung-RokJang + Sung-HyonMyaeng 91–95 W17-1911 10.18653/v1/W17-1911 @@ -3112,9 +3112,9 @@ is able to handle phenomena related to scope by means of an higher-order type th TTCS<tex-math>^{\mathcal{E}}</tex-math>: a Vectorial Resource for Computing Conceptual Similarity - Enrico Mensa - Daniele P. Radicioni - Antonio Lieto + EnricoMensa + Daniele P.Radicioni + AntonioLieto 96–101 W17-1912 10.18653/v1/W17-1912 @@ -3122,8 +3122,8 @@ is able to handle phenomena related to scope by means of an higher-order type th Measuring the <fixed-case>I</fixed-case>talian-<fixed-case>E</fixed-case>nglish lexical gap for action verbs and its impact on translation - Lorenzo Gregori - Alessandro Panunzi + LorenzoGregori + AlessandroPanunzi 102–109 W17-1913 10.18653/v1/W17-1913 @@ -3131,9 +3131,9 @@ is able to handle phenomena related to scope by means of an higher-order type th Word Sense Filtering Improves Embedding-Based Lexical Substitution - Anne Cocos - Marianna Apidianaki - Chris Callison-Burch + AnneCocos + MariannaApidianaki + ChrisCallison-Burch 110–119 W17-1914 10.18653/v1/W17-1914 @@ -3141,8 +3141,8 @@ is able to handle phenomena related to scope by means of an higher-order type th Supervised and Unsupervised Word Sense Disambiguation on Word Embedding Vectors of Unambigous Synonyms - Aleksander Wawer - Agnieszka Mykowiecka + AleksanderWawer + AgnieszkaMykowiecka 120–125 W17-1915 10.18653/v1/W17-1915 @@ -3168,10 +3168,10 @@ is able to handle phenomena related to scope by means of an higher-order type th The <fixed-case>BURCHAK</fixed-case> corpus: a Challenge Data Set for Interactive Learning of Visually Grounded Word Meanings - Yanchao Yu - Arash Eshghi - Gregory Mills - Oliver Lemon + YanchaoYu + ArashEshghi + GregoryMills + OliverLemon 1–10 W17-2001 10.18653/v1/W17-2001 @@ -3179,8 +3179,8 @@ is able to handle phenomena related to scope by means of an higher-order type th The Use of Object Labels and Spatial Prepositions as Keywords in a Web-Retrieval-Based Image Caption Generation System - Brandon Birmingham - Adrian Muscat + BrandonBirmingham + AdrianMuscat 11–20 W17-2002 10.18653/v1/W17-2002 @@ -3188,9 +3188,9 @@ is able to handle phenomena related to scope by means of an higher-order type th Learning to Recognize Animals by Watching Documentaries: Using Subtitles as Weak Supervision - Aparna Nurani Venkitasubramanian - Tinne Tuytelaars - Marie-Francine Moens + AparnaNurani Venkitasubramanian + TinneTuytelaars + Marie-FrancineMoens 21–30 W17-2003 10.18653/v1/W17-2003 @@ -3198,11 +3198,11 @@ is able to handle phenomena related to scope by means of an higher-order type th Human Evaluation of Multi-modal Neural Machine Translation: A Case-Study on E-Commerce Listing Titles - Iacer Calixto - Daniel Stein - Evgeny Matusov - Sheila Castilho - Andy Way + IacerCalixto + DanielStein + EvgenyMatusov + SheilaCastilho + AndyWay 31–37 W17-2004 10.18653/v1/W17-2004 @@ -3210,10 +3210,10 @@ is able to handle phenomena related to scope by means of an higher-order type th The <fixed-case>B</fixed-case>reaking<fixed-case>N</fixed-case>ews Dataset - Arnau Ramisa - Fei Yan - Francesc Moreno-Noguer - Krystian Mikolajczyk + ArnauRamisa + FeiYan + FrancescMoreno-Noguer + KrystianMikolajczyk 38–39 W17-2005 10.18653/v1/W17-2005 @@ -3221,9 +3221,9 @@ is able to handle phenomena related to scope by means of an higher-order type th Automatic identification of head movements in video-recorded conversations: can words help? - Patrizia Paggio - Costanza Navarretta - Bart Jongejan + PatriziaPaggio + CostanzaNavarretta + BartJongejan 40–42 W17-2006 10.18653/v1/W17-2006 @@ -3231,10 +3231,10 @@ is able to handle phenomena related to scope by means of an higher-order type th Multi-Modal Fashion Product Retrieval - Antonio Rubio Romano - LongLong Yu - Edgar Simo-Serra - Francesc Moreno-Noguer + AntonioRubio Romano + LongLongYu + EdgarSimo-Serra + FrancescMoreno-Noguer 43–45 W17-2007 10.18653/v1/W17-2007 @@ -3262,10 +3262,10 @@ is able to handle phenomena related to scope by means of an higher-order type th Metaphor Detection in a Poetry Corpus - Vaibhav Kesarwani - Diana Inkpen - Stan Szpakowicz - Chris Tanasescu + VaibhavKesarwani + DianaInkpen + StanSzpakowicz + ChrisTanasescu 1–9 W17-2201 10.18653/v1/W17-2201 @@ -3273,10 +3273,10 @@ is able to handle phenomena related to scope by means of an higher-order type th Machine Translation and Automated Analysis of the <fixed-case>S</fixed-case>umerian Language - Émilie Pagé-Perron - Maria Sukhareva - Ilya Khait - Christian Chiarcos + ÉmiliePagé-Perron + MariaSukhareva + IlyaKhait + ChristianChiarcos 10–16 W17-2202 10.18653/v1/W17-2202 @@ -3284,9 +3284,9 @@ is able to handle phenomena related to scope by means of an higher-order type th Investigating the Relationship between Literary Genres and Emotional Plot Development - Evgeny Kim - Sebastian Padó - Roman Klinger + EvgenyKim + SebastianPadó + RomanKlinger 17–26 W17-2203 10.18653/v1/W17-2203 @@ -3294,10 +3294,10 @@ is able to handle phenomena related to scope by means of an higher-order type th Enjambment Detection in a Large Diachronic Corpus of <fixed-case>S</fixed-case>panish Sonnets - Pablo Ruiz - Clara Martínez Cantón - Thierry Poibeau - Elena González-Blanco + PabloRuiz + ClaraMartínez Cantón + ThierryPoibeau + ElenaGonzález-Blanco 27–32 W17-2204 10.18653/v1/W17-2204 @@ -3305,8 +3305,8 @@ is able to handle phenomena related to scope by means of an higher-order type th Plotting <fixed-case>M</fixed-case>arkson’s “Mistress” - Conor Kelleher - Mark Keane + ConorKelleher + MarkKeane 33–39 W17-2205 10.18653/v1/W17-2205 @@ -3314,11 +3314,11 @@ is able to handle phenomena related to scope by means of an higher-order type th Annotation Challenges for Reconstructing the Structural Elaboration of Middle Low <fixed-case>G</fixed-case>erman - Nina Seemann - Marie-Luis Merten - Michaela Geierhos - Doris Tophinke - Eyke Hüllermeier + NinaSeemann + Marie-LuisMerten + MichaelaGeierhos + DorisTophinke + EykeHüllermeier 40–45 W17-2206 10.18653/v1/W17-2206 @@ -3326,7 +3326,7 @@ is able to handle phenomena related to scope by means of an higher-order type th Phonological Soundscapes in Medieval Poetry - Christopher Hench + ChristopherHench 46–56 W17-2207 10.18653/v1/W17-2207 @@ -3334,10 +3334,10 @@ is able to handle phenomena related to scope by means of an higher-order type th An End-to-end Environment for Research Question-Driven Entity Extraction and Network Analysis - Andre Blessing - Nora Echelmeyer - Markus John - Nils Reiter + AndreBlessing + NoraEchelmeyer + MarkusJohn + NilsReiter 57–67 W17-2208 10.18653/v1/W17-2208 @@ -3347,8 +3347,8 @@ is able to handle phenomena related to scope by means of an higher-order type th Modeling intra-textual variation with entropy and surprisal: topical vs. stylistic patterns - Stefania Degaetano-Ortlieb - Elke Teich + StefaniaDegaetano-Ortlieb + ElkeTeich 68–77 W17-2209 10.18653/v1/W17-2209 @@ -3356,8 +3356,8 @@ is able to handle phenomena related to scope by means of an higher-order type th Finding a Character’s Voice: Stylome Classification on Literary Characters - Liviu P. Dinu - Ana Sabina Uban + Liviu P.Dinu + Ana SabinaUban 78–82 W17-2210 10.18653/v1/W17-2210 @@ -3365,7 +3365,7 @@ is able to handle phenomena related to scope by means of an higher-order type th An Ontology-Based Method for Extracting and Classifying Domain-Specific Compositional Nominal Compounds - Maria Pia di Buono + Maria Piadi Buono 83–88 W17-2211 10.18653/v1/W17-2211 @@ -3373,8 +3373,8 @@ is able to handle phenomena related to scope by means of an higher-order type th Speeding up corpus development for linguistic research: language documentation and acquisition in <fixed-case>R</fixed-case>omansh Tuatschin - Géraldine Walther - Benoît Sagot + GéraldineWalther + BenoîtSagot 89–94 W17-2212 10.18653/v1/W17-2212 @@ -3382,12 +3382,12 @@ is able to handle phenomena related to scope by means of an higher-order type th Distantly Supervised <fixed-case>POS</fixed-case> Tagging of Low-Resource Languages under Extreme Data Sparsity: The Case of <fixed-case>H</fixed-case>ittite - Maria Sukhareva - Francesco Fuscagni - Johannes Daxenberger - Susanne Görke - Doris Prechel - Iryna Gurevych + MariaSukhareva + FrancescoFuscagni + JohannesDaxenberger + SusanneGörke + DorisPrechel + IrynaGurevych 95–104 W17-2213 10.18653/v1/W17-2213 @@ -3395,9 +3395,9 @@ is able to handle phenomena related to scope by means of an higher-order type th A Dataset for <fixed-case>S</fixed-case>anskrit Word Segmentation - Amrith Krishna - Pavan Kumar Satuluri - Pawan Goyal + AmrithKrishna + Pavan KumarSatuluri + PawanGoyal 105–114 W17-2214 10.18653/v1/W17-2214 @@ -3405,8 +3405,8 @@ is able to handle phenomena related to scope by means of an higher-order type th Lexical Correction of Polish Twitter Political Data - Maciej Ogrodniczuk - Mateusz Kopeć + MaciejOgrodniczuk + MateuszKopeć 115–125 W17-2215 10.18653/v1/W17-2215 @@ -3432,11 +3432,11 @@ is able to handle phenomena related to scope by means of an higher-order type th Target word prediction and paraphasia classification in spoken discourse - Joel Adams - Steven Bedrick - Gerasimos Fergadiotis - Kyle Gorman - Jan van Santen + JoelAdams + StevenBedrick + GerasimosFergadiotis + KyleGorman + Janvan Santen 1–8 W17-2301 10.18653/v1/W17-2301 @@ -3444,9 +3444,9 @@ is able to handle phenomena related to scope by means of an higher-order type th Extracting Drug-Drug Interactions with Attention <fixed-case>CNN</fixed-case>s - Masaki Asada - Makoto Miwa - Yutaka Sasaki + MasakiAsada + MakotoMiwa + YutakaSasaki 9–18 W17-2302 10.18653/v1/W17-2302 @@ -3454,9 +3454,9 @@ is able to handle phenomena related to scope by means of an higher-order type th Insights into Analogy Completion from the Biomedical Domain - Denis Newman-Griffis - Albert Lai - Eric Fosler-Lussier + DenisNewman-Griffis + AlbertLai + EricFosler-Lussier 19–28 W17-2303 10.18653/v1/W17-2303 @@ -3464,8 +3464,8 @@ is able to handle phenomena related to scope by means of an higher-order type th Deep learning for extracting protein-protein interactions from biomedical literature - Yifan Peng - Zhiyong Lu + YifanPeng + ZhiyongLu 29–38 W17-2304 10.18653/v1/W17-2304 @@ -3473,9 +3473,9 @@ is able to handle phenomena related to scope by means of an higher-order type th Stacking With Auxiliary Features for Entity Linking in the Medical Domain - Nazneen Fatema Rajani - Mihaela Bornea - Ken Barker + Nazneen FatemaRajani + MihaelaBornea + KenBarker 39–47 W17-2305 10.18653/v1/W17-2305 @@ -3483,11 +3483,11 @@ is able to handle phenomena related to scope by means of an higher-order type th Results of the fifth edition of the <fixed-case>B</fixed-case>io<fixed-case>ASQ</fixed-case> Challenge - Anastasios Nentidis - Konstantinos Bougiatiotis - Anastasia Krithara - Georgios Paliouras - Ioannis Kakadiaris + AnastasiosNentidis + KonstantinosBougiatiotis + AnastasiaKrithara + GeorgiosPaliouras + IoannisKakadiaris 48–57 W17-2306 10.18653/v1/W17-2306 @@ -3496,12 +3496,12 @@ is able to handle phenomena related to scope by means of an higher-order type th Tackling Biomedical Text Summarization: <fixed-case>OAQA</fixed-case> at <fixed-case>B</fixed-case>io<fixed-case>ASQ</fixed-case> 5<fixed-case>B</fixed-case> - Khyathi Chandu - Aakanksha Naik - Aditya Chandrasekar - Zi Yang - Niloy Gupta - Eric Nyberg + KhyathiChandu + AakankshaNaik + AdityaChandrasekar + ZiYang + NiloyGupta + EricNyberg 58–66 W17-2307 10.18653/v1/W17-2307 @@ -3509,7 +3509,7 @@ is able to handle phenomena related to scope by means of an higher-order type th <fixed-case>M</fixed-case>acquarie University at <fixed-case>B</fixed-case>io<fixed-case>ASQ</fixed-case> 5b – Query-based Summarisation Techniques for Selecting the Ideal Answers - Diego Mollá + DiegoMollá 67–75 W17-2308 10.18653/v1/W17-2308 @@ -3517,9 +3517,9 @@ is able to handle phenomena related to scope by means of an higher-order type th Neural Question Answering at <fixed-case>B</fixed-case>io<fixed-case>ASQ</fixed-case> 5<fixed-case>B</fixed-case> - Georg Wiese - Dirk Weissenborn - Mariana Neves + GeorgWiese + DirkWeissenborn + MarianaNeves 76–79 W17-2309 10.18653/v1/W17-2309 @@ -3528,12 +3528,12 @@ is able to handle phenomena related to scope by means of an higher-order type th End-to-End System for Bacteria Habitat Extraction - Farrokh Mehryary - Kai Hakala - Suwisa Kaewphan - Jari Björne - Tapio Salakoski - Filip Ginter + FarrokhMehryary + KaiHakala + SuwisaKaewphan + JariBjörne + TapioSalakoski + FilipGinter 80–90 W17-2310 10.18653/v1/W17-2310 @@ -3542,11 +3542,11 @@ is able to handle phenomena related to scope by means of an higher-order type th Creation and evaluation of a dictionary-based tagger for virus species and proteins - Helen Cook - Rūdolfs Bērziņš - Cristina Leal Rodrıguez - Juan Miguel Cejuela - Lars Juhl Jensen + HelenCook + RūdolfsBērziņš + Cristina LealRodrıguez + Juan MiguelCejuela + Lars JuhlJensen 91–98 W17-2311 10.18653/v1/W17-2311 @@ -3554,9 +3554,9 @@ is able to handle phenomena related to scope by means of an higher-order type th Representation of complex terms in a vector space structured by an ontology for a normalization task - Arnaud Ferré - Pierre Zweigenbaum - Claire Nédellec + ArnaudFerré + PierreZweigenbaum + ClaireNédellec 99–106 W17-2312 10.18653/v1/W17-2312 @@ -3564,8 +3564,8 @@ is able to handle phenomena related to scope by means of an higher-order type th Improving Correlation with Human Judgments by Integrating Semantic Similarity with Second–Order Vectors - Bridget McInnes - Ted Pedersen + BridgetMcInnes + TedPedersen 107–116 W17-2313 10.18653/v1/W17-2313 @@ -3573,9 +3573,9 @@ is able to handle phenomena related to scope by means of an higher-order type th Proactive Learning for Named Entity Recognition - Maolin Li - Nhung Nguyen - Sophia Ananiadou + MaolinLi + NhungNguyen + SophiaAnaniadou 117–125 W17-2314 10.18653/v1/W17-2314 @@ -3583,10 +3583,10 @@ is able to handle phenomena related to scope by means of an higher-order type th Biomedical Event Extraction using Abstract Meaning Representation - Sudha Rao - Daniel Marcu - Kevin Knight - Hal Daumé III + SudhaRao + DanielMarcu + KevinKnight + HalDaumé III 126–135 W17-2315 10.18653/v1/W17-2315 @@ -3594,11 +3594,11 @@ is able to handle phenomena related to scope by means of an higher-order type th Detecting Personal Medication Intake in Twitter: An Annotated Corpus and Baseline Classification System - Ari Klein - Abeed Sarker - Masoud Rouhizadeh - Karen O’Connor - Graciela Gonzalez + AriKlein + AbeedSarker + MasoudRouhizadeh + KarenO’Connor + GracielaGonzalez 136–142 W17-2316 10.18653/v1/W17-2316 @@ -3606,9 +3606,9 @@ is able to handle phenomena related to scope by means of an higher-order type th Unsupervised Context-Sensitive Spelling Correction of Clinical Free-Text with Word and Character N-Gram Embeddings - Pieter Fivez - Simon Šuster - Walter Daelemans + PieterFivez + SimonŠuster + WalterDaelemans 143–148 W17-2317 10.18653/v1/W17-2317 @@ -3616,11 +3616,11 @@ is able to handle phenomena related to scope by means of an higher-order type th Characterization of Divergence in Impaired Speech of <fixed-case>ALS</fixed-case> Patients - Archna Bhatia - Bonnie Dorr - Kristy Hollingshead - Samuel L. Phillips - Barbara McKenzie + ArchnaBhatia + BonnieDorr + KristyHollingshead + Samuel L.Phillips + BarbaraMcKenzie 149–158 W17-2318 10.18653/v1/W17-2318 @@ -3628,11 +3628,11 @@ is able to handle phenomena related to scope by means of an higher-order type th Deep Learning for Punctuation Restoration in Medical Reports - Wael Salloum - Greg Finley - Erik Edwards - Mark Miller - David Suendermann-Oeft + WaelSalloum + GregFinley + ErikEdwards + MarkMiller + DavidSuendermann-Oeft 159–164 W17-2319 10.18653/v1/W17-2319 @@ -3640,10 +3640,10 @@ is able to handle phenomena related to scope by means of an higher-order type th Unsupervised Domain Adaptation for Clinical Negation Detection - Timothy Miller - Steven Bethard - Hadi Amiri - Guergana Savova + TimothyMiller + StevenBethard + HadiAmiri + GuerganaSavova 165–170 W17-2320 10.18653/v1/W17-2320 @@ -3651,13 +3651,13 @@ is able to handle phenomena related to scope by means of an higher-order type th <fixed-case>B</fixed-case>io<fixed-case>C</fixed-case>reative <fixed-case>VI</fixed-case> Precision Medicine Track: creating a training corpus for mining protein-protein interactions affected by mutations - Rezarta Islamaj Doğan - Andrew Chatr-aryamontri - Sun Kim - Chih-Hsuan Wei - Yifan Peng - Donald Comeau - Zhiyong Lu + RezartaIslamaj Doğan + AndrewChatr-aryamontri + SunKim + Chih-HsuanWei + YifanPeng + DonaldComeau + ZhiyongLu 171–175 W17-2321 10.18653/v1/W17-2321 @@ -3665,8 +3665,8 @@ is able to handle phenomena related to scope by means of an higher-order type th Painless Relation Extraction with Kindred - Jake Lever - Steven Jones + JakeLever + StevenJones 176–183 W17-2322 10.18653/v1/W17-2322 @@ -3674,9 +3674,9 @@ is able to handle phenomena related to scope by means of an higher-order type th Noise Reduction Methods for Distantly Supervised Biomedical Relation Extraction - Gang Li - Cathy Wu - K. Vijay-Shanker + GangLi + CathyWu + K.Vijay-Shanker 184–193 W17-2323 10.18653/v1/W17-2323 @@ -3684,11 +3684,11 @@ is able to handle phenomena related to scope by means of an higher-order type th Role-Preserving Redaction of Medical Records to Enable Ontology-Driven Processing - Seth Polsley - Atif Tahir - Muppala Raju - Akintayo Akinleye - Duane Steward + SethPolsley + AtifTahir + MuppalaRaju + AkintayoAkinleye + DuaneSteward 194–199 W17-2324 10.18653/v1/W17-2324 @@ -3696,10 +3696,10 @@ is able to handle phenomena related to scope by means of an higher-order type th Annotation of pain and anesthesia events for surgery-related processes and outcomes extraction - Wen-wai Yim - Dario Tedesco - Catherine Curtin - Tina Hernandez-Boussard + Wen-waiYim + DarioTedesco + CatherineCurtin + TinaHernandez-Boussard 200–205 W17-2325 10.18653/v1/W17-2325 @@ -3707,11 +3707,11 @@ is able to handle phenomena related to scope by means of an higher-order type th Identifying Comparative Structures in Biomedical Text - Samir Gupta - A.S.M. Ashique Mahmood - Karen Ross - Cathy Wu - K. Vijay-Shanker + SamirGupta + A.S.M. AshiqueMahmood + KarenRoss + CathyWu + K.Vijay-Shanker 206–215 W17-2326 10.18653/v1/W17-2326 @@ -3719,13 +3719,13 @@ is able to handle phenomena related to scope by means of an higher-order type th Tagging Funding Agencies and Grants in Scientific Articles using Sequential Learning Models - Subhradeep Kayal - Zubair Afzal - George Tsatsaronis - Sophia Katrenko - Pascal Coupet - Marius Doornenbal - Michelle Gregory + SubhradeepKayal + ZubairAfzal + GeorgeTsatsaronis + SophiaKatrenko + PascalCoupet + MariusDoornenbal + MichelleGregory 216–221 W17-2327 10.18653/v1/W17-2327 @@ -3733,10 +3733,10 @@ is able to handle phenomena related to scope by means of an higher-order type th Deep Learning for Biomedical Information Retrieval: Learning Textual Relevance from Click Logs - Sunil Mohan - Nicolas Fiorini - Sun Kim - Zhiyong Lu + SunilMohan + NicolasFiorini + SunKim + ZhiyongLu 222–231 W17-2328 10.18653/v1/W17-2328 @@ -3744,10 +3744,10 @@ is able to handle phenomena related to scope by means of an higher-order type th Detecting Dementia through Retrospective Analysis of Routine Blog Posts by Bloggers with Dementia - Vaden Masrani - Gabriel Murray - Thalia Field - Giuseppe Carenini + VadenMasrani + GabrielMurray + ThaliaField + GiuseppeCarenini 232–237 W17-2329 10.18653/v1/W17-2329 @@ -3755,9 +3755,9 @@ is able to handle phenomena related to scope by means of an higher-order type th Protein Word Detection using Text Segmentation Techniques - Devi Ganesan - Ashish V. Tendulkar - Sutanu Chakraborti + DeviGanesan + Ashish V.Tendulkar + SutanuChakraborti 238–246 W17-2330 10.18653/v1/W17-2330 @@ -3765,8 +3765,8 @@ is able to handle phenomena related to scope by means of an higher-order type th External Evaluation of Event Extraction Classifiers for Automatic Pathway Curation: An extended study of the m<fixed-case>TOR</fixed-case> pathway - Wojciech Kusa - Michael Spranger + WojciechKusa + MichaelSpranger 247–256 W17-2331 10.18653/v1/W17-2331 @@ -3774,8 +3774,8 @@ is able to handle phenomena related to scope by means of an higher-order type th Toward Automated Early Sepsis Alerting: Identifying Infection Patients from Nursing Notes - Emilia Apostolova - Tom Velez + EmiliaApostolova + TomVelez 257–262 W17-2332 10.18653/v1/W17-2332 @@ -3783,10 +3783,10 @@ is able to handle phenomena related to scope by means of an higher-order type th Enhancing Automatic <fixed-case>ICD</fixed-case>-9-<fixed-case>CM</fixed-case> Code Assignment for Medical Texts with <fixed-case>P</fixed-case>ub<fixed-case>M</fixed-case>ed - Danchen Zhang - Daqing He - Sanqiang Zhao - Lei Li + DanchenZhang + DaqingHe + SanqiangZhao + LeiLi 263–271 W17-2333 10.18653/v1/W17-2333 @@ -3794,9 +3794,9 @@ is able to handle phenomena related to scope by means of an higher-order type th Evaluating Feature Extraction Methods for Knowledge-based Biomedical Word Sense Disambiguation - Sam Henry - Clint Cuffy - Bridget McInnes + SamHenry + ClintCuffy + BridgetMcInnes 272–281 W17-2334 10.18653/v1/W17-2334 @@ -3804,11 +3804,11 @@ is able to handle phenomena related to scope by means of an higher-order type th Investigating the Documentation of Electronic Cigarette Use in the Veteran Affairs Electronic Health Record: A Pilot Study - Danielle Mowery - Brett South - Olga Patterson - Shu-Hong Zhu - Mike Conway + DanielleMowery + BrettSouth + OlgaPatterson + Shu-HongZhu + MikeConway 282–286 W17-2335 10.18653/v1/W17-2335 @@ -3816,11 +3816,11 @@ is able to handle phenomena related to scope by means of an higher-order type th Automated Preamble Detection in Dictated Medical Reports - Wael Salloum - Greg Finley - Erik Edwards - Mark Miller - David Suendermann-Oeft + WaelSalloum + GregFinley + ErikEdwards + MarkMiller + DavidSuendermann-Oeft 287–295 W17-2336 10.18653/v1/W17-2336 @@ -3828,8 +3828,8 @@ is able to handle phenomena related to scope by means of an higher-order type th A Biomedical Question Answering System in <fixed-case>B</fixed-case>io<fixed-case>ASQ</fixed-case> 2017 - Mourad Sarrouti - Said Ouatik El Alaoui + MouradSarrouti + SaidOuatik El Alaoui 296–301 W17-2337 10.18653/v1/W17-2337 @@ -3837,11 +3837,11 @@ is able to handle phenomena related to scope by means of an higher-order type th Adapting Pre-trained Word Embeddings For Use In Medical Coding - Kevin Patel - Divya Patel - Mansi Golakiya - Pushpak Bhattacharyya - Nilesh Birari + KevinPatel + DivyaPatel + MansiGolakiya + PushpakBhattacharyya + NileshBirari 302–306 W17-2338 10.18653/v1/W17-2338 @@ -3849,8 +3849,8 @@ is able to handle phenomena related to scope by means of an higher-order type th Initializing neural networks for hierarchical multi-label text classification - Simon Baker - Anna Korhonen + SimonBaker + AnnaKorhonen 307–315 W17-2339 10.18653/v1/W17-2339 @@ -3858,9 +3858,9 @@ is able to handle phenomena related to scope by means of an higher-order type th Biomedical Event Trigger Identification Using Bidirectional Recurrent Neural Network Based Models - Rahul V S S Patchigolla - Sunil Sahu - Ashish Anand + RahulV S S Patchigolla + SunilSahu + AshishAnand 316–321 W17-2340 10.18653/v1/W17-2340 @@ -3868,11 +3868,11 @@ is able to handle phenomena related to scope by means of an higher-order type th Representations of Time Expressions for Temporal Relation Extraction with Convolutional Neural Networks - Chen Lin - Timothy Miller - Dmitriy Dligach - Steven Bethard - Guergana Savova + ChenLin + TimothyMiller + DmitriyDligach + StevenBethard + GuerganaSavova 322–327 W17-2341 10.18653/v1/W17-2341 @@ -3880,10 +3880,10 @@ is able to handle phenomena related to scope by means of an higher-order type th Automatic Diagnosis Coding of Radiology Reports: A Comparison of Deep Learning and Conventional Classification Methods - Sarvnaz Karimi - Xiang Dai - Hamed Hassanzadeh - Anthony Nguyen + SarvnazKarimi + XiangDai + HamedHassanzadeh + AnthonyNguyen 328–332 W17-2342 10.18653/v1/W17-2342 @@ -3891,9 +3891,9 @@ is able to handle phenomena related to scope by means of an higher-order type th Automatic classification of doctor-patient questions for a virtual patient record query task - Leonardo Campillos Llanos - Sophie Rosset - Pierre Zweigenbaum + LeonardoCampillos Llanos + SophieRosset + PierreZweigenbaum 333–341 W17-2343 10.18653/v1/W17-2343 @@ -3901,10 +3901,10 @@ is able to handle phenomena related to scope by means of an higher-order type th Assessing the performance of <fixed-case>O</fixed-case>lelo, a real-time biomedical question answering application - Mariana Neves - Fabian Eckert - Hendrik Folkerts - Matthias Uflacker + MarianaNeves + FabianEckert + HendrikFolkerts + MatthiasUflacker 342–350 W17-2344 10.18653/v1/W17-2344 @@ -3913,8 +3913,8 @@ is able to handle phenomena related to scope by means of an higher-order type th Clinical Event Detection with Hybrid Neural Architecture - Adyasha Maharana - Meliha Yetisgen + AdyashaMaharana + MelihaYetisgen 351–355 W17-2345 10.18653/v1/W17-2345 @@ -3922,9 +3922,9 @@ is able to handle phenomena related to scope by means of an higher-order type th Extracting Personal Medical Events for User Timeline Construction using Minimal Supervision - Aakanksha Naik - Chris Bogart - Carolyn Rose + AakankshaNaik + ChrisBogart + CarolynRose 356–364 W17-2346 10.18653/v1/W17-2346 @@ -3932,13 +3932,13 @@ is able to handle phenomena related to scope by means of an higher-order type th Detecting mentions of pain and acute confusion in <fixed-case>F</fixed-case>innish clinical text - Hans Moen - Kai Hakala - Farrokh Mehryary - Laura-Maria Peltonen - Tapio Salakoski - Filip Ginter - Sanna Salanterä + HansMoen + KaiHakala + FarrokhMehryary + Laura-MariaPeltonen + TapioSalakoski + FilipGinter + SannaSalanterä 365–372 W17-2347 10.18653/v1/W17-2347 @@ -3946,11 +3946,11 @@ is able to handle phenomena related to scope by means of an higher-order type th A Multi-strategy Query Processing Approach for Biomedical Question Answering: <fixed-case>USTB</fixed-case>_<fixed-case>PRIR</fixed-case> at <fixed-case>B</fixed-case>io<fixed-case>ASQ</fixed-case> 2017 Task 5<fixed-case>B</fixed-case> - Zan-Xia Jin - Bo-Wen Zhang - Fan Fang - Le-Le Zhang - Xu-Cheng Yin + Zan-XiaJin + Bo-WenZhang + FanFang + Le-LeZhang + Xu-ChengYin 373–380 W17-2348 10.18653/v1/W17-2348 @@ -3976,11 +3976,11 @@ is able to handle phenomena related to scope by means of an higher-order type th On the “Calligraphy” of Books - Vanessa Queiroz Marinho - Henrique Ferraz de Arruda - Thales Sinelli - Luciano da Fontoura Costa - Diego Raphael Amancio + Vanessa QueirozMarinho + Henrique Ferrazde Arruda + ThalesSinelli + Luciano da FontouraCosta + Diego RaphaelAmancio 1–10 W17-2401 10.18653/v1/W17-2401 @@ -3988,11 +3988,11 @@ is able to handle phenomena related to scope by means of an higher-order type th Adapting predominant and novel sense discovery algorithms for identifying corpus-specific sense differences - Binny Mathew - Suman Kalyan Maity - Pratip Sarkar - Animesh Mukherjee - Pawan Goyal + BinnyMathew + Suman KalyanMaity + PratipSarkar + AnimeshMukherjee + PawanGoyal 11–20 W17-2402 10.18653/v1/W17-2402 @@ -4000,9 +4000,9 @@ is able to handle phenomena related to scope by means of an higher-order type th Merging knowledge bases in different languages - Jerónimo Hernández-González - Estevam R. Hruschka Jr. - Tom M. Mitchell + JerónimoHernández-González + Estevam R.Hruschka Jr. + Tom M.Mitchell 21–29 W17-2403 10.18653/v1/W17-2403 @@ -4010,8 +4010,8 @@ is able to handle phenomena related to scope by means of an higher-order type th Parameter Free Hierarchical Graph-Based Clustering for Analyzing Continuous Word Embeddings - Thomas Alexander Trost - Dietrich Klakow + Thomas AlexanderTrost + DietrichKlakow 30–38 W17-2404 10.18653/v1/W17-2404 @@ -4019,9 +4019,9 @@ is able to handle phenomena related to scope by means of an higher-order type th Spectral Graph-Based Method of Multimodal Word Embedding - Kazuki Fukui - Takamasa Oshikiri - Hidetoshi Shimodaira + KazukiFukui + TakamasaOshikiri + HidetoshiShimodaira 39–44 W17-2405 10.18653/v1/W17-2405 @@ -4029,8 +4029,8 @@ is able to handle phenomena related to scope by means of an higher-order type th Graph Methods for Multilingual <fixed-case>F</fixed-case>rame<fixed-case>N</fixed-case>ets - Collin Baker - Michael Ellsworth + CollinBaker + MichaelEllsworth 45–50 W17-2406 10.18653/v1/W17-2406 @@ -4039,8 +4039,8 @@ is able to handle phenomena related to scope by means of an higher-order type th Extract with Order for Coherent Multi-Document Summarization - Mir Tafseer Nayeem - Yllias Chali + Mir TafseerNayeem + YlliasChali 51–56 W17-2407 10.18653/v1/W17-2407 @@ -4048,8 +4048,8 @@ is able to handle phenomena related to scope by means of an higher-order type th Work Hard, Play Hard: Email Classification on the Avocado and <fixed-case>E</fixed-case>nron Corpora - Sakhar Alkhereyf - Owen Rambow + SakharAlkhereyf + OwenRambow 57–65 W17-2408 10.18653/v1/W17-2408 @@ -4057,13 +4057,13 @@ is able to handle phenomena related to scope by means of an higher-order type th A Graph Based Semi-Supervised Approach for Analysis of Derivational Nouns in <fixed-case>S</fixed-case>anskrit - Amrith Krishna - Pavankumar Satuluri - Harshavardhan Ponnada - Muneeb Ahmed - Gulab Arora - Kaustubh Hiware - Pawan Goyal + AmrithKrishna + PavankumarSatuluri + HarshavardhanPonnada + MuneebAhmed + GulabArora + KaustubhHiware + PawanGoyal 66–75 W17-2409 10.18653/v1/W17-2409 @@ -4071,8 +4071,8 @@ is able to handle phenomena related to scope by means of an higher-order type th Evaluating text coherence based on semantic similarity graph - Jan Wira Gotama Putra - Takenobu Tokunaga + Jan Wira GotamaPutra + TakenobuTokunaga 76–85 W17-2410 10.18653/v1/W17-2410 @@ -4084,9 +4084,9 @@ is able to handle phenomena related to scope by means of an higher-order type th Proceedings of the 10th Workshop on Building and Using Comparable Corpora W17-25 - Serge Sharoff - Pierre Zweigenbaum - Reinhard Rapp + SergeSharoff + PierreZweigenbaum + ReinhardRapp 10.18653/v1/W17-25 Association for Computational Linguistics
Vancouver, Canada
@@ -4098,7 +4098,7 @@ is able to handle phenomena related to scope by means of an higher-order type th Users and Data: The Two Neglected Children of Bilingual Natural Language Processing Research - Phillippe Langlais + PhillippeLanglais 1–5 W17-2501 10.18653/v1/W17-2501 @@ -4106,10 +4106,10 @@ is able to handle phenomena related to scope by means of an higher-order type th Deep Investigation of Cross-Language Plagiarism Detection Methods - Jérémy Ferrero - Laurent Besacier - Didier Schwab - Frédéric Agnès + JérémyFerrero + LaurentBesacier + DidierSchwab + FrédéricAgnès 6–15 W17-2502 10.18653/v1/W17-2502 @@ -4118,8 +4118,8 @@ is able to handle phenomena related to scope by means of an higher-order type th Sentence Alignment using Unfolding Recursive Autoencoders - Jeenu Grover - Pabitra Mitra + JeenuGrover + PabitraMitra 16–20 W17-2503 10.18653/v1/W17-2503 @@ -4127,8 +4127,8 @@ is able to handle phenomena related to scope by means of an higher-order type th Acquisition of Translation Lexicons for Historically Unwritten Languages via Bridging Loanwords - Michael Bloodgood - Benjamin Strauss + MichaelBloodgood + BenjaminStrauss 21–25 W17-2504 10.18653/v1/W17-2504 @@ -4137,7 +4137,7 @@ is able to handle phenomena related to scope by means of an higher-order type th Toward a Comparable Corpus of <fixed-case>L</fixed-case>atvian, <fixed-case>R</fixed-case>ussian and <fixed-case>E</fixed-case>nglish Tweets - Dmitrijs Milajevs + DmitrijsMilajevs 26–30 W17-2505 10.18653/v1/W17-2505 @@ -4145,9 +4145,9 @@ is able to handle phenomena related to scope by means of an higher-order type th Automatic Extraction of Parallel Speech Corpora from Dubbed Movies - Alp Öktem - Mireia Farrús - Leo Wanner + AlpÖktem + MireiaFarrús + LeoWanner 31–35 W17-2506 10.18653/v1/W17-2506 @@ -4156,7 +4156,7 @@ is able to handle phenomena related to scope by means of an higher-order type th A parallel collection of clinical trials in <fixed-case>P</fixed-case>ortuguese and <fixed-case>E</fixed-case>nglish - Mariana Neves + MarianaNeves 36–40 W17-2507 10.18653/v1/W17-2507 @@ -4165,9 +4165,9 @@ is able to handle phenomena related to scope by means of an higher-order type th Weighted Set-Theoretic Alignment of Comparable Sentences - Andoni Azpeitia - Thierry Etchegoyhen - Eva Martínez Garcia + AndoniAzpeitia + ThierryEtchegoyhen + EvaMartínez Garcia 41–45 W17-2508 10.18653/v1/W17-2508 @@ -4175,8 +4175,8 @@ is able to handle phenomena related to scope by means of an higher-order type th <fixed-case>BUCC</fixed-case> 2017 Shared Task: a First Attempt Toward a Deep Learning Framework for Identifying Parallel Sentences in Comparable Corpora - Francis Grégoire - Philippe Langlais + FrancisGrégoire + PhilippeLanglais 46–50 W17-2509 10.18653/v1/W17-2509 @@ -4184,8 +4184,8 @@ is able to handle phenomena related to scope by means of an higher-order type th z<fixed-case>NLP</fixed-case>: Identifying Parallel Sentences in <fixed-case>C</fixed-case>hinese-<fixed-case>E</fixed-case>nglish Comparable Corpora - Zheng Zhang - Pierre Zweigenbaum + ZhengZhang + PierreZweigenbaum 51–55 W17-2510 10.18653/v1/W17-2510 @@ -4193,9 +4193,9 @@ is able to handle phenomena related to scope by means of an higher-order type th <fixed-case>BUCC</fixed-case>2017: A Hybrid Approach for Identifying Parallel Sentences in Comparable Corpora - Sainik Mahata - Dipankar Das - Sivaji Bandyopadhyay + SainikMahata + DipankarDas + SivajiBandyopadhyay 56–59 W17-2511 10.18653/v1/W17-2511 @@ -4203,9 +4203,9 @@ is able to handle phenomena related to scope by means of an higher-order type th Overview of the Second <fixed-case>BUCC</fixed-case> Shared Task: Spotting Parallel Sentences in Comparable Corpora - Pierre Zweigenbaum - Serge Sharoff - Reinhard Rapp + PierreZweigenbaum + SergeSharoff + ReinhardRapp 60–67 W17-2512 10.18653/v1/W17-2512 @@ -4237,7 +4237,7 @@ is able to handle phenomena related to scope by means of an higher-order type th Sense Contextualization in a Dependency-Based Compositional Distributional Model - Pablo Gamallo + PabloGamallo 1–9 W17-2601 10.18653/v1/W17-2601 @@ -4245,7 +4245,7 @@ is able to handle phenomena related to scope by means of an higher-order type th Context encoders as a simple but powerful extension of word2vec - Franziska Horn + FranziskaHorn 10–14 W17-2602 10.18653/v1/W17-2602 @@ -4253,14 +4253,14 @@ is able to handle phenomena related to scope by means of an higher-order type th Machine Comprehension by Text-to-Text Neural Question Generation - Xingdi Yuan - Tong Wang - Caglar Gulcehre - Alessandro Sordoni - Philip Bachman - Saizheng Zhang - Sandeep Subramanian - Adam Trischler + XingdiYuan + TongWang + CaglarGulcehre + AlessandroSordoni + PhilipBachman + SaizhengZhang + SandeepSubramanian + AdamTrischler 15–25 W17-2603 10.18653/v1/W17-2603 @@ -4268,10 +4268,10 @@ is able to handle phenomena related to scope by means of an higher-order type th Emergent Predication Structure in Hidden State Vectors of Neural Readers - Hai Wang - Takeshi Onishi - Kevin Gimpel - David McAllester + HaiWang + TakeshiOnishi + KevinGimpel + DavidMcAllester 26–36 W17-2604 10.18653/v1/W17-2604 @@ -4279,8 +4279,8 @@ is able to handle phenomena related to scope by means of an higher-order type th Towards Harnessing Memory Networks for Coreference Resolution - Joe Cheri - Pushpak Bhattacharyya + JoeCheri + PushpakBhattacharyya 37–42 W17-2605 10.18653/v1/W17-2605 @@ -4288,9 +4288,9 @@ is able to handle phenomena related to scope by means of an higher-order type th Combining Word-Level and Character-Level Representations for Relation Classification of Informal Text - Dongyun Liang - Weiran Xu - Yinge Zhao + DongyunLiang + WeiranXu + YingeZhao 43–47 W17-2606 10.18653/v1/W17-2606 @@ -4298,10 +4298,10 @@ is able to handle phenomena related to scope by means of an higher-order type th Transfer Learning for Neural Semantic Parsing - Xing Fan - Emilio Monti - Lambert Mathias - Markus Dreyer + XingFan + EmilioMonti + LambertMathias + MarkusDreyer 48–56 W17-2607 10.18653/v1/W17-2607 @@ -4309,10 +4309,10 @@ is able to handle phenomena related to scope by means of an higher-order type th Modeling Large-Scale Structured Relationships with Shared Memory for Knowledge Base Completion - Yelong Shen - Po-Sen Huang - Ming-Wei Chang - Jianfeng Gao + YelongShen + Po-SenHuang + Ming-WeiChang + JianfengGao 57–68 W17-2608 10.18653/v1/W17-2608 @@ -4320,9 +4320,9 @@ is able to handle phenomena related to scope by means of an higher-order type th Knowledge Base Completion: Baselines Strike Back - Rudolf Kadlec - Ondrej Bajgar - Jan Kleindienst + RudolfKadlec + OndrejBajgar + JanKleindienst 69–74 W17-2609 10.18653/v1/W17-2609 @@ -4330,9 +4330,9 @@ is able to handle phenomena related to scope by means of an higher-order type th Sequential Attention: A Context-Aware Alignment Function for Machine Reading - Sebastian Brarda - Philip Yeres - Samuel Bowman + SebastianBrarda + PhilipYeres + SamuelBowman 75–80 W17-2610 10.18653/v1/W17-2610 @@ -4340,12 +4340,12 @@ is able to handle phenomena related to scope by means of an higher-order type th Semantic Vector Encoding and Similarity Search Using Fulltext Search Engines - Jan Rygl - Jan Pomikálek - Radim Řehůřek - Michal Růžička - Vít Novotný - Petr Sojka + JanRygl + JanPomikálek + RadimŘehůřek + MichalRůžička + VítNovotný + PetrSojka 81–90 W17-2611 10.18653/v1/W17-2611 @@ -4353,8 +4353,8 @@ is able to handle phenomena related to scope by means of an higher-order type th Multi-task Domain Adaptation for Sequence Tagging - Nanyun Peng - Mark Dredze + NanyunPeng + MarkDredze 91–100 W17-2612 10.18653/v1/W17-2612 @@ -4362,11 +4362,11 @@ is able to handle phenomena related to scope by means of an higher-order type th Beyond Bilingual: Multi-sense Word Embeddings using Multilingual Context - Shyam Upadhyay - Kai-Wei Chang - Matt Taddy - Adam Kalai - James Zou + ShyamUpadhyay + Kai-WeiChang + MattTaddy + AdamKalai + JamesZou 101–110 W17-2613 10.18653/v1/W17-2613 @@ -4374,10 +4374,10 @@ is able to handle phenomena related to scope by means of an higher-order type th <fixed-case>D</fixed-case>oc<fixed-case>T</fixed-case>ag2<fixed-case>V</fixed-case>ec: An Embedding Based Multi-label Learning Approach for Document Tagging - Sheng Chen - Akshay Soni - Aasish Pappu - Yashar Mehdad + ShengChen + AkshaySoni + AasishPappu + YasharMehdad 111–120 W17-2614 10.18653/v1/W17-2614 @@ -4385,8 +4385,8 @@ is able to handle phenomena related to scope by means of an higher-order type th Binary Paragraph Vectors - Karol Grzegorczyk - Marcin Kurdziel + KarolGrzegorczyk + MarcinKurdziel 121–130 W17-2615 10.18653/v1/W17-2615 @@ -4394,9 +4394,9 @@ is able to handle phenomena related to scope by means of an higher-order type th Representing Compositionality based on Multiple Timescales Gated Recurrent Neural Networks with Adaptive Temporal Hierarchy for Character-Level Language Models - Dennis Singh Moirangthem - Jegyung Son - Minho Lee + Dennis SinghMoirangthem + JegyungSon + MinhoLee 131–138 W17-2616 10.18653/v1/W17-2616 @@ -4404,8 +4404,8 @@ is able to handle phenomena related to scope by means of an higher-order type th Learning Bilingual Projections of Embeddings for Vocabulary Expansion in Machine Translation - Pranava Swaroop Madhyastha - Cristina España-Bonet + Pranava SwaroopMadhyastha + CristinaEspaña-Bonet 139–145 W17-2617 10.18653/v1/W17-2617 @@ -4413,9 +4413,9 @@ is able to handle phenomena related to scope by means of an higher-order type th Prediction of Frame-to-Frame Relations in the <fixed-case>F</fixed-case>rame<fixed-case>N</fixed-case>et Hierarchy with Frame Embeddings - Teresa Botschen - Hatem Mousselly-Sergieh - Iryna Gurevych + TeresaBotschen + HatemMousselly-Sergieh + IrynaGurevych 146–156 W17-2618 10.18653/v1/W17-2618 @@ -4423,8 +4423,8 @@ is able to handle phenomena related to scope by means of an higher-order type th Learning Joint Multilingual Sentence Representations with Neural Machine Translation - Holger Schwenk - Matthijs Douze + HolgerSchwenk + MatthijsDouze 157–167 W17-2619 10.18653/v1/W17-2619 @@ -4432,12 +4432,12 @@ is able to handle phenomena related to scope by means of an higher-order type th Transfer Learning for Speech Recognition on a Budget - Julius Kunze - Louis Kirsch - Ilia Kurenkov - Andreas Krug - Jens Johannsmeier - Sebastian Stober + JuliusKunze + LouisKirsch + IliaKurenkov + AndreasKrug + JensJohannsmeier + SebastianStober 168–177 W17-2620 10.18653/v1/W17-2620 @@ -4445,8 +4445,8 @@ is able to handle phenomena related to scope by means of an higher-order type th Gradual Learning of Matrix-Space Models of Language for Sentiment Analysis - Shima Asaadi - Sebastian Rudolph + ShimaAsaadi + SebastianRudolph 178–185 W17-2621 10.18653/v1/W17-2621 @@ -4454,9 +4454,9 @@ is able to handle phenomena related to scope by means of an higher-order type th Improving Language Modeling using Densely Connected Recurrent Neural Networks - Fréderic Godin - Joni Dambre - Wesley De Neve + FrédericGodin + JoniDambre + WesleyDe Neve 186–190 W17-2622 10.18653/v1/W17-2622 @@ -4464,13 +4464,13 @@ is able to handle phenomena related to scope by means of an higher-order type th <fixed-case>N</fixed-case>ews<fixed-case>QA</fixed-case>: A Machine Comprehension Dataset - Adam Trischler - Tong Wang - Xingdi Yuan - Justin Harris - Alessandro Sordoni - Philip Bachman - Kaheer Suleman + AdamTrischler + TongWang + XingdiYuan + JustinHarris + AlessandroSordoni + PhilipBachman + KaheerSuleman 191–200 W17-2623 10.18653/v1/W17-2623 @@ -4478,11 +4478,11 @@ is able to handle phenomena related to scope by means of an higher-order type th Intrinsic and Extrinsic Evaluation of Spatiotemporal Text Representations in Twitter Streams - Lawrence Phillips - Kyle Shaffer - Dustin Arendt - Nathan Hodas - Svitlana Volkova + LawrencePhillips + KyleShaffer + DustinArendt + NathanHodas + SvitlanaVolkova 201–210 W17-2624 10.18653/v1/W17-2624 @@ -4490,11 +4490,11 @@ is able to handle phenomena related to scope by means of an higher-order type th Rethinking Skip-thought: A Neighborhood based Approach - Shuai Tang - Hailin Jin - Chen Fang - Zhaowen Wang - Virginia de Sa + ShuaiTang + HailinJin + ChenFang + ZhaowenWang + Virginiade Sa 211–218 W17-2625 10.18653/v1/W17-2625 @@ -4503,10 +4503,10 @@ is able to handle phenomena related to scope by means of an higher-order type th A Frame Tracking Model for Memory-Enhanced Dialogue Systems - Hannes Schulz - Jeremie Zumer - Layla El Asri - Shikhar Sharma + HannesSchulz + JeremieZumer + LaylaEl Asri + ShikharSharma 219–227 W17-2626 10.18653/v1/W17-2626 @@ -4514,10 +4514,10 @@ is able to handle phenomena related to scope by means of an higher-order type th Plan, Attend, Generate: Character-Level Neural Machine Translation with Planning - Caglar Gulcehre - Francis Dutil - Adam Trischler - Yoshua Bengio + CaglarGulcehre + FrancisDutil + AdamTrischler + YoshuaBengio 228–234 W17-2627 10.18653/v1/W17-2627 @@ -4525,9 +4525,9 @@ is able to handle phenomena related to scope by means of an higher-order type th Does the Geometry of Word Embeddings Help Document Classification? A Case Study on Persistent Homology-Based Representations - Paul Michel - Abhilasha Ravichander - Shruti Rijhwani + PaulMichel + AbhilashaRavichander + ShrutiRijhwani 235–240 W17-2628 10.18653/v1/W17-2628 @@ -4535,11 +4535,11 @@ is able to handle phenomena related to scope by means of an higher-order type th Adversarial Generation of Natural Language - Sandeep Subramanian - Sai Rajeswar - Francis Dutil - Chris Pal - Aaron Courville + SandeepSubramanian + SaiRajeswar + FrancisDutil + ChrisPal + AaronCourville 241–251 W17-2629 10.18653/v1/W17-2629 @@ -4547,11 +4547,11 @@ is able to handle phenomena related to scope by means of an higher-order type th Deep Active Learning for Named Entity Recognition - Yanyao Shen - Hyokun Yun - Zachary Lipton - Yakov Kronrod - Animashree Anandkumar + YanyaoShen + HyokunYun + ZacharyLipton + YakovKronrod + AnimashreeAnandkumar 252–256 W17-2630 10.18653/v1/W17-2630 @@ -4559,8 +4559,8 @@ is able to handle phenomena related to scope by means of an higher-order type th Learning when to skim and when to read - Alexander Johansen - Richard Socher + AlexanderJohansen + RichardSocher 257–264 W17-2631 10.18653/v1/W17-2631 @@ -4568,9 +4568,9 @@ is able to handle phenomena related to scope by means of an higher-order type th Learning to Embed Words in Context for Syntactic Tasks - Lifu Tu - Kevin Gimpel - Karen Livescu + LifuTu + KevinGimpel + KarenLivescu 265–275 W17-2632 10.18653/v1/W17-2632 @@ -4600,8 +4600,8 @@ is able to handle phenomena related to scope by means of an higher-order type th news<fixed-case>L</fixed-case>ens: building and visualizing long-ranging news stories - Philippe Laban - Marti Hearst + PhilippeLaban + MartiHearst 1–9 W17-2701 10.18653/v1/W17-2701 @@ -4609,8 +4609,8 @@ is able to handle phenomena related to scope by means of an higher-order type th Detecting Changes in Twitter Streams using Temporal Clusters of Hashtags - Yunli Wang - Cyril Goutte + YunliWang + CyrilGoutte 10–14 W17-2702 10.18653/v1/W17-2702 @@ -4618,9 +4618,9 @@ is able to handle phenomena related to scope by means of an higher-order type th Event Detection Using Frame-Semantic Parser - Evangelia Spiliopoulou - Eduard Hovy - Teruko Mitamura + EvangeliaSpiliopoulou + EduardHovy + TerukoMitamura 15–20 W17-2703 10.18653/v1/W17-2703 @@ -4628,8 +4628,8 @@ is able to handle phenomena related to scope by means of an higher-order type th Improving Shared Argument Identification in <fixed-case>J</fixed-case>apanese Event Knowledge Acquisition - Yin Jou Huang - Sadao Kurohashi + Yin JouHuang + SadaoKurohashi 21–30 W17-2704 10.18653/v1/W17-2704 @@ -4637,9 +4637,9 @@ is able to handle phenomena related to scope by means of an higher-order type th Tracing armed conflicts with diachronic word embedding models - Andrey Kutuzov - Erik Velldal - Lilja Øvrelid + AndreyKutuzov + ErikVelldal + LiljaØvrelid 31–36 W17-2705 10.18653/v1/W17-2705 @@ -4648,9 +4648,9 @@ is able to handle phenomena related to scope by means of an higher-order type th The Circumstantial Event Ontology (<fixed-case>CEO</fixed-case>) - Roxane Segers - Tommaso Caselli - Piek Vossen + RoxaneSegers + TommasoCaselli + PiekVossen 37–41 W17-2706 10.18653/v1/W17-2706 @@ -4658,15 +4658,15 @@ is able to handle phenomena related to scope by means of an higher-order type th Event Detection and Semantic Storytelling: Generating a Travelogue from a large Collection of Personal Letters - Georg Rehm - Julian Moreno Schneider - Peter Bourgonje - Ankit Srivastava - Jan Nehring - Armin Berger - Luca König - Sören Räuchle - Jens Gerth + GeorgRehm + JulianMoreno Schneider + PeterBourgonje + AnkitSrivastava + JanNehring + ArminBerger + LucaKönig + SörenRäuchle + JensGerth 42–51 W17-2707 10.18653/v1/W17-2707 @@ -4674,9 +4674,9 @@ is able to handle phenomena related to scope by means of an higher-order type th Inference of Fine-Grained Event Causality from Blogs and Films - Zhichao Hu - Elahe Rahimtoroghi - Marilyn Walker + ZhichaoHu + ElaheRahimtoroghi + MarilynWalker 52–58 W17-2708 10.18653/v1/W17-2708 @@ -4684,10 +4684,10 @@ is able to handle phenomena related to scope by means of an higher-order type th On the Creation of a Security-Related Event Corpus - Martin Atkinson - Jakub Piskorski - Hristo Tanev - Vanni Zavarella + MartinAtkinson + JakubPiskorski + HristoTanev + VanniZavarella 59–65 W17-2709 10.18653/v1/W17-2709 @@ -4695,7 +4695,7 @@ is able to handle phenomena related to scope by means of an higher-order type th Inducing Event Types and Roles in Reverse: Using Function to Discover Theme - Natalie Ahn + NatalieAhn 66–76 W17-2710 10.18653/v1/W17-2710 @@ -4703,8 +4703,8 @@ is able to handle phenomena related to scope by means of an higher-order type th The Event <fixed-case>S</fixed-case>tory<fixed-case>L</fixed-case>ine Corpus: A New Benchmark for Causal and Temporal Relation Extraction - Tommaso Caselli - Piek Vossen + TommasoCaselli + PiekVossen 77–86 W17-2711 10.18653/v1/W17-2711 @@ -4712,10 +4712,10 @@ is able to handle phenomena related to scope by means of an higher-order type th The Rich Event Ontology - Susan Brown - Claire Bonial - Leo Obrst - Martha Palmer + SusanBrown + ClaireBonial + LeoObrst + MarthaPalmer 87–97 W17-2712 10.18653/v1/W17-2712 @@ -4723,9 +4723,9 @@ is able to handle phenomena related to scope by means of an higher-order type th Integrating Decompositional Event Structures into Storylines - William Croft - Pavlína Pešková - Michael Regan + WilliamCroft + PavlínaPešková + MichaelRegan 98–109 W17-2713 10.18653/v1/W17-2713 @@ -4752,10 +4752,10 @@ is able to handle phenomena related to scope by means of an higher-order type th Grounding Language for Interactive Task Learning - Peter Lindes - Aaron Mininger - James R. Kirk - John E. Laird + PeterLindes + AaronMininger + James R.Kirk + John E.Laird 1–9 W17-2801 10.18653/v1/W17-2801 @@ -4765,9 +4765,9 @@ is able to handle phenomena related to scope by means of an higher-order type th Learning how to Learn: An Adaptive Dialogue Agent for Incrementally Learning Visually Grounded Word Meanings - Yanchao Yu - Arash Eshghi - Oliver Lemon + YanchaoYu + ArashEshghi + OliverLemon 10–19 W17-2802 10.18653/v1/W17-2802 @@ -4775,9 +4775,9 @@ is able to handle phenomena related to scope by means of an higher-order type th Guiding Interaction Behaviors for Multi-modal Grounded Language Learning - Jesse Thomason - Jivko Sinapov - Raymond Mooney + JesseThomason + JivkoSinapov + RaymondMooney 20–24 W17-2803 10.18653/v1/W17-2803 @@ -4785,10 +4785,10 @@ is able to handle phenomena related to scope by means of an higher-order type th Structured Learning for Context-aware Spoken Language Understanding of Robotic Commands - Andrea Vanzo - Danilo Croce - Roberto Basili - Daniele Nardi + AndreaVanzo + DaniloCroce + RobertoBasili + DanieleNardi 25–34 W17-2804 10.18653/v1/W17-2804 @@ -4796,11 +4796,11 @@ is able to handle phenomena related to scope by means of an higher-order type th Natural Language Grounding and Grammar Induction for Robotic Manipulation Commands - Muhannad Alomari - Paul Duckworth - Majd Hawasly - David C. Hogg - Anthony G. Cohn + MuhannadAlomari + PaulDuckworth + MajdHawasly + David C.Hogg + Anthony G.Cohn 35–43 W17-2805 10.18653/v1/W17-2805 @@ -4808,8 +4808,8 @@ is able to handle phenomena related to scope by means of an higher-order type th Communication with Robots using Multilayer Recurrent Networks - Bedřich Pišl - David Mareček + BedřichPišl + DavidMareček 44–48 W17-2806 10.18653/v1/W17-2806 @@ -4817,10 +4817,10 @@ is able to handle phenomena related to scope by means of an higher-order type th Grounding Symbols in Multi-Modal Instructions - Yordan Hristov - Svetlin Penkov - Alex Lascarides - Subramanian Ramamoorthy + YordanHristov + SvetlinPenkov + AlexLascarides + SubramanianRamamoorthy 49–57 W17-2807 10.18653/v1/W17-2807 @@ -4828,15 +4828,15 @@ is able to handle phenomena related to scope by means of an higher-order type th Exploring Variation of Natural Human Commands to a Robot in a Collaborative Navigation Task - Matthew Marge - Claire Bonial - Ashley Foots - Cory Hayes - Cassidy Henry - Kimberly Pollard - Ron Artstein - Clare Voss - David Traum + MatthewMarge + ClaireBonial + AshleyFoots + CoryHayes + CassidyHenry + KimberlyPollard + RonArtstein + ClareVoss + DavidTraum 58–66 W17-2808 10.18653/v1/W17-2808 @@ -4845,13 +4845,13 @@ is able to handle phenomena related to scope by means of an higher-order type th A Tale of Two <fixed-case>DRAGGN</fixed-case>s: A Hybrid Approach for Interpreting Action-Oriented and Goal-Oriented Instructions - Siddharth Karamcheti - Edward Clem Williams - Dilip Arumugam - Mina Rhee - Nakul Gopalan - Lawson L.S. Wong - Stefanie Tellex + SiddharthKaramcheti + Edward ClemWilliams + DilipArumugam + MinaRhee + NakulGopalan + Lawson L.S.Wong + StefanieTellex 67–75 W17-2809 10.18653/v1/W17-2809 @@ -4859,8 +4859,8 @@ is able to handle phenomena related to scope by means of an higher-order type th Are Distributional Representations Ready for the Real World? Evaluating Word Vectors for Grounded Perceptual Meaning - Li Lucy - Jon Gauthier + LiLucy + JonGauthier 76–85 W17-2810 10.18653/v1/W17-2810 @@ -4868,10 +4868,10 @@ is able to handle phenomena related to scope by means of an higher-order type th Sympathy Begins with a Smile, Intelligence Begins with a Word: Use of Multimodal Features in Spoken Human-Robot Interaction - Jekaterina Novikova - Christian Dondrup - Ioannis Papaioannou - Oliver Lemon + JekaterinaNovikova + ChristianDondrup + IoannisPapaioannou + OliverLemon 86–94 W17-2811 10.18653/v1/W17-2811 @@ -4879,16 +4879,16 @@ is able to handle phenomena related to scope by means of an higher-order type th Towards Problem Solving Agents that Communicate and Learn - Anjali Narayan-Chen - Colin Graber - Mayukh Das - Md Rakibul Islam - Soham Dan - Sriraam Natarajan - Janardhan Rao Doppa - Julia Hockenmaier - Martha Palmer - Dan Roth + AnjaliNarayan-Chen + ColinGraber + MayukhDas + Md RakibulIslam + SohamDan + SriraamNatarajan + Janardhan RaoDoppa + JuliaHockenmaier + MarthaPalmer + DanRoth 95–103 W17-2812 10.18653/v1/W17-2812 @@ -4917,9 +4917,9 @@ is able to handle phenomena related to scope by means of an higher-order type th Language-independent Gender Prediction on Twitter - Nikola Ljubešić - Darja Fišer - Tomaž Erjavec + NikolaLjubešić + DarjaFišer + TomažErjavec 1–6 W17-2901 10.18653/v1/W17-2901 @@ -4927,8 +4927,8 @@ is able to handle phenomena related to scope by means of an higher-order type th When does a compliment become sexist? Analysis and classification of ambivalent sexism using twitter data - Akshita Jha - Radhika Mamidi + AkshitaJha + RadhikaMamidi 7–16 W17-2902 10.18653/v1/W17-2902 @@ -4936,9 +4936,9 @@ is able to handle phenomena related to scope by means of an higher-order type th Personality Driven Differences in Paraphrase Preference - Daniel Preoţiuc-Pietro - Jordan Carpenter - Lyle Ungar + DanielPreoţiuc-Pietro + JordanCarpenter + LyleUngar 17–26 W17-2903 10.18653/v1/W17-2903 @@ -4947,7 +4947,7 @@ is able to handle phenomena related to scope by means of an higher-order type th community2vec: Vector representations of online communities encode semantic relationships - Trevor Martin + TrevorMartin 27–31 W17-2904 10.18653/v1/W17-2904 @@ -4955,10 +4955,10 @@ is able to handle phenomena related to scope by means of an higher-order type th Telling Apart Tweets Associated with Controversial versus Non-Controversial Topics - Aseel Addawood - Rezvaneh Rezapour - Omid Abdar - Jana Diesner + AseelAddawood + RezvanehRezapour + OmidAbdar + JanaDiesner 32–41 W17-2905 10.18653/v1/W17-2905 @@ -4966,9 +4966,9 @@ is able to handle phenomena related to scope by means of an higher-order type th Cross-Lingual Classification of Topics in Political Texts - Goran Glavaš - Federico Nanni - Simone Paolo Ponzetto + GoranGlavaš + FedericoNanni + Simone PaoloPonzetto 42–46 W17-2906 10.18653/v1/W17-2906 @@ -4976,8 +4976,8 @@ is able to handle phenomena related to scope by means of an higher-order type th Mining Social Science Publications for Survey Variables - Andrea Zielinski - Peter Mutschke + AndreaZielinski + PeterMutschke 47–52 W17-2907 10.18653/v1/W17-2907 @@ -4985,12 +4985,12 @@ is able to handle phenomena related to scope by means of an higher-order type th Linguistic Markers of Influence in Informal Interactions - Shrimai Prabhumoye - Samridhi Choudhary - Evangelia Spiliopoulou - Christopher Bogart - Carolyn Rose - Alan W Black + ShrimaiPrabhumoye + SamridhiChoudhary + EvangeliaSpiliopoulou + ChristopherBogart + CarolynRose + Alan WBlack 53–62 W17-2908 10.18653/v1/W17-2908 @@ -4998,10 +4998,10 @@ is able to handle phenomena related to scope by means of an higher-order type th Non-lexical Features Encode Political Affiliation on Twitter - Rachael Tatman - Leo Stewart - Amandalynne Paullada - Emma Spiro + RachaelTatman + LeoStewart + AmandalynnePaullada + EmmaSpiro 63–67 W17-2909 10.18653/v1/W17-2909 @@ -5009,7 +5009,7 @@ is able to handle phenomena related to scope by means of an higher-order type th Modelling Participation in Small Group Social Sequences with <fixed-case>M</fixed-case>arkov Rewards Analysis - Gabriel Murray + GabrielMurray 68–72 W17-2910 10.18653/v1/W17-2910 @@ -5017,10 +5017,10 @@ is able to handle phenomena related to scope by means of an higher-order type th Code-Switching as a Social Act: The Case of <fixed-case>A</fixed-case>rabic <fixed-case>W</fixed-case>ikipedia Talk Pages - Michael Yoder - Shruti Rijhwani - Carolyn Rosé - Lori Levin + MichaelYoder + ShrutiRijhwani + CarolynRosé + LoriLevin 73–82 W17-2911 10.18653/v1/W17-2911 @@ -5028,10 +5028,10 @@ is able to handle phenomena related to scope by means of an higher-order type th How Does Twitter User Behavior Vary Across Demographic Groups? - Zach Wood-Doughty - Michael Smith - David Broniatowski - Mark Dredze + ZachWood-Doughty + MichaelSmith + DavidBroniatowski + MarkDredze 83–89 W17-2912 10.18653/v1/W17-2912 @@ -5039,9 +5039,9 @@ is able to handle phenomena related to scope by means of an higher-order type th Ideological Phrase Indicators for Classification of Political Discourse Framing on Twitter - Kristen Johnson - I-Ta Lee - Dan Goldwasser + KristenJohnson + I-TaLee + DanGoldwasser 90–99 W17-2913 10.18653/v1/W17-2913 @@ -5067,8 +5067,8 @@ is able to handle phenomena related to scope by means of an higher-order type th Dimensions of Abusive Language on Twitter - Isobelle Clarke - Jack Grieve + IsobelleClarke + JackGrieve 1–10 W17-3001 10.18653/v1/W17-3001 @@ -5076,8 +5076,8 @@ is able to handle phenomena related to scope by means of an higher-order type th Constructive Language in News Comments - Varada Kolhatkar - Maite Taboada + VaradaKolhatkar + MaiteTaboada 11–17 W17-3002 10.18653/v1/W17-3002 @@ -5085,10 +5085,10 @@ is able to handle phenomena related to scope by means of an higher-order type th Rephrasing Profanity in <fixed-case>C</fixed-case>hinese Text - Hui-Po Su - Zhen-Jie Huang - Hao-Tsung Chang - Chuan-Jie Lin + Hui-PoSu + Zhen-JieHuang + Hao-TsungChang + Chuan-JieLin 18–24 W17-3003 10.18653/v1/W17-3003 @@ -5096,9 +5096,9 @@ is able to handle phenomena related to scope by means of an higher-order type th Deep Learning for User Comment Moderation - John Pavlopoulos - Prodromos Malakasiotis - Ion Androutsopoulos + JohnPavlopoulos + ProdromosMalakasiotis + IonAndroutsopoulos 25–35 W17-3004 10.18653/v1/W17-3004 @@ -5106,12 +5106,12 @@ is able to handle phenomena related to scope by means of an higher-order type th Class-based Prediction Errors to Detect Hate Speech with Out-of-vocabulary Words - Joan Serrà - Ilias Leontiadis - Dimitris Spathis - Gianluca Stringhini - Jeremy Blackburn - Athena Vakali + JoanSerrà + IliasLeontiadis + DimitrisSpathis + GianlucaStringhini + JeremyBlackburn + AthenaVakali 36–40 W17-3005 10.18653/v1/W17-3005 @@ -5119,8 +5119,8 @@ is able to handle phenomena related to scope by means of an higher-order type th One-step and Two-step Classification for Abusive Language Detection on Twitter - Ji Ho Park - Pascale Fung + Ji HoPark + PascaleFung 41–45 W17-3006 10.18653/v1/W17-3006 @@ -5128,9 +5128,9 @@ is able to handle phenomena related to scope by means of an higher-order type th Legal Framework, Dataset and Annotation Schema for Socially Unacceptable Online Discourse Practices in <fixed-case>S</fixed-case>lovene - Darja Fišer - Tomaž Erjavec - Nikola Ljubešić + DarjaFišer + TomažErjavec + NikolaLjubešić 46–51 W17-3007 10.18653/v1/W17-3007 @@ -5138,9 +5138,9 @@ is able to handle phenomena related to scope by means of an higher-order type th Abusive Language Detection on <fixed-case>A</fixed-case>rabic Social Media - Hamdy Mubarak - Kareem Darwish - Walid Magdy + HamdyMubarak + KareemDarwish + WalidMagdy 52–56 W17-3008 10.18653/v1/W17-3008 @@ -5148,11 +5148,11 @@ is able to handle phenomena related to scope by means of an higher-order type th Vectors for Counterspeech on Twitter - Lucas Wright - Derek Ruths - Kelly P Dillon - Haji Mohammad Saleem - Susan Benesch + LucasWright + DerekRuths + Kelly PDillon + Haji MohammadSaleem + SusanBenesch 57–62 W17-3009 10.18653/v1/W17-3009 @@ -5160,11 +5160,11 @@ is able to handle phenomena related to scope by means of an higher-order type th Detecting Nastiness in Social Media - Niloofar Safi Samghabadi - Suraj Maharjan - Alan Sprague - Raquel Diaz-Sprague - Thamar Solorio + NiloofarSafi Samghabadi + SurajMaharjan + AlanSprague + RaquelDiaz-Sprague + ThamarSolorio 63–72 W17-3010 10.18653/v1/W17-3010 @@ -5172,13 +5172,13 @@ is able to handle phenomena related to scope by means of an higher-order type th Technology Solutions to Combat Online Harassment - George Kennedy - Andrew McCollough - Edward Dixon - Alexei Bastidas - John Ryan - Chris Loo - Saurav Sahay + GeorgeKennedy + AndrewMcCollough + EdwardDixon + AlexeiBastidas + JohnRyan + ChrisLoo + SauravSahay 73–77 W17-3011 10.18653/v1/W17-3011 @@ -5186,10 +5186,10 @@ is able to handle phenomena related to scope by means of an higher-order type th Understanding Abuse: A Typology of Abusive Language Detection Subtasks - Zeerak Waseem - Thomas Davidson - Dana Warmsley - Ingmar Weber + ZeerakWaseem + ThomasDavidson + DanaWarmsley + IngmarWeber 78–84 W17-3012 10.18653/v1/W17-3012 @@ -5197,8 +5197,8 @@ is able to handle phenomena related to scope by means of an higher-order type th Using Convolutional Neural Networks to Classify Hate-Speech - Björn Gambäck - Utpal Kumar Sikdar + BjörnGambäck + Utpal KumarSikdar 85–90 W17-3013 10.18653/v1/W17-3013 @@ -5206,9 +5206,9 @@ is able to handle phenomena related to scope by means of an higher-order type th Illegal is not a Noun: Linguistic Form for Detection of Pejorative Nominalizations - Alexis Palmer - Melissa Robinson - Kristy K. Phillips + AlexisPalmer + MelissaRobinson + Kristy K.Phillips 91–100 W17-3014 10.18653/v1/W17-3014 @@ -5233,9 +5233,9 @@ is able to handle phenomena related to scope by means of an higher-order type th A Cross-modal Review of Indicators for Depression Detection Systems - Michelle Morales - Stefan Scherer - Rivka Levitan + MichelleMorales + StefanScherer + RivkaLevitan 1–12 W17-3101 10.18653/v1/W17-3101 @@ -5243,11 +5243,11 @@ is able to handle phenomena related to scope by means of an higher-order type th In your wildest dreams: the language and psychological features of dreams - Kate Niederhoffer - Jonathan Schler - Patrick Crutchley - Kate Loveys - Glen Coppersmith + KateNiederhoffer + JonathanSchler + PatrickCrutchley + KateLoveys + GlenCoppersmith 13–25 W17-3102 10.18653/v1/W17-3102 @@ -5255,11 +5255,11 @@ is able to handle phenomena related to scope by means of an higher-order type th A Corpus Analysis of Social Connections and Social Isolation in Adolescents Suffering from Depressive Disorders - Jia-Wen Guo - Danielle L Mowery - Djin Lai - Katherine Sward - Mike Conway + Jia-WenGuo + Danielle LMowery + DjinLai + KatherineSward + MikeConway 26–31 W17-3103 10.18653/v1/W17-3103 @@ -5267,10 +5267,10 @@ is able to handle phenomena related to scope by means of an higher-order type th Monitoring Tweets for Depression to Detect At-risk Users - Zunaira Jamil - Diana Inkpen - Prasadith Buddhitha - Kenton White + ZunairaJamil + DianaInkpen + PrasadithBuddhitha + KentonWhite 32–40 W17-3104 10.18653/v1/W17-3104 @@ -5278,9 +5278,9 @@ is able to handle phenomena related to scope by means of an higher-order type th Investigating Patient Attitudes Towards the use of Social Media Data to Augment Depression Diagnosis and Treatment: a Qualitative Study - Jude Mikal - Samantha Hurst - Mike Conway + JudeMikal + SamanthaHurst + MikeConway 41–47 W17-3105 10.18653/v1/W17-3105 @@ -5288,9 +5288,9 @@ is able to handle phenomena related to scope by means of an higher-order type th Natural-language Interactive Narratives in Imaginal Exposure Therapy for Obsessive-Compulsive Disorder - Melissa Roemmele - Paola Mardo - Andrew Gordon + MelissaRoemmele + PaolaMardo + AndrewGordon 48–57 W17-3106 10.18653/v1/W17-3106 @@ -5298,8 +5298,8 @@ is able to handle phenomena related to scope by means of an higher-order type th Detecting Anxiety through <fixed-case>R</fixed-case>eddit - Judy Hanwen Shen - Frank Rudzicz + Judy HanwenShen + FrankRudzicz 58–65 W17-3107 10.18653/v1/W17-3107 @@ -5307,9 +5307,9 @@ is able to handle phenomena related to scope by means of an higher-order type th Detecting and Explaining Crisis - Rohan Kshirsagar - Robert Morris - Samuel Bowman + RohanKshirsagar + RobertMorris + SamuelBowman 66–73 W17-3108 10.18653/v1/W17-3108 @@ -5317,8 +5317,8 @@ is able to handle phenomena related to scope by means of an higher-order type th A Dictionary-Based Comparison of Autobiographies by People and Murderous Monsters - Micah Iserman - Molly Ireland + MicahIserman + MollyIreland 74–84 W17-3109 10.18653/v1/W17-3109 @@ -5326,10 +5326,10 @@ is able to handle phenomena related to scope by means of an higher-order type th Small but Mighty: Affective Micropatterns for Quantifying Mental Health from Social Media Language - Kate Loveys - Patrick Crutchley - Emily Wyatt - Glen Coppersmith + KateLoveys + PatrickCrutchley + EmilyWyatt + GlenCoppersmith 85–95 W17-3110 10.18653/v1/W17-3110 @@ -5355,8 +5355,8 @@ is able to handle phenomena related to scope by means of an higher-order type th An Empirical Study of Adequate Vision Span for Attention-Based Neural Machine Translation - Raphael Shu - Hideki Nakayama + RaphaelShu + HidekiNakayama 1–10 W17-3201 10.18653/v1/W17-3201 @@ -5364,10 +5364,10 @@ is able to handle phenomena related to scope by means of an higher-order type th Analyzing Neural <fixed-case>MT</fixed-case> Search and Model Performance - Jan Niehues - Eunah Cho - Thanh-Le Ha - Alex Waibel + JanNiehues + EunahCho + Thanh-LeHa + AlexWaibel 11–17 W17-3202 10.18653/v1/W17-3202 @@ -5375,8 +5375,8 @@ is able to handle phenomena related to scope by means of an higher-order type th Stronger Baselines for Trustable Results in Neural Machine Translation - Michael Denkowski - Graham Neubig + MichaelDenkowski + GrahamNeubig 18–27 W17-3203 10.18653/v1/W17-3203 @@ -5384,8 +5384,8 @@ is able to handle phenomena related to scope by means of an higher-order type th Six Challenges for Neural Machine Translation - Philipp Koehn - Rebecca Knowles + PhilippKoehn + RebeccaKnowles 28–39 W17-3204 10.18653/v1/W17-3204 @@ -5393,10 +5393,10 @@ is able to handle phenomena related to scope by means of an higher-order type th Cost Weighting for Neural Machine Translation Domain Adaptation - Boxing Chen - Colin Cherry - George Foster - Samuel Larkin + BoxingChen + ColinCherry + GeorgeFoster + SamuelLarkin 40–46 W17-3205 10.18653/v1/W17-3205 @@ -5404,8 +5404,8 @@ is able to handle phenomena related to scope by means of an higher-order type th Detecting Untranslated Content for Neural Machine Translation - Isao Goto - Hideki Tanaka + IsaoGoto + HidekiTanaka 47–55 W17-3206 10.18653/v1/W17-3206 @@ -5413,8 +5413,8 @@ is able to handle phenomena related to scope by means of an higher-order type th Beam Search Strategies for Neural Machine Translation - Markus Freitag - Yaser Al-Onaizan + MarkusFreitag + YaserAl-Onaizan 56–60 W17-3207 10.18653/v1/W17-3207 @@ -5422,12 +5422,12 @@ is able to handle phenomena related to scope by means of an higher-order type th An Empirical Study of Mini-Batch Creation Strategies for Neural Machine Translation - Makoto Morishita - Yusuke Oda - Graham Neubig - Koichiro Yoshino - Katsuhito Sudoh - Satoshi Nakamura + MakotoMorishita + YusukeOda + GrahamNeubig + KoichiroYoshino + KatsuhitoSudoh + SatoshiNakamura 61–68 W17-3208 10.18653/v1/W17-3208 @@ -5435,9 +5435,9 @@ is able to handle phenomena related to scope by means of an higher-order type th Detecting Cross-Lingual Semantic Divergence for Neural Machine Translation - Marine Carpuat - Yogarshi Vyas - Xing Niu + MarineCarpuat + YogarshiVyas + XingNiu 69–79 W17-3209 10.18653/v1/W17-3209 @@ -5448,9 +5448,9 @@ is able to handle phenomena related to scope by means of an higher-order type th Proceedings of the 15th Meeting on the Mathematics of Language W17-34 - Makoto Kanazawa - Philippe de Groote - Mehrnoosh Sadrzadeh + MakotoKanazawa + Philippede Groote + MehrnooshSadrzadeh 10.18653/v1/W17-34 Association for Computational Linguistics
London, UK
@@ -5462,109 +5462,109 @@ is able to handle phenomena related to scope by means of an higher-order type th <fixed-case>BE</fixed-case> Is Not the Unique Homomorphism That Makes the Partee Triangle Commute - Junri Shimada + JunriShimada 1–10 W17-3401 10.18653/v1/W17-3401 How Many Stemmata with Root Degree k? - Armin Hoenen - Steffen Eger - Ralf Gehrke + ArminHoenen + SteffenEger + RalfGehrke 11–21 W17-3402 10.18653/v1/W17-3402 On the Logical Complexity of Autosegmental Representations - Adam Jardine + AdamJardine 22–35 W17-3403 10.18653/v1/W17-3403 Extracting Forbidden Factors from Regular Stringsets - James Rogers - Dakotah Lambert + JamesRogers + DakotahLambert 36–46 W17-3404 10.18653/v1/W17-3404 Latent-Variable <fixed-case>PCFG</fixed-case>s: Background and Applications - Shay Cohen + ShayCohen 47–58 W17-3405 10.18653/v1/W17-3405 A Proof-Theoretic Semantics for Transitive Verbs with an Implicit Object - Nissim Francez + NissimFrancez 59–67 W17-3406 10.18653/v1/W17-3406 Why We Speak - Rohit Parikh + RohitParikh 68–74 W17-3407 10.18653/v1/W17-3407 A Monotonicity Calculus and Its Completeness - Thomas Icard - Lawrence Moss - William Tune + ThomasIcard + LawrenceMoss + WilliamTune 75–87 W17-3408 10.18653/v1/W17-3408 <fixed-case>DAG</fixed-case> Automata for Meaning Representation - Frank Drewes + FrankDrewes 88–99 W17-3409 10.18653/v1/W17-3409 (Re)introducing Regular Graph Languages - Sorcha Gilroy - Adam Lopez - Sebastian Maneth - Pijus Simonaitis + SorchaGilroy + AdamLopez + SebastianManeth + PijusSimonaitis 100–113 W17-3410 10.18653/v1/W17-3410 Graph Transductions and Typological Gaps in Morphological Paradigms - Thomas Graf + ThomasGraf 114–126 W17-3411 10.18653/v1/W17-3411 Introducing Structure into Neural Network-Based Semantic Models - Stephen Clark + StephenClark 127 W17-3412 10.18653/v1/W17-3412 Count-Invariance Including Exponentials - Stepan Kuznetsov - Glyn Morrill - Oriol Valentín + StepanKuznetsov + GlynMorrill + OriolValentín 128–139 W17-3413 10.18653/v1/W17-3413 Conjunctive Categorial Grammars - Stepan Kuznetsov - Alexander Okhotin + StepanKuznetsov + AlexanderOkhotin 140–151 W17-3414 10.18653/v1/W17-3414 @@ -5588,10 +5588,10 @@ is able to handle phenomena related to scope by means of an higher-order type th Linguistic realisation as machine translation: Comparing different <fixed-case>MT</fixed-case> models for <fixed-case>AMR</fixed-case>-to-text generation - Thiago Castro Ferreira - Iacer Calixto - Sander Wubben - Emiel Krahmer + ThiagoCastro Ferreira + IacerCalixto + SanderWubben + EmielKrahmer 1–10 W17-3501 10.18653/v1/W17-3501 @@ -5599,7 +5599,7 @@ is able to handle phenomena related to scope by means of an higher-order type th A Survey on Intelligent Poetry Generation: Languages, Features, Techniques, Reutilisation and Evaluation - Hugo Gonçalo Oliveira + HugoGonçalo Oliveira 11–20 W17-3502 10.18653/v1/W17-3502 @@ -5607,9 +5607,9 @@ is able to handle phenomena related to scope by means of an higher-order type th Cross-linguistic differences and similarities in image descriptions - Emiel van Miltenburg - Desmond Elliott - Piek Vossen + Emielvan Miltenburg + DesmondElliott + PiekVossen 21–30 W17-3503 10.18653/v1/W17-3503 @@ -5617,10 +5617,10 @@ is able to handle phenomena related to scope by means of an higher-order type th Content Selection for Real-time Sports News Construction from Commentary Texts - Jin-ge Yao - Jianmin Zhang - Xiaojun Wan - Jianguo Xiao + Jin-geYao + JianminZhang + XiaojunWan + JianguoXiao 31–40 W17-3504 10.18653/v1/W17-3504 @@ -5628,9 +5628,9 @@ is able to handle phenomena related to scope by means of an higher-order type th Improving the Naturalness and Expressivity of Language Generation for <fixed-case>S</fixed-case>panish - Cristina Barros - Dimitra Gkatzia - Elena Lloret + CristinaBarros + DimitraGkatzia + ElenaLloret 41–50 W17-3505 10.18653/v1/W17-3505 @@ -5638,9 +5638,9 @@ is able to handle phenomena related to scope by means of an higher-order type th What is the Role of Recurrent Neural Networks (<fixed-case>RNN</fixed-case>s) in an Image Caption Generator? - Marc Tanti - Albert Gatt - Kenneth Camilleri + MarcTanti + AlbertGatt + KennethCamilleri 51–60 W17-3506 10.18653/v1/W17-3506 @@ -5648,11 +5648,11 @@ is able to handle phenomena related to scope by means of an higher-order type th Exploring the Behavior of Classic <fixed-case>REG</fixed-case> Algorithms in the Description of Characters in 3<fixed-case>D</fixed-case> Images - Gonzalo Méndez - Raquel Hervás - Susana Bautista - Adrián Rabadán - Teresa Rodríguez + GonzaloMéndez + RaquelHervás + SusanaBautista + AdriánRabadán + TeresaRodríguez 61–69 W17-3507 10.18653/v1/W17-3507 @@ -5660,9 +5660,9 @@ is able to handle phenomena related to scope by means of an higher-order type th Co-<fixed-case>P</fixed-case>oe<fixed-case>T</fixed-case>ry<fixed-case>M</fixed-case>e: a Co-Creative Interface for the Composition of Poetry - Hugo Gonçalo Oliveira - Tiago Mendes - Ana Boavida + HugoGonçalo Oliveira + TiagoMendes + AnaBoavida 70–71 W17-3508 10.18653/v1/W17-3508 @@ -5670,9 +5670,9 @@ is able to handle phenomena related to scope by means of an higher-order type th Refer-i<fixed-case>TTS</fixed-case>: A System for Referring in Spoken Installments to Objects in Real-World Images - Sina Zarrieß - M. Soledad López Gambino - David Schlangen + SinaZarrieß + M. SoledadLópez Gambino + DavidSchlangen 72–73 W17-3509 10.18653/v1/W17-3509 @@ -5680,7 +5680,7 @@ is able to handle phenomena related to scope by means of an higher-order type th Finding the “right” answers for customers - Frank Schilder + FrankSchilder 74 W17-3510 10.18653/v1/W17-3510 @@ -5688,8 +5688,8 @@ is able to handle phenomena related to scope by means of an higher-order type th Referring Expression Generation under Uncertainty: Algorithm and Evaluation Framework - Tom Williams - Matthias Scheutz + TomWilliams + MatthiasScheutz 75–84 W17-3511 10.18653/v1/W17-3511 @@ -5697,8 +5697,8 @@ is able to handle phenomena related to scope by means of an higher-order type th Natural Language Descriptions for Human Activities in Video Streams - Nouf Alharbi - Yoshihiko Gotoh + NoufAlharbi + YoshihikoGotoh 85–94 W17-3512 10.18653/v1/W17-3512 @@ -5706,9 +5706,9 @@ is able to handle phenomena related to scope by means of an higher-order type th <fixed-case>PASS</fixed-case>: A <fixed-case>D</fixed-case>utch data-to-text system for soccer, targeted towards specific audiences - Chris van der Lee - Emiel Krahmer - Sander Wubben + Chrisvan der Lee + EmielKrahmer + SanderWubben 95–104 W17-3513 10.18653/v1/W17-3513 @@ -5716,9 +5716,9 @@ is able to handle phenomena related to scope by means of an higher-order type th Evaluation of a <fixed-case>R</fixed-case>unyankore grammar engine for healthcare messages - Joan Byamugisha - C. Maria Keet - Brian DeRenzi + JoanByamugisha + C. MariaKeet + BrianDeRenzi 105–113 W17-3514 10.18653/v1/W17-3514 @@ -5726,7 +5726,7 @@ is able to handle phenomena related to scope by means of an higher-order type th Talking about the world with a distributed model - Gemma Boleda + GemmaBoleda 114 W17-3515 10.18653/v1/W17-3515 @@ -5734,9 +5734,9 @@ is able to handle phenomena related to scope by means of an higher-order type th The <fixed-case>C</fixed-case>ode2<fixed-case>T</fixed-case>ext Challenge: Text Generation in Source Libraries - Kyle Richardson - Sina Zarrieß - Jonas Kuhn + KyleRichardson + SinaZarrieß + JonasKuhn 115–119 W17-3516 10.18653/v1/W17-3516 @@ -5744,10 +5744,10 @@ is able to handle phenomena related to scope by means of an higher-order type th Shared Task Proposal: Multilingual Surface Realization Using Universal Dependency Trees - Simon Mille - Bernd Bohnet - Leo Wanner - Anja Belz + SimonMille + BerndBohnet + LeoWanner + AnjaBelz 120–123 W17-3517 10.18653/v1/W17-3517 @@ -5755,10 +5755,10 @@ is able to handle phenomena related to scope by means of an higher-order type th The <fixed-case>W</fixed-case>eb<fixed-case>NLG</fixed-case> Challenge: Generating Text from <fixed-case>RDF</fixed-case> Data - Claire Gardent - Anastasia Shimorina - Shashi Narayan - Laura Perez-Beltrachini + ClaireGardent + AnastasiaShimorina + ShashiNarayan + LauraPerez-Beltrachini 124–133 W17-3518 10.18653/v1/W17-3518 @@ -5766,7 +5766,7 @@ is able to handle phenomena related to scope by means of an higher-order type th A Commercial Perspective on Reference - Ehud Reiter + EhudReiter 134–138 W17-3519 10.18653/v1/W17-3519 @@ -5774,8 +5774,8 @@ is able to handle phenomena related to scope by means of an higher-order type th Integrated sentence generation using charts - Alexander Koller - Nikos Engonopoulos + AlexanderKoller + NikosEngonopoulos 139–143 W17-3520 10.18653/v1/W17-3520 @@ -5783,9 +5783,9 @@ is able to handle phenomena related to scope by means of an higher-order type th Adapting <fixed-case>S</fixed-case>imple<fixed-case>NLG</fixed-case> to <fixed-case>S</fixed-case>panish - Alejandro Ramos-Soto - Julio Janeiro-Gallardo - Alberto Bugarín Diz + AlejandroRamos-Soto + JulioJaneiro-Gallardo + AlbertoBugarín Diz 144–148 W17-3521 10.18653/v1/W17-3521 @@ -5793,9 +5793,9 @@ is able to handle phenomena related to scope by means of an higher-order type th G-<fixed-case>TUNA</fixed-case>: a corpus of referring expressions in <fixed-case>G</fixed-case>erman, including duration information - David Howcroft - Jorrig Vogels - Vera Demberg + DavidHowcroft + JorrigVogels + VeraDemberg 149–153 W17-3522 10.18653/v1/W17-3522 @@ -5803,9 +5803,9 @@ is able to handle phenomena related to scope by means of an higher-order type th Toward an <fixed-case>NLG</fixed-case> System for <fixed-case>B</fixed-case>antu languages: first steps with <fixed-case>R</fixed-case>unyankore (demo) - Joan Byamugisha - C. Maria Keet - Brian DeRenzi + JoanByamugisha + C. MariaKeet + BrianDeRenzi 154–155 W17-3523 10.18653/v1/W17-3523 @@ -5813,8 +5813,8 @@ is able to handle phenomena related to scope by means of an higher-order type th A working, non-trivial, topically indifferent <fixed-case>NLG</fixed-case> System for 17 languages - Robert Weißgraeber - Andreas Madsack + RobertWeißgraeber + AndreasMadsack 156–157 W17-3524 10.18653/v1/W17-3524 @@ -5822,9 +5822,9 @@ is able to handle phenomena related to scope by means of an higher-order type th Generating titles for millions of browse pages on an e-Commerce site - Prashant Mathur - Nicola Ueffing - Gregor Leusch + PrashantMathur + NicolaUeffing + GregorLeusch 158–167 W17-3525 10.18653/v1/W17-3525 @@ -5832,8 +5832,8 @@ is able to handle phenomena related to scope by means of an higher-order type th Towards Automatic Generation of Product Reviews from Aspect-Sentiment Scores - Hongyu Zang - Xiaojun Wan + HongyuZang + XiaojunWan 168–177 W17-3526 10.18653/v1/W17-3526 @@ -5841,8 +5841,8 @@ is able to handle phenomena related to scope by means of an higher-order type th A model of suspense for narrative generation - Richard Doust - Paul Piwek + RichardDoust + PaulPiwek 178–187 W17-3527 10.18653/v1/W17-3527 @@ -5850,10 +5850,10 @@ is able to handle phenomena related to scope by means of an higher-order type th Data-Driven News Generation for Automated Journalism - Leo Leppänen - Myriam Munezero - Mark Granroth-Wilding - Hannu Toivonen + LeoLeppänen + MyriamMunezero + MarkGranroth-Wilding + HannuToivonen 188–197 W17-3528 10.18653/v1/W17-3528 @@ -5861,9 +5861,9 @@ is able to handle phenomena related to scope by means of an higher-order type th Data Augmentation for Visual Question Answering - Kushal Kafle - Mohammed Yousefhussien - Christopher Kanan + KushalKafle + MohammedYousefhussien + ChristopherKanan 198–202 W17-3529 10.18653/v1/W17-3529 @@ -5871,7 +5871,7 @@ is able to handle phenomena related to scope by means of an higher-order type th Personalized Questions, Answers and Grammars: Aiding the Search for Relevant Web Information - Marta Gatius + MartaGatius 203–207 W17-3530 10.18653/v1/W17-3530 @@ -5879,11 +5879,11 @@ is able to handle phenomena related to scope by means of an higher-order type th A Comparison of Neural Models for Word Ordering - Eva Hasler - Felix Stahlberg - Marcus Tomalin - Adrià de Gispert - Bill Byrne + EvaHasler + FelixStahlberg + MarcusTomalin + Adriàde Gispert + BillByrne 208–212 W17-3531 10.18653/v1/W17-3531 @@ -5891,12 +5891,12 @@ is able to handle phenomena related to scope by means of an higher-order type th Investigating the content and form of referring expressions in <fixed-case>M</fixed-case>andarin: introducing the Mtuna corpus - Kees van Deemter - Le Sun - Rint Sybesma - Xiao Li - Bo Chen - Muyun Yang + Keesvan Deemter + LeSun + RintSybesma + XiaoLi + BoChen + MuyunYang 213–217 W17-3532 10.18653/v1/W17-3532 @@ -5904,8 +5904,8 @@ is able to handle phenomena related to scope by means of an higher-order type th Realization of long sentences using chunking - Ewa Muszyńska - Ann Copestake + EwaMuszyńska + AnnCopestake 218–222 W17-3533 10.18653/v1/W17-3533 @@ -5913,10 +5913,10 @@ is able to handle phenomena related to scope by means of an higher-order type th <fixed-case>S</fixed-case>a<fixed-case>T</fixed-case>o<fixed-case>S</fixed-case>: Assessing and Summarising Terms of Services from <fixed-case>G</fixed-case>erman Webshops - Daniel Braun - Elena Scepankova - Patrick Holl - Florian Matthes + DanielBraun + ElenaScepankova + PatrickHoll + FlorianMatthes 223–227 W17-3534 10.18653/v1/W17-3534 @@ -5924,9 +5924,9 @@ is able to handle phenomena related to scope by means of an higher-order type th Textually Summarising Incomplete Data - Stephanie Inglis - Ehud Reiter - Somayajulu Sripada + StephanieInglis + EhudReiter + SomayajuluSripada 228–232 W17-3535 10.18653/v1/W17-3535 @@ -5934,8 +5934,8 @@ is able to handle phenomena related to scope by means of an higher-order type th Improving the generation of personalised descriptions - Thiago Castro Ferreira - Ivandré Paraboni + ThiagoCastro Ferreira + IvandréParaboni 233–237 W17-3536 10.18653/v1/W17-3536 @@ -5943,8 +5943,8 @@ is able to handle phenomena related to scope by means of an higher-order type th Analysing Data-To-Text Generation Benchmarks - Laura Perez-Beltrachini - Claire Gardent + LauraPerez-Beltrachini + ClaireGardent 238–242 W17-3537 10.18653/v1/W17-3537 @@ -5952,9 +5952,9 @@ is able to handle phenomena related to scope by means of an higher-order type th Linguistic Description of Complex Phenomena with the r<fixed-case>LDCP</fixed-case> R Package - Jose Alonso - Patricia Conde-Clemente - Gracian Trivino + JoseAlonso + PatriciaConde-Clemente + GracianTrivino 243–244 W17-3538 10.18653/v1/W17-3538 @@ -5962,8 +5962,8 @@ is able to handle phenomena related to scope by means of an higher-order type th A demo of <fixed-case>FORG</fixed-case>e: the <fixed-case>P</fixed-case>ompeu <fixed-case>F</fixed-case>abra Open Rule-based Generator - Simon Mille - Leo Wanner + SimonMille + LeoWanner 245–246 W17-3539 10.18653/v1/W17-3539 @@ -5971,9 +5971,9 @@ is able to handle phenomena related to scope by means of an higher-order type th Referential Success of Set Referring Expressions with Fuzzy Properties - Nicolás Marín - Gustavo Rivas-Gervilla - Daniel Sánchez + NicolásMarín + GustavoRivas-Gervilla + DanielSánchez 247–251 W17-3540 10.18653/v1/W17-3540 @@ -5981,10 +5981,10 @@ is able to handle phenomena related to scope by means of an higher-order type th Neural Response Generation for Customer Service based on Personality Traits - Jonathan Herzig - Michal Shmueli-Scheuer - Tommy Sandbank - David Konopnicki + JonathanHerzig + MichalShmueli-Scheuer + TommySandbank + DavidKonopnicki 252–256 W17-3541 10.18653/v1/W17-3541 @@ -5992,8 +5992,8 @@ is able to handle phenomena related to scope by means of an higher-order type th Neural Paraphrase Generation using Transfer Learning - Florin Brad - Traian Rebedea + FlorinBrad + TraianRebedea 257–261 W17-3542 10.18653/v1/W17-3542 @@ -6022,91 +6022,91 @@ is able to handle phenomena related to scope by means of an higher-order type th Deliberation as Genre: Mapping Argumentation through Relational Discourse Structure - Oier Imaz - Mikel Iruskieta + OierImaz + MikelIruskieta 1–10 W17-3601 10.18653/v1/W17-3601 The Good, the Bad, and the Disagreement: Complex ground truth in rhetorical structure analysis - Debopam Das - Manfred Stede - Maite Taboada + DebopamDas + ManfredStede + MaiteTaboada 11–19 W17-3602 10.18653/v1/W17-3602 A Distributional View of Discourse Encapsulation: Multifactorial Prediction of Coreference Density in <fixed-case>RST</fixed-case> - Amir Zeldes + AmirZeldes 20–28 W17-3603 10.18653/v1/W17-3603 Rhetorical relations markers in <fixed-case>R</fixed-case>ussian <fixed-case>RST</fixed-case> Treebank - Svetlana Toldova - Dina Pisarevskaya - Margarita Ananyeva - Maria Kobozeva - Alexander Nasedkin - Sofia Nikiforova - Irina Pavlova - Alexey Shelepov + SvetlanaToldova + DinaPisarevskaya + MargaritaAnanyeva + MariaKobozeva + AlexanderNasedkin + SofiaNikiforova + IrinaPavlova + AlexeyShelepov 29–33 W17-3604 10.18653/v1/W17-3604 Applying the Rhetorical Structure Theory in <fixed-case>A</fixed-case>lzheimer patients’ speech - Anayeli Paulino - Gerardo Sierra + AnayeliPaulino + GerardoSierra 34–38 W17-3605 10.18653/v1/W17-3605 Using lexical level information in discourse structures for Basque sentiment analysis - Jon Alkorta - Koldo Gojenola - Mikel Iruskieta - Maite Taboada + JonAlkorta + KoldoGojenola + MikelIruskieta + MaiteTaboada 39–47 W17-3606 10.18653/v1/W17-3606 Framework for the Analysis of Simplified Texts Taking Discourse into Account: the Basque Causal Relations as Case Study - Itziar Gonzalez-Dios - Arantza Diaz de Ilarraza - Mikel Iruskieta + ItziarGonzalez-Dios + ArantzaDiaz de Ilarraza + MikelIruskieta 48–57 W17-3607 10.18653/v1/W17-3607 Using Rhetorical Structure Theory for Detection of Fake Online Reviews - Olu Popoola + OluPopoola 58–63 W17-3608 10.18653/v1/W17-3608 “Haters gonna hate”: challenges for sentiment analysis of <fixed-case>F</fixed-case>acebook comments in <fixed-case>B</fixed-case>razilian <fixed-case>P</fixed-case>ortuguese - Juliano D. Antonio - Ana Carolina L. Santin + Juliano D.Antonio + Ana Carolina L.Santin 64–72 W17-3609 10.18653/v1/W17-3609 Discourse Segmentation for Building a <fixed-case>RST</fixed-case> <fixed-case>C</fixed-case>hinese Treebank - Shuyuan Cao - Nianwen Xue - Iria da Cunha - Mikel Iruskieta - Chuan Wang + ShuyuanCao + NianwenXue + Iriada Cunha + MikelIruskieta + ChuanWang 73–81 W17-3610 10.18653/v1/W17-3610 @@ -6129,30 +6129,30 @@ is able to handle phenomena related to scope by means of an higher-order type th Two Challenges for <fixed-case>CI</fixed-case> Trustworthiness and How to Address Them - Kevin Baum - Maximilian A. Köhl - Eva Schmidt + KevinBaum + Maximilian A.Köhl + EvaSchmidt W17-3701 10.18653/v1/W17-3701 A Simple Method for Clarifying Sentences with Coordination Ambiguities - Michael White - Manjuan Duan - David L. King + MichaelWhite + ManjuanDuan + David L.King W17-3702 10.18653/v1/W17-3702 Requirements for Conceptual Representations of Explanations and How Reasoning Systems Can Serve Them - Helmut Horacek + HelmutHoracek W17-3703 10.18653/v1/W17-3703 An Essay on Self-explanatory Computational Intelligence: A Linguistic Model of Data Processing Systems - Jose M. Alonso - Gracian Trivino + Jose M.Alonso + GracianTrivino W17-3704 10.18653/v1/W17-3704 @@ -6175,66 +6175,66 @@ is able to handle phenomena related to scope by means of an higher-order type th From <fixed-case>FOAF</fixed-case> to <fixed-case>E</fixed-case>nglish: Linguistic Contribution to Web Semantics - Max Silberztein + MaxSilberztein 1–9 W17-3801 10.18653/v1/W17-3801 Lexicon for Natural Language Generation in <fixed-case>S</fixed-case>panish Adapted to Alternative and Augmentative Communication - Silvia García-Méndez - Milagros Fernández-Gavilanes - Enrique Costa-Montenegro - Jonathan Juncal-Martínez - Francisco J. González-Castaño + SilviaGarcía-Méndez + MilagrosFernández-Gavilanes + EnriqueCosta-Montenegro + JonathanJuncal-Martínez + Francisco J.González-Castaño 11-15 W17-3802 10.18653/v1/W17-3802 Generating Answering Patterns from Factoid <fixed-case>A</fixed-case>rabic Questions - Essia Bessaies - Slim Mesfar - Henda Ben Ghezala + EssiaBessaies + SlimMesfar + HendaBen Ghezala 17–24 W17-3803 10.18653/v1/W17-3803 Language Generation from <fixed-case>DB</fixed-case> Query - Kristina Kocijan - Božo Bekavac - Krešimir Šojat + KristinaKocijan + BožoBekavac + KrešimirŠojat 25–32 W17-3804 10.18653/v1/W17-3804 Using Electronic Dictionaries and <fixed-case>N</fixed-case>oo<fixed-case>J</fixed-case> to Generate Sentences Containing <fixed-case>E</fixed-case>nglish Phrasal Verbs - Peter A. Machonis + Peter A.Machonis 33–37 W17-3805 10.18653/v1/W17-3805 Generating Text with Correct Verb Conjugation: Proposal for a New Automatic Conjugator with <fixed-case>N</fixed-case>oo<fixed-case>J</fixed-case> - Héla Fehri - Sondes Dardour + HélaFehri + SondesDardour 39–42 W17-3806 10.18653/v1/W17-3806 Formalization of Speech Verbs with <fixed-case>N</fixed-case>oo<fixed-case>J</fixed-case> for Machine Translation: the <fixed-case>F</fixed-case>rench Verb accuser - Jouda Ghorbel + JoudaGhorbel 43–47 W17-3807 10.18653/v1/W17-3807 Using Serious Games to Correct <fixed-case>F</fixed-case>rench Dictations: Proposal for a New <fixed-case>U</fixed-case>nity3<fixed-case>D</fixed-case>/<fixed-case>N</fixed-case>oo<fixed-case>J</fixed-case> Connector - Ikram Bououd - Rania Fafi + IkramBououd + RaniaFafi 49–52 W17-3808 10.18653/v1/W17-3808 @@ -6244,10 +6244,10 @@ is able to handle phenomena related to scope by means of an higher-order type th Proceedings of the Workshop on Computational Creativity in Natural Language Generation (CC-NLG 2017) W17-39 - Hugo Gonçalo Oliveira - Ben Burtenshaw - Mike Kestemont - Tom De Smedt + HugoGonçalo Oliveira + BenBurtenshaw + MikeKestemont + TomDe Smedt Association for Computational Linguistics
Santiago de Compostela, Spain
September @@ -6259,43 +6259,43 @@ is able to handle phenomena related to scope by means of an higher-order type th Poet’s Little Helper: A methodology for computer-based poetry generation. A case study for the Basque language - Aitzol Astigarraga - José María Martínez-Otzeta - Igor Rodriguez - Basilio Sierra - Elena Lazkano + AitzolAstigarraga + JoséMaría Martínez-Otzeta + IgorRodriguez + BasilioSierra + ElenaLazkano 2–10 W17-3901 10.18653/v1/W17-3901 O Poeta Artificial 2.0: Increasing Meaningfulness in a Poetry Generation Twitter bot - Hugo Gonçalo Oliveira + HugoGonçalo Oliveira 11–20 W17-3902 10.18653/v1/W17-3902 Template-Free Construction of Rhyming Poems with Thematic Cohesion - Pablo Gervás + PabloGervás 21–28 W17-3903 10.18653/v1/W17-3903 Synthetic Literature: Writing Science Fiction in a Co-Creative Process - Enrique Manjavacas - Folgert Karsdorp - Ben Burtenshaw - Mike Kestemont + EnriqueManjavacas + FolgertKarsdorp + BenBurtenshaw + MikeKestemont 29–37 W17-3904 10.18653/v1/W17-3904 Constructing narrative using a generative model and continuous action policies - Emmanouil Theofanis Chourdakis - Joshua Reiss + EmmanouilTheofanis Chourdakis + JoshuaReiss 38–43 W17-3905 10.18653/v1/W17-3905 @@ -6418,8 +6418,8 @@ is able to handle phenomena related to scope by means of an higher-order type th Character and Subword-Based Word Representation for Neural Language Modeling Prediction - Matthieu Labeau - Alexandre Allauzen + MatthieuLabeau + AlexandreAllauzen 1–13 W17-4101 10.18653/v1/W17-4101 @@ -6427,8 +6427,8 @@ is able to handle phenomena related to scope by means of an higher-order type th Learning variable length units for <fixed-case>SMT</fixed-case> between related languages via Byte Pair Encoding - Anoop Kunchukuttan - Pushpak Bhattacharyya + AnoopKunchukuttan + PushpakBhattacharyya 14–24 W17-4102 10.18653/v1/W17-4102 @@ -6436,8 +6436,8 @@ is able to handle phenomena related to scope by means of an higher-order type th Character Based Pattern Mining for Neology Detection - Gaël Lejeune - Emmanuel Cartier + GaëlLejeune + EmmanuelCartier 25–30 W17-4103 10.18653/v1/W17-4103 @@ -6445,10 +6445,10 @@ is able to handle phenomena related to scope by means of an higher-order type th Automated Word Stress Detection in <fixed-case>R</fixed-case>ussian - Maria Ponomareva - Kirill Milintsevich - Ekaterina Chernyak - Anatoly Starostin + MariaPonomareva + KirillMilintsevich + EkaterinaChernyak + AnatolyStarostin 31–35 W17-4104 10.18653/v1/W17-4104 @@ -6456,10 +6456,10 @@ is able to handle phenomena related to scope by means of an higher-order type th A Syllable-based Technique for Word Embeddings of <fixed-case>K</fixed-case>orean Words - Sanghyuk Choi - Taeuk Kim - Jinseok Seol - Sang-goo Lee + SanghyukChoi + TaeukKim + JinseokSeol + Sang-gooLee 36–40 W17-4105 10.18653/v1/W17-4105 @@ -6467,8 +6467,8 @@ is able to handle phenomena related to scope by means of an higher-order type th Supersense Tagging with a Combination of Character, Subword, and Word-level Representations - Youhyun Shin - Sang-goo Lee + YouhyunShin + Sang-gooLee 41–45 W17-4106 10.18653/v1/W17-4106 @@ -6476,8 +6476,8 @@ is able to handle phenomena related to scope by means of an higher-order type th Weakly supervised learning of allomorphy - Miikka Silfverberg - Mans Hulden + MiikkaSilfverberg + MansHulden 46–56 W17-4107 10.18653/v1/W17-4107 @@ -6485,8 +6485,8 @@ is able to handle phenomena related to scope by means of an higher-order type th Character-based recurrent neural networks for morphological relational reasoning - Olof Mogren - Richard Johansson + OlofMogren + RichardJohansson 57–63 W17-4108 10.18653/v1/W17-4108 @@ -6508,8 +6508,8 @@ is able to handle phenomena related to scope by means of an higher-order type th Glyph-aware Embedding of <fixed-case>C</fixed-case>hinese Characters - Falcon Dai - Zheng Cai + FalconDai + ZhengCai 64–69 W17-4109 10.18653/v1/W17-4109 @@ -6517,8 +6517,8 @@ is able to handle phenomena related to scope by means of an higher-order type th Exploring Cross-Lingual Transfer of Morphological Knowledge In Sequence-to-Sequence Models - Huiming Jin - Katharina Kann + HuimingJin + KatharinaKann 70–75 W17-4110 10.18653/v1/W17-4110 @@ -6526,8 +6526,8 @@ is able to handle phenomena related to scope by means of an higher-order type th Unlabeled Data for Morphological Generation With Character-Based Sequence-to-Sequence Models - Katharina Kann - Hinrich Schütze + KatharinaKann + HinrichSchütze 76–81 W17-4111 10.18653/v1/W17-4111 @@ -6535,8 +6535,8 @@ is able to handle phenomena related to scope by means of an higher-order type th Vowel and Consonant Classification through Spectral Decomposition - Patricia Thaine - Gerald Penn + PatriciaThaine + GeraldPenn 82–91 W17-4112 10.18653/v1/W17-4112 @@ -6545,10 +6545,10 @@ is able to handle phenomena related to scope by means of an higher-order type th Syllable-level Neural Language Model for Agglutinative Language - Seunghak Yu - Nilesh Kulkarni - Haejun Lee - Jihie Kim + SeunghakYu + NileshKulkarni + HaejunLee + JihieKim 92–96 W17-4113 10.18653/v1/W17-4113 @@ -6557,10 +6557,10 @@ is able to handle phenomena related to scope by means of an higher-order type th Character-based Bidirectional <fixed-case>LSTM</fixed-case>-<fixed-case>CRF</fixed-case> with words and characters for <fixed-case>J</fixed-case>apanese Named Entity Recognition - Shotaro Misawa - Motoki Taniguchi - Yasuhide Miura - Tomoko Ohkuma + ShotaroMisawa + MotokiTaniguchi + YasuhideMiura + TomokoOhkuma 97–102 W17-4114 10.18653/v1/W17-4114 @@ -6568,10 +6568,10 @@ is able to handle phenomena related to scope by means of an higher-order type th Word Representation Models for Morphologically Rich Languages in Neural Machine Translation - Ekaterina Vylomova - Trevor Cohn - Xuanli He - Gholamreza Haffari + EkaterinaVylomova + TrevorCohn + XuanliHe + GholamrezaHaffari 103–108 W17-4115 10.18653/v1/W17-4115 @@ -6579,8 +6579,8 @@ is able to handle phenomena related to scope by means of an higher-order type th Spell-Checking based on Syllabification and Character-level Graphs for a <fixed-case>P</fixed-case>eruvian Agglutinative Language - Carlo Alva - Arturo Oncevay + CarloAlva + ArturoOncevay 109–116 W17-4116 10.18653/v1/W17-4116 @@ -6588,9 +6588,9 @@ is able to handle phenomena related to scope by means of an higher-order type th What do we need to know about an unknown word when parsing <fixed-case>G</fixed-case>erman - Bich-Ngoc Do - Ines Rehbein - Anette Frank + Bich-NgocDo + InesRehbein + AnetteFrank 117–123 W17-4117 10.18653/v1/W17-4117 @@ -6598,9 +6598,9 @@ is able to handle phenomena related to scope by means of an higher-order type th A General-Purpose Tagger with Convolutional Neural Networks - Xiang Yu - Agnieszka Falenska - Ngoc Thang Vu + XiangYu + AgnieszkaFalenska + Ngoc ThangVu 124–129 W17-4118 10.18653/v1/W17-4118 @@ -6608,7 +6608,7 @@ is able to handle phenomena related to scope by means of an higher-order type th Reconstruction of Word Embeddings from Sub-Word Parameters - Karl Stratos + KarlStratos 130–135 W17-4119 10.18653/v1/W17-4119 @@ -6616,9 +6616,9 @@ is able to handle phenomena related to scope by means of an higher-order type th Inflection Generation for <fixed-case>S</fixed-case>panish Verbs using Supervised Learning - Cristina Barros - Dimitra Gkatzia - Elena Lloret + CristinaBarros + DimitraGkatzia + ElenaLloret 136–141 W17-4120 10.18653/v1/W17-4120 @@ -6626,11 +6626,11 @@ is able to handle phenomena related to scope by means of an higher-order type th Neural Paraphrase Identification of Questions with Noisy Pretraining - Gaurav Singh Tomar - Thyago Duque - Oscar Täckström - Jakob Uszkoreit - Dipanjan Das + Gaurav SinghTomar + ThyagoDuque + OscarTäckström + JakobUszkoreit + DipanjanDas 142–147 W17-4121 10.18653/v1/W17-4121 @@ -6638,9 +6638,9 @@ is able to handle phenomena related to scope by means of an higher-order type th Sub-character Neural Language Modelling in <fixed-case>J</fixed-case>apanese - Viet Nguyen - Julian Brooke - Timothy Baldwin + VietNguyen + JulianBrooke + TimothyBaldwin 148–153 W17-4122 10.18653/v1/W17-4122 @@ -6648,9 +6648,9 @@ is able to handle phenomena related to scope by means of an higher-order type th Byte-based Neural Machine Translation - Marta R. Costa-jussà - Carlos Escolano - José A. R. Fonollosa + Marta R.Costa-jussà + CarlosEscolano + José A. R.Fonollosa 154–158 W17-4123 10.18653/v1/W17-4123 @@ -6658,8 +6658,8 @@ is able to handle phenomena related to scope by means of an higher-order type th Improving Opinion-Target Extraction with Character-Level Word Embeddings - Soufian Jebbara - Philipp Cimiano + SoufianJebbara + PhilippCimiano 159–167 W17-4124 10.18653/v1/W17-4124 @@ -6670,8 +6670,8 @@ is able to handle phenomena related to scope by means of an higher-order type th Proceedings of the 2017 EMNLP Workshop: Natural Language Processing meets Journalism W17-42 - Octavian Popescu - Carlo Strapparava + OctavianPopescu + CarloStrapparava 10.18653/v1/W17-42 Association for Computational Linguistics
Copenhagen, Denmark
@@ -6683,12 +6683,12 @@ is able to handle phenomena related to scope by means of an higher-order type th Predicting News Values from Headline Text and Emotions - Maria Pia di Buono - Jan Šnajder - Bojana Dalbelo Bašić - Goran Glavaš - Martin Tutek - Natasa Milic-Frayling + Maria Piadi Buono + JanŠnajder + BojanaDalbelo Bašić + GoranGlavaš + MartinTutek + NatasaMilic-Frayling 1–6 W17-4201 10.18653/v1/W17-4201 @@ -6696,8 +6696,8 @@ is able to handle phenomena related to scope by means of an higher-order type th Predicting User Views in Online News - Daniel Hardt - Owen Rambow + DanielHardt + OwenRambow 7–12 W17-4202 10.18653/v1/W17-4202 @@ -6705,11 +6705,11 @@ is able to handle phenomena related to scope by means of an higher-order type th Tracking Bias in News Sources Using Social Media: the Russia-<fixed-case>U</fixed-case>kraine Maidan Crisis of 2013–2014 - Peter Potash - Alexey Romanov - Mikhail Gronas - Anna Rumshisky - Mikhail Gronas + PeterPotash + AlexeyRomanov + MikhailGronas + AnnaRumshisky + MikhailGronas 13–18 W17-4203 10.18653/v1/W17-4203 @@ -6717,10 +6717,10 @@ is able to handle phenomena related to scope by means of an higher-order type th What to Write? A topic recommender for journalists - Alessandro Cucchiarelli - Christian Morbidoni - Giovanni Stilo - Paola Velardi + AlessandroCucchiarelli + ChristianMorbidoni + GiovanniStilo + PaolaVelardi 19–24 W17-4204 10.18653/v1/W17-4204 @@ -6728,11 +6728,11 @@ is able to handle phenomena related to scope by means of an higher-order type th Comparing Attitudes to Climate Change in the Media using sentiment analysis based on Latent <fixed-case>D</fixed-case>irichlet Allocation - Ye Jiang - Xingyi Song - Jackie Harrison - Shaun Quegan - Diana Maynard + YeJiang + XingyiSong + JackieHarrison + ShaunQuegan + DianaMaynard 25–30 W17-4205 10.18653/v1/W17-4205 @@ -6740,12 +6740,12 @@ is able to handle phenomena related to scope by means of an higher-order type th Language-based Construction of Explorable News Graphs for Journalists - Rémi Bois - Guillaume Gravier - Eric Jamet - Emmanuel Morin - Pascale Sébillot - Maxime Robert + RémiBois + GuillaumeGravier + EricJamet + EmmanuelMorin + PascaleSébillot + MaximeRobert 31–36 W17-4206 10.18653/v1/W17-4206 @@ -6753,13 +6753,13 @@ is able to handle phenomena related to scope by means of an higher-order type th <fixed-case>S</fixed-case>toryteller: Visual Analytics of Perspectives on Rich Text Interpretations - Maarten van Meersbergen - Piek Vossen - Janneke van der Zwaan - Antske Fokkens - Willem van Hage - Inger Leemans - Isa Maks + Maartenvan Meersbergen + PiekVossen + Jannekevan der Zwaan + AntskeFokkens + Willemvan Hage + IngerLeemans + IsaMaks 37–45 W17-4207 10.18653/v1/W17-4207 @@ -6767,10 +6767,10 @@ is able to handle phenomena related to scope by means of an higher-order type th Analyzing the Revision Logs of a <fixed-case>J</fixed-case>apanese Newspaper for Article Quality Assessment - Hideaki Tamori - Yuta Hitomi - Naoaki Okazaki - Kentaro Inui + HideakiTamori + YutaHitomi + NaoakiOkazaki + KentaroInui 46–50 W17-4208 10.18653/v1/W17-4208 @@ -6778,10 +6778,10 @@ is able to handle phenomena related to scope by means of an higher-order type th Improved Abusive Comment Moderation with User Embeddings - John Pavlopoulos - Prodromos Malakasiotis - Juli Bakagianni - Ion Androutsopoulos + JohnPavlopoulos + ProdromosMalakasiotis + JuliBakagianni + IonAndroutsopoulos 51–55 W17-4209 10.18653/v1/W17-4209 @@ -6789,10 +6789,10 @@ is able to handle phenomena related to scope by means of an higher-order type th Incongruent Headlines: Yet Another Way to Mislead Your Readers - Sophie Chesney - Maria Liakata - Massimo Poesio - Matthew Purver + SophieChesney + MariaLiakata + MassimoPoesio + MatthewPurver 56–61 W17-4210 10.18653/v1/W17-4210 @@ -6800,9 +6800,9 @@ is able to handle phenomena related to scope by means of an higher-order type th Unsupervised Event Clustering and Aggregation from Newswire and Web Articles - Swen Ribeiro - Olivier Ferret - Xavier Tannier + SwenRibeiro + OlivierFerret + XavierTannier 62–67 W17-4211 10.18653/v1/W17-4211 @@ -6810,11 +6810,11 @@ is able to handle phenomena related to scope by means of an higher-order type th Semantic Storytelling, Cross-lingual Event Detection and other Semantic Services for a Newsroom Content Curation Dashboard - Julian Moreno-Schneider - Ankit Srivastava - Peter Bourgonje - David Wabnitz - Georg Rehm + JulianMoreno-Schneider + AnkitSrivastava + PeterBourgonje + DavidWabnitz + GeorgRehm 68–73 W17-4212 10.18653/v1/W17-4212 @@ -6822,7 +6822,7 @@ is able to handle phenomena related to scope by means of an higher-order type th Deception Detection in News Reports in the <fixed-case>R</fixed-case>ussian Language: Lexics and Discourse - Dina Pisarevskaya + DinaPisarevskaya 74–79 W17-4213 10.18653/v1/W17-4213 @@ -6830,12 +6830,12 @@ is able to handle phenomena related to scope by means of an higher-order type th Fake news stance detection using stacked ensemble of classifiers - James Thorne - Mingjie Chen - Giorgos Myrianthous - Jiashu Pu - Xiaoxuan Wang - Andreas Vlachos + JamesThorne + MingjieChen + GiorgosMyrianthous + JiashuPu + XiaoxuanWang + AndreasVlachos 80–83 W17-4214 10.18653/v1/W17-4214 @@ -6843,9 +6843,9 @@ is able to handle phenomena related to scope by means of an higher-order type th From Clickbait to Fake News Detection: An Approach based on Detecting the Stance of Headlines to Articles - Peter Bourgonje - Julian Moreno Schneider - Georg Rehm + PeterBourgonje + JulianMoreno Schneider + GeorgRehm 84–89 W17-4215 10.18653/v1/W17-4215 @@ -6853,8 +6853,8 @@ is able to handle phenomena related to scope by means of an higher-order type th ‘Fighting’ or ‘Conflict’? An Approach to Revealing Concepts of Terms in Political Discourse - Linyuan Tang - Kyo Kageura + LinyuanTang + KyoKageura 90–94 W17-4216 10.18653/v1/W17-4216 @@ -6862,9 +6862,9 @@ is able to handle phenomena related to scope by means of an higher-order type th A News Chain Evaluation Methodology along with a Lattice-based Approach for News Chain Construction - Mustafa Toprak - Özer Özkahraman - Selma Tekir + MustafaToprak + ÖzerÖzkahraman + SelmaTekir 95–99 W17-4217 10.18653/v1/W17-4217 @@ -6872,8 +6872,8 @@ is able to handle phenomena related to scope by means of an higher-order type th Using New York Times Picks to Identify Constructive Comments - Varada Kolhatkar - Maite Taboada + VaradaKolhatkar + MaiteTaboada 100–105 W17-4218 10.18653/v1/W17-4218 @@ -6881,9 +6881,9 @@ is able to handle phenomena related to scope by means of an higher-order type th An <fixed-case>NLP</fixed-case> Analysis of Exaggerated Claims in Science News - Yingya Li - Jieke Zhang - Bei Yu + YingyaLi + JiekeZhang + BeiYu 106–111 W17-4219 10.18653/v1/W17-4219 @@ -6909,8 +6909,8 @@ is able to handle phenomena related to scope by means of an higher-order type th Dependency Parsing with Dilated Iterated Graph <fixed-case>CNN</fixed-case>s - Emma Strubell - Andrew McCallum + EmmaStrubell + AndrewMcCallum 1–6 W17-4301 10.18653/v1/W17-4301 @@ -6918,7 +6918,7 @@ is able to handle phenomena related to scope by means of an higher-order type th Entity Identification as Multitasking - Karl Stratos + KarlStratos 7–11 W17-4302 10.18653/v1/W17-4302 @@ -6926,8 +6926,8 @@ is able to handle phenomena related to scope by means of an higher-order type th Towards Neural Machine Translation with Latent Tree Attention - James Bradbury - Richard Socher + JamesBradbury + RichardSocher 12–16 W17-4303 10.18653/v1/W17-4303 @@ -6936,8 +6936,8 @@ is able to handle phenomena related to scope by means of an higher-order type th Structured Prediction via Learning to Search under Bandit Feedback - Amr Sharaf - Hal Daumé III + AmrSharaf + HalDaumé III 17–26 W17-4304 10.18653/v1/W17-4304 @@ -6945,11 +6945,11 @@ is able to handle phenomena related to scope by means of an higher-order type th Syntax Aware <fixed-case>LSTM</fixed-case> model for Semantic Role Labeling - Feng Qian - Lei Sha - Baobao Chang - Lu-chen Liu - Ming Zhang + FengQian + LeiSha + BaobaoChang + Lu-chenLiu + MingZhang 27–32 W17-4305 10.18653/v1/W17-4305 @@ -6957,9 +6957,9 @@ is able to handle phenomena related to scope by means of an higher-order type th Spatial Language Understanding with Multimodal Graphs using Declarative Learning based Programming - Parisa Kordjamshidi - Taher Rahgooy - Umar Manzoor + ParisaKordjamshidi + TaherRahgooy + UmarManzoor 33–43 W17-4306 10.18653/v1/W17-4306 @@ -6967,8 +6967,8 @@ is able to handle phenomena related to scope by means of an higher-order type th Boosting Information Extraction Systems with Character-level Neural Networks and Free Noisy Supervision - Philipp Meerkamp - Zhengyi Zhou + PhilippMeerkamp + ZhengyiZhou 44–51 W17-4307 10.18653/v1/W17-4307 @@ -6976,10 +6976,10 @@ is able to handle phenomena related to scope by means of an higher-order type th Piecewise Latent Variables for Neural Variational Text Processing - Iulian Vlad Serban - Alexander Ororbia II - Joelle Pineau - Aaron Courville + Iulian VladSerban + AlexanderOrorbia II + JoellePineau + AaronCourville 52–62 W17-4308 10.18653/v1/W17-4308 @@ -7005,7 +7005,7 @@ is able to handle phenomena related to scope by means of an higher-order type th Boundary-based <fixed-case>MWE</fixed-case> segmentation with text partitioning - Jake Williams + JakeWilliams 1–10 W17-4401 10.18653/v1/W17-4401 @@ -7013,11 +7013,11 @@ is able to handle phenomena related to scope by means of an higher-order type th Towards the Understanding of Gaming Audiences by Modeling Twitch Emotes - Francesco Barbieri - Luis Espinosa-Anke - Miguel Ballesteros - Juan Soler-Company - Horacio Saggion + FrancescoBarbieri + LuisEspinosa-Anke + MiguelBallesteros + JuanSoler-Company + HoracioSaggion 11–20 W17-4402 10.18653/v1/W17-4402 @@ -7025,9 +7025,9 @@ is able to handle phenomena related to scope by means of an higher-order type th Churn Identification in Microblogs using Convolutional Neural Networks with Structured Logical Knowledge - Mourad Gridach - Hatem Haddad - Hala Mulki + MouradGridach + HatemHaddad + HalaMulki 21–30 W17-4403 10.18653/v1/W17-4403 @@ -7036,9 +7036,9 @@ is able to handle phenomena related to scope by means of an higher-order type th To normalize, or not to normalize: The impact of normalization on Part-of-Speech tagging - Rob van der Goot - Barbara Plank - Malvina Nissim + Robvan der Goot + BarbaraPlank + MalvinaNissim 31–39 W17-4404 10.18653/v1/W17-4404 @@ -7046,10 +7046,10 @@ is able to handle phenomena related to scope by means of an higher-order type th Constructing an Alias List for Named Entities during an Event - Anietie Andy - Mark Dredze - Mugizi Rwebangira - Chris Callison-Burch + AnietieAndy + MarkDredze + MugiziRwebangira + ChrisCallison-Burch 40–44 W17-4405 10.18653/v1/W17-4405 @@ -7057,8 +7057,8 @@ is able to handle phenomena related to scope by means of an higher-order type th Incorporating Metadata into Content-Based User Embeddings - Linzi Xing - Michael J. Paul + LinziXing + Michael J.Paul 45–49 W17-4406 10.18653/v1/W17-4406 @@ -7066,9 +7066,9 @@ is able to handle phenomena related to scope by means of an higher-order type th Simple Queries as Distant Labels for Predicting Gender on Twitter - Chris Emmery - Grzegorz Chrupała - Walter Daelemans + ChrisEmmery + GrzegorzChrupała + WalterDaelemans 50–55 W17-4407 10.18653/v1/W17-4407 @@ -7076,9 +7076,9 @@ is able to handle phenomena related to scope by means of an higher-order type th A Dataset and Classifier for Recognizing Social Media <fixed-case>E</fixed-case>nglish - Su Lin Blodgett - Johnny Wei - Brendan O’Connor + Su LinBlodgett + JohnnyWei + BrendanO’Connor 56–61 W17-4408 10.18653/v1/W17-4408 @@ -7086,8 +7086,8 @@ is able to handle phenomena related to scope by means of an higher-order type th Evaluating hypotheses in geolocation on a very large sample of Twitter - Bahar Salehi - Anders Søgaard + BaharSalehi + AndersSøgaard 62–67 W17-4409 10.18653/v1/W17-4409 @@ -7095,9 +7095,9 @@ is able to handle phenomena related to scope by means of an higher-order type th The Effect of Error Rate in Artificially Generated Data for Automatic Preposition and Determiner Correction - Fraser Bowen - Jon Dehdari - Josef van Genabith + FraserBowen + JonDehdari + Josefvan Genabith 68–76 W17-4410 10.18653/v1/W17-4410 @@ -7105,10 +7105,10 @@ is able to handle phenomena related to scope by means of an higher-order type th An Entity Resolution Approach to Isolate Instances of Human Trafficking Online - Chirag Nagpal - Kyle Miller - Benedikt Boecking - Artur Dubrawski + ChiragNagpal + KyleMiller + BenediktBoecking + ArturDubrawski 77–84 W17-4411 10.18653/v1/W17-4411 @@ -7116,8 +7116,8 @@ is able to handle phenomena related to scope by means of an higher-order type th Noisy <fixed-case>U</fixed-case>yghur Text Normalization - Osman Tursun - Ruket Cakici + OsmanTursun + RuketCakici 85–93 W17-4412 10.18653/v1/W17-4412 @@ -7125,9 +7125,9 @@ is able to handle phenomena related to scope by means of an higher-order type th Crowdsourcing Multiple Choice Science Questions - Johannes Welbl - Nelson F. Liu - Matt Gardner + JohannesWelbl + Nelson F.Liu + MattGardner 94–106 W17-4413 10.18653/v1/W17-4413 @@ -7135,11 +7135,11 @@ is able to handle phenomena related to scope by means of an higher-order type th A Text Normalisation System for Non-Standard <fixed-case>E</fixed-case>nglish Words - Emma Flint - Elliot Ford - Olivia Thomas - Andrew Caines - Paula Buttery + EmmaFlint + ElliotFord + OliviaThomas + AndrewCaines + PaulaButtery 107–115 W17-4414 10.18653/v1/W17-4414 @@ -7147,10 +7147,10 @@ is able to handle phenomena related to scope by means of an higher-order type th Huntsville, hospitals, and hockey teams: Names can reveal your location - Bahar Salehi - Dirk Hovy - Eduard Hovy - Anders Søgaard + BaharSalehi + DirkHovy + EduardHovy + AndersSøgaard 116–121 W17-4415 10.18653/v1/W17-4415 @@ -7158,9 +7158,9 @@ is able to handle phenomena related to scope by means of an higher-order type th Improving Document Clustering by Removing Unnatural Language - Myungha Jang - Jinho D. Choi - James Allan + MyunghaJang + Jinho D.Choi + JamesAllan 122–130 W17-4416 10.18653/v1/W17-4416 @@ -7168,9 +7168,9 @@ is able to handle phenomena related to scope by means of an higher-order type th Lithium <fixed-case>NLP</fixed-case>: A System for Rich Information Extraction from Noisy User Generated Text on Social Media - Preeti Bhargava - Nemanja Spasojevic - Guoning Hu + PreetiBhargava + NemanjaSpasojevic + GuoningHu 131–139 W17-4417 10.18653/v1/W17-4417 @@ -7178,10 +7178,10 @@ is able to handle phenomena related to scope by means of an higher-order type th Results of the <fixed-case>WNUT</fixed-case>2017 Shared Task on Novel and Emerging Entity Recognition - Leon Derczynski - Eric Nichols - Marieke van Erp - Nut Limsopatham + LeonDerczynski + EricNichols + Mariekevan Erp + NutLimsopatham 140–147 W17-4418 10.18653/v1/W17-4418 @@ -7189,10 +7189,10 @@ is able to handle phenomena related to scope by means of an higher-order type th A Multi-task Approach for Named Entity Recognition in Social Media Data - Gustavo Aguilar - Suraj Maharjan - Adrian Pastor López-Monroy - Thamar Solorio + GustavoAguilar + SurajMaharjan + Adrian PastorLópez-Monroy + ThamarSolorio 148–153 W17-4419 10.18653/v1/W17-4419 @@ -7200,8 +7200,8 @@ is able to handle phenomena related to scope by means of an higher-order type th Distributed Representation, <fixed-case>LDA</fixed-case> Topic Modelling and Deep Learning for Emerging Named Entity Recognition from Social Media - Patrick Jansson - Shuhua Liu + PatrickJansson + ShuhuaLiu 154–159 W17-4420 10.18653/v1/W17-4420 @@ -7209,10 +7209,10 @@ is able to handle phenomena related to scope by means of an higher-order type th Multi-channel <fixed-case>B</fixed-case>i<fixed-case>LSTM</fixed-case>-<fixed-case>CRF</fixed-case> Model for Emerging Named Entity Recognition in Social Media - Bill Y. Lin - Frank Xu - Zhiyi Luo - Kenny Zhu + Bill Y.Lin + FrankXu + ZhiyiLuo + KennyZhu 160–165 W17-4421 10.18653/v1/W17-4421 @@ -7220,8 +7220,8 @@ is able to handle phenomena related to scope by means of an higher-order type th Transfer Learning and Sentence Level Features for Named Entity Recognition on Tweets - Pius von Däniken - Mark Cieliebak + Piusvon Däniken + MarkCieliebak 166–171 W17-4422 10.18653/v1/W17-4422 @@ -7229,8 +7229,8 @@ is able to handle phenomena related to scope by means of an higher-order type th Context-Sensitive Recognition for Emerging and Rare Entities - Jake Williams - Giovanni Santia + JakeWilliams + GiovanniSantia 172–176 W17-4423 10.18653/v1/W17-4423 @@ -7238,8 +7238,8 @@ is able to handle phenomena related to scope by means of an higher-order type th A Feature-based Ensemble Approach to Recognition of Emerging and Rare Named Entities - Utpal Kumar Sikdar - Björn Gambäck + Utpal KumarSikdar + BjörnGambäck 177–181 W17-4424 10.18653/v1/W17-4424 @@ -7265,8 +7265,8 @@ is able to handle phenomena related to scope by means of an higher-order type th Video Highlights Detection and Summarization with Lag-Calibration based on Concept-Emotion Mapping of Crowdsourced Time-Sync Comments - Qing Ping - Chaomei Chen + QingPing + ChaomeiChen 1–11 W17-4501 10.18653/v1/W17-4501 @@ -7274,8 +7274,8 @@ is able to handle phenomena related to scope by means of an higher-order type th Multimedia Summary Generation from Online Conversations: Current Approaches and Future Directions - Enamul Hoque - Giuseppe Carenini + EnamulHoque + GiuseppeCarenini 12–19 W17-4502 10.18653/v1/W17-4502 @@ -7283,8 +7283,8 @@ is able to handle phenomena related to scope by means of an higher-order type th Low-Resource Neural Headline Generation - Ottokar Tilk - Tanel Alumäe + OttokarTilk + TanelAlumäe 20–26 W17-4503 10.18653/v1/W17-4503 @@ -7292,9 +7292,9 @@ is able to handle phenomena related to scope by means of an higher-order type th Towards Improving Abstractive Summarization via Entailment Generation - Ramakanth Pasunuru - Han Guo - Mohit Bansal + RamakanthPasunuru + HanGuo + MohitBansal 27–32 W17-4504 10.18653/v1/W17-4504 @@ -7302,8 +7302,8 @@ is able to handle phenomena related to scope by means of an higher-order type th Coarse-to-Fine Attention Models for Document Summarization - Jeffrey Ling - Alexander Rush + JeffreyLing + AlexanderRush 33–42 W17-4505 10.18653/v1/W17-4505 @@ -7311,11 +7311,11 @@ is able to handle phenomena related to scope by means of an higher-order type th Automatic Community Creation for Abstractive Spoken Conversations Summarization - Karan Singla - Evgeny Stepanov - Ali Orkan Bayer - Giuseppe Carenini - Giuseppe Riccardi + KaranSingla + EvgenyStepanov + Ali OrkanBayer + GiuseppeCarenini + GiuseppeRiccardi 43–47 W17-4506 10.18653/v1/W17-4506 @@ -7323,9 +7323,9 @@ is able to handle phenomena related to scope by means of an higher-order type th Combining Graph Degeneracy and Submodularity for Unsupervised Extractive Summarization - Antoine Tixier - Polykarpos Meladianos - Michalis Vazirgiannis + AntoineTixier + PolykarposMeladianos + MichalisVazirgiannis 48–58 W17-4507 10.18653/v1/W17-4507 @@ -7333,10 +7333,10 @@ is able to handle phenomena related to scope by means of an higher-order type th <fixed-case>TL</fixed-case>;<fixed-case>DR</fixed-case>: Mining <fixed-case>R</fixed-case>eddit to Learn Automatic Summarization - Michael Völske - Martin Potthast - Shahbaz Syed - Benno Stein + MichaelVölske + MartinPotthast + ShahbazSyed + BennoStein 59–63 W17-4508 10.18653/v1/W17-4508 @@ -7344,8 +7344,8 @@ is able to handle phenomena related to scope by means of an higher-order type th Topic Model Stability for Hierarchical Summarization - John Miller - Kathleen McCoy + JohnMiller + KathleenMcCoy 64–73 W17-4509 10.18653/v1/W17-4509 @@ -7354,9 +7354,9 @@ is able to handle phenomena related to scope by means of an higher-order type th Learning to Score System Summaries for Better Content Selection Evaluation. - Maxime Peyrard - Teresa Botschen - Iryna Gurevych + MaximePeyrard + TeresaBotschen + IrynaGurevych 74–84 W17-4510 10.18653/v1/W17-4510 @@ -7364,7 +7364,7 @@ is able to handle phenomena related to scope by means of an higher-order type th Revisiting the Centroid-based Method: A Strong Baseline for Multi-Document Summarization - Demian Gholipour Ghalandari + DemianGholipour Ghalandari 85–90 W17-4511 10.18653/v1/W17-4511 @@ -7372,9 +7372,9 @@ is able to handle phenomena related to scope by means of an higher-order type th Reader-Aware Multi-Document Summarization: An Enhanced Model and The First Dataset - Piji Li - Lidong Bing - Wai Lam + PijiLi + LidongBing + WaiLam 91–99 W17-4512 10.18653/v1/W17-4512 @@ -7382,8 +7382,8 @@ is able to handle phenomena related to scope by means of an higher-order type th A Pilot Study of Domain Adaptation Effect for Neural Abstractive Summarization - Xinyu Hua - Lu Wang + XinyuHua + LuWang 100–106 W17-4513 10.18653/v1/W17-4513 @@ -7407,10 +7407,10 @@ is able to handle phenomena related to scope by means of an higher-order type th Functions of Silences towards Information Flow in Spoken Conversation - Shammur Absar Chowdhury - Evgeny Stepanov - Morena Danieli - Giuseppe Riccardi + Shammur AbsarChowdhury + EvgenyStepanov + MorenaDanieli + GiuseppeRiccardi 1–9 W17-4601 10.18653/v1/W17-4601 @@ -7418,8 +7418,8 @@ is able to handle phenomena related to scope by means of an higher-order type th Encoding Word Confusion Networks with Recurrent Neural Networks for Dialog State Tracking - Glorianna Jagfeld - Ngoc Thang Vu + GloriannaJagfeld + Ngoc ThangVu 10–17 W17-4602 10.18653/v1/W17-4602 @@ -7427,8 +7427,8 @@ is able to handle phenomena related to scope by means of an higher-order type th Analyzing Human and Machine Performance In Resolving Ambiguous Spoken Sentences - Hussein Ghaly - Michael Mandel + HusseinGhaly + MichaelMandel 18–26 W17-4603 10.18653/v1/W17-4603 @@ -7437,9 +7437,9 @@ is able to handle phenomena related to scope by means of an higher-order type th Parsing transcripts of speech - Andrew Caines - Michael McCarthy - Paula Buttery + AndrewCaines + MichaelMcCarthy + PaulaButtery 27–36 W17-4604 10.18653/v1/W17-4604 @@ -7447,8 +7447,8 @@ is able to handle phenomena related to scope by means of an higher-order type th Enriching <fixed-case>ASR</fixed-case> Lattices with <fixed-case>POS</fixed-case> Tags for Dependency Parsing - Moritz Stiefel - Ngoc Thang Vu + MoritzStiefel + Ngoc ThangVu 37–47 W17-4605 10.18653/v1/W17-4605 @@ -7456,10 +7456,10 @@ is able to handle phenomena related to scope by means of an higher-order type th End-to-End Information Extraction without Token-Level Supervision - Rasmus Berg Palm - Dirk Hovy - Florian Laws - Ole Winther + Rasmus BergPalm + DirkHovy + FlorianLaws + OleWinther 48–52 W17-4606 10.18653/v1/W17-4606 @@ -7467,11 +7467,11 @@ is able to handle phenomena related to scope by means of an higher-order type th Spoken Term Discovery for Language Documentation using Translations - Antonios Anastasopoulos - Sameer Bansal - David Chiang - Sharon Goldwater - Adam Lopez + AntoniosAnastasopoulos + SameerBansal + DavidChiang + SharonGoldwater + AdamLopez 53–58 W17-4607 10.18653/v1/W17-4607 @@ -7479,9 +7479,9 @@ is able to handle phenomena related to scope by means of an higher-order type th <fixed-case>A</fixed-case>mharic-<fixed-case>E</fixed-case>nglish Speech Translation in Tourism Domain - Michael Melese - Laurent Besacier - Million Meshesha + MichaelMelese + LaurentBesacier + MillionMeshesha 59–66 W17-4608 10.18653/v1/W17-4608 @@ -7489,9 +7489,9 @@ is able to handle phenomena related to scope by means of an higher-order type th Speech- and Text-driven Features for Automated Scoring of <fixed-case>E</fixed-case>nglish Speaking Tasks - Anastassia Loukina - Nitin Madnani - Aoife Cahill + AnastassiaLoukina + NitinMadnani + AoifeCahill 67–77 W17-4609 10.18653/v1/W17-4609 @@ -7499,10 +7499,10 @@ is able to handle phenomena related to scope by means of an higher-order type th Improving coreference resolution with automatically predicted prosodic information - Ina Roesiger - Sabrina Stehwien - Arndt Riester - Ngoc Thang Vu + InaRoesiger + SabrinaStehwien + ArndtRiester + Ngoc ThangVu 78–83 W17-4610 10.18653/v1/W17-4610 @@ -7534,456 +7534,456 @@ is able to handle phenomena related to scope by means of an higher-order type th Sense-Aware Statistical Machine Translation using Adaptive Context-Dependent Clustering - Xiao Pu - Nikolaos Pappas - Andrei Popescu-Belis + XiaoPu + NikolaosPappas + AndreiPopescu-Belis 1–10 W17-4701 10.18653/v1/W17-4701 Improving Word Sense Disambiguation in Neural Machine Translation with Sense Embeddings - Annette Rios Gonzales - Laura Mascarell - Rico Sennrich + AnnetteRios Gonzales + LauraMascarell + RicoSennrich 11–19 W17-4702 10.18653/v1/W17-4702 Word Representations in Factored Neural Machine Translation - Franck Burlot - Mercedes García-Martínez - Loïc Barrault - Fethi Bougares - François Yvon + FranckBurlot + MercedesGarcía-Martínez + LoïcBarrault + FethiBougares + FrançoisYvon 20–31 W17-4703 10.18653/v1/W17-4703 Modeling Target-Side Inflection in Neural Machine Translation - Aleš Tamchyna - Marion Weller-Di Marco - Alexander Fraser + AlešTamchyna + MarionWeller-Di Marco + AlexanderFraser 32–42 W17-4704 10.18653/v1/W17-4704 Evaluating the morphological competence of Machine Translation Systems - Franck Burlot - François Yvon + FranckBurlot + FrançoisYvon 43–55 W17-4705 10.18653/v1/W17-4705 Target-side Word Segmentation Strategies for Neural Machine Translation - Matthias Huck - Simon Riess - Alexander Fraser + MatthiasHuck + SimonRiess + AlexanderFraser 56–67 W17-4706 10.18653/v1/W17-4706 Predicting Target Language <fixed-case>CCG</fixed-case> Supertags Improves Neural Machine Translation - Maria Nădejde - Siva Reddy - Rico Sennrich - Tomasz Dwojak - Marcin Junczys-Dowmunt - Philipp Koehn - Alexandra Birch + MariaNădejde + SivaReddy + RicoSennrich + TomaszDwojak + MarcinJunczys-Dowmunt + PhilippKoehn + AlexandraBirch 68–79 W17-4707 10.18653/v1/W17-4707 Exploiting Linguistic Resources for Neural Machine Translation Using Multi-task Learning - Jan Niehues - Eunah Cho + JanNiehues + EunahCho 80–89 W17-4708 10.18653/v1/W17-4708 Tree as a Pivot: Syntactic Matching Methods in Pivot Translation - Akiva Miura - Graham Neubig - Katsuhito Sudoh - Satoshi Nakamura + AkivaMiura + GrahamNeubig + KatsuhitoSudoh + SatoshiNakamura 90–98 W17-4709 10.18653/v1/W17-4709 Deep architectures for Neural Machine Translation - Antonio Valerio Miceli Barone - Jindřich Helcl - Rico Sennrich - Barry Haddow - Alexandra Birch + Antonio ValerioMiceli Barone + JindřichHelcl + RicoSennrich + BarryHaddow + AlexandraBirch 99–107 W17-4710 10.18653/v1/W17-4710 Biasing Attention-Based Recurrent Neural Networks Using External Alignment Information - Tamer Alkhouli - Hermann Ney + TamerAlkhouli + HermannNey 108–117 W17-4711 10.18653/v1/W17-4711 Effective Domain Mixing for Neural Machine Translation - Denny Britz - Quoc Le - Reid Pryzant + DennyBritz + QuocLe + ReidPryzant 118–126 W17-4712 10.18653/v1/W17-4712 Multi-Domain Neural Machine Translation through Unsupervised Adaptation - M. Amin Farajian - Marco Turchi - Matteo Negri - Marcello Federico + M. AminFarajian + MarcoTurchi + MatteoNegri + MarcelloFederico 127–137 W17-4713 10.18653/v1/W17-4713 Adapting Neural Machine Translation with Parallel Synthetic Data - Mara Chinea-Ríos - Álvaro Peris - Francisco Casacuberta + MaraChinea-Ríos + ÁlvaroPeris + FranciscoCasacuberta 138–147 W17-4714 10.18653/v1/W17-4714 Copied Monolingual Data Improves Low-Resource Neural Machine Translation - Anna Currey - Antonio Valerio Miceli Barone - Kenneth Heafield + AnnaCurrey + Antonio ValerioMiceli Barone + KennethHeafield 148–156 W17-4715 10.18653/v1/W17-4715 Guiding Neural Machine Translation Decoding with External Knowledge - Rajen Chatterjee - Matteo Negri - Marco Turchi - Marcello Federico - Lucia Specia - Frédéric Blain + RajenChatterjee + MatteoNegri + MarcoTurchi + MarcelloFederico + LuciaSpecia + FrédéricBlain 157–168 W17-4716 10.18653/v1/W17-4716 Findings of the 2017 Conference on Machine Translation (<fixed-case>WMT</fixed-case>17) - Ondřej Bojar - Rajen Chatterjee - Christian Federmann - Yvette Graham - Barry Haddow - Shujian Huang - Matthias Huck - Philipp Koehn - Qun Liu - Varvara Logacheva - Christof Monz - Matteo Negri - Matt Post - Raphael Rubino - Lucia Specia - Marco Turchi + OndřejBojar + RajenChatterjee + ChristianFedermann + YvetteGraham + BarryHaddow + ShujianHuang + MatthiasHuck + PhilippKoehn + QunLiu + VarvaraLogacheva + ChristofMonz + MatteoNegri + MattPost + RaphaelRubino + LuciaSpecia + MarcoTurchi 169–214 W17-4717 10.18653/v1/W17-4717 Findings of the Second Shared Task on Multimodal Machine Translation and Multilingual Image Description - Desmond Elliott - Stella Frank - Loïc Barrault - Fethi Bougares - Lucia Specia + DesmondElliott + StellaFrank + LoïcBarrault + FethiBougares + LuciaSpecia 215–233 W17-4718 10.18653/v1/W17-4718 Findings of the <fixed-case>WMT</fixed-case> 2017 Biomedical Translation Shared Task - Antonio Jimeno Yepes - Aurélie Névéol - Mariana Neves - Karin Verspoor - Ondřej Bojar - Arthur Boyer - Cristian Grozea - Barry Haddow - Madeleine Kittner - Yvonne Lichtblau - Pavel Pecina - Roland Roller - Rudolf Rosa - Amy Siu - Philippe Thomas - Saskia Trescher + AntonioJimeno Yepes + AurélieNévéol + MarianaNeves + KarinVerspoor + OndřejBojar + ArthurBoyer + CristianGrozea + BarryHaddow + MadeleineKittner + YvonneLichtblau + PavelPecina + RolandRoller + RudolfRosa + AmySiu + PhilippeThomas + SaskiaTrescher 234–247 W17-4719 10.18653/v1/W17-4719 <fixed-case>CUNI</fixed-case> submission in <fixed-case>WMT</fixed-case>17: Chimera goes neural - Roman Sudarikov - David Mareček - Tom Kocmi - Dušan Variš - Ondřej Bojar + RomanSudarikov + DavidMareček + TomKocmi + DušanVariš + OndřejBojar 248–256 W17-4720 10.18653/v1/W17-4720 LIMSI@WMT’17 - Franck Burlot - Pooyan Safari - Matthieu Labeau - Alexandre Allauzen - François Yvon + FranckBurlot + PooyanSafari + MatthieuLabeau + AlexandreAllauzen + FrançoisYvon 257–264 W17-4721 10.18653/v1/W17-4721 <fixed-case>SYSTRAN</fixed-case> Purely Neural <fixed-case>MT</fixed-case> Engines for <fixed-case>WMT</fixed-case>2017 - Yongchao Deng - Jungi Kim - Guillaume Klein - Catherine Kobus - Natalia Segal - Christophe Servan - Bo Wang - Dakun Zhang - Josep Crego - Jean Senellart + YongchaoDeng + JungiKim + GuillaumeKlein + CatherineKobus + NataliaSegal + ChristopheServan + BoWang + DakunZhang + JosepCrego + JeanSenellart 265–270 W17-4722 10.18653/v1/W17-4722 <fixed-case>FBK</fixed-case>’s Participation to the <fixed-case>E</fixed-case>nglish-to-<fixed-case>G</fixed-case>erman News Translation Task of <fixed-case>WMT</fixed-case> 2017 - Mattia Antonino Di Gangi - Nicola Bertoldi - Marcello Federico + Mattia AntoninoDi Gangi + NicolaBertoldi + MarcelloFederico 271–275 W17-4723 10.18653/v1/W17-4723 The <fixed-case>JHU</fixed-case> Machine Translation Systems for <fixed-case>WMT</fixed-case> 2017 - Shuoyang Ding - Huda Khayrallah - Philipp Koehn - Matt Post - Gaurav Kumar - Kevin Duh + ShuoyangDing + HudaKhayrallah + PhilippKoehn + MattPost + GauravKumar + KevinDuh 276–282 W17-4724 10.18653/v1/W17-4724 The <fixed-case>TALP</fixed-case>-<fixed-case>UPC</fixed-case> Neural Machine Translation System for <fixed-case>G</fixed-case>erman/<fixed-case>F</fixed-case>innish-<fixed-case>E</fixed-case>nglish Using the Inverse Direction Model in Rescoring - Carlos Escolano - Marta R. Costa-jussà - José A. R. Fonollosa + CarlosEscolano + Marta R.Costa-jussà + José A. R.Fonollosa 283–287 W17-4725 10.18653/v1/W17-4725 <fixed-case>LIUM</fixed-case> Machine Translation Systems for <fixed-case>WMT</fixed-case>17 News Translation Task - Mercedes García-Martínez - Ozan Caglayan - Walid Aransa - Adrien Bardet - Fethi Bougares - Loïc Barrault + MercedesGarcía-Martínez + OzanCaglayan + WalidAransa + AdrienBardet + FethiBougares + LoïcBarrault 288–295 W17-4726 10.18653/v1/W17-4726 Extending hybrid word-character neural machine translation with multi-task learning of morphological analysis - Stig-Arne Grönroos - Sami Virpioja - Mikko Kurimo + Stig-ArneGrönroos + SamiVirpioja + MikkoKurimo 296–302 W17-4727 10.18653/v1/W17-4727 The <fixed-case>AFRL-MITLL</fixed-case> <fixed-case>WMT17</fixed-case> Systems: Old, New, Borrowed, <fixed-case>BLEU</fixed-case> - Jeremy Gwinnup - Timothy Anderson - Grant Erdmann - Katherine Young - Michaeel Kazi - Elizabeth Salesky - Brian Thompson - Jonathan Taylor + JeremyGwinnup + TimothyAnderson + GrantErdmann + KatherineYoung + MichaeelKazi + ElizabethSalesky + BrianThompson + JonathanTaylor 303–309 W17-4728 10.18653/v1/W17-4728 University of Rochester <fixed-case>WMT</fixed-case> 2017 <fixed-case>NMT</fixed-case> System Submission - Chester Holtz - Chuyang Ke - Daniel Gildea + ChesterHoltz + ChuyangKe + DanielGildea 310–314 W17-4729 10.18653/v1/W17-4729 <fixed-case>LMU</fixed-case> Munich’s Neural Machine Translation Systems for News Articles and Health Information Texts - Matthias Huck - Fabienne Braune - Alexander Fraser + MatthiasHuck + FabienneBraune + AlexanderFraser 315–322 W17-4730 10.18653/v1/W17-4730 Rule-based Machine translation from <fixed-case>E</fixed-case>nglish to <fixed-case>F</fixed-case>innish - Arvi Hurskainen - Jörg Tiedemann + ArviHurskainen + JörgTiedemann 323–329 W17-4731 10.18653/v1/W17-4731 <fixed-case>NRC</fixed-case> Machine Translation System for <fixed-case>WMT</fixed-case> 2017 - Chi-kiu Lo - Boxing Chen - Colin Cherry - George Foster - Samuel Larkin - Darlene Stewart - Roland Kuhn + Chi-kiuLo + BoxingChen + ColinCherry + GeorgeFoster + SamuelLarkin + DarleneStewart + RolandKuhn 330–337 W17-4732 10.18653/v1/W17-4732 The <fixed-case>H</fixed-case>elsinki Neural Machine Translation System - Robert Östling - Yves Scherrer - Jörg Tiedemann - Gongbo Tang - Tommi Nieminen + RobertÖstling + YvesScherrer + JörgTiedemann + GongboTang + TommiNieminen 338–347 W17-4733 10.18653/v1/W17-4733 The <fixed-case>QT</fixed-case>21 Combined Machine Translation System for <fixed-case>E</fixed-case>nglish to <fixed-case>L</fixed-case>atvian - Jan-Thorsten Peter - Hermann Ney - Ondřej Bojar - Ngoc-Quan Pham - Jan Niehues - Alex Waibel - Franck Burlot - François Yvon - Mārcis Pinnis - Valters Šics - Joost Bastings - Miguel Rios - Wilker Aziz - Philip Williams - Frédéric Blain - Lucia Specia + Jan-ThorstenPeter + HermannNey + OndřejBojar + Ngoc-QuanPham + JanNiehues + AlexWaibel + FranckBurlot + FrançoisYvon + MārcisPinnis + ValtersŠics + JoostBastings + MiguelRios + WilkerAziz + PhilipWilliams + FrédéricBlain + LuciaSpecia 348–357 W17-4734 10.18653/v1/W17-4734 The <fixed-case>RWTH</fixed-case> Aachen University <fixed-case>E</fixed-case>nglish-<fixed-case>G</fixed-case>erman and <fixed-case>G</fixed-case>erman-<fixed-case>E</fixed-case>nglish Machine Translation System for <fixed-case>WMT</fixed-case> 2017 - Jan-Thorsten Peter - Andreas Guta - Tamer Alkhouli - Parnia Bahar - Jan Rosendahl - Nick Rossenbach - Miguel Graça - Hermann Ney + Jan-ThorstenPeter + AndreasGuta + TamerAlkhouli + ParniaBahar + JanRosendahl + NickRossenbach + MiguelGraça + HermannNey 358–365 W17-4735 10.18653/v1/W17-4735 The Karlsruhe Institute of Technology Systems for the News Translation Task in <fixed-case>WMT</fixed-case> 2017 - Ngoc-Quan Pham - Jan Niehues - Thanh-Le Ha - Eunah Cho - Matthias Sperber - Alexander Waibel + Ngoc-QuanPham + JanNiehues + Thanh-LeHa + EunahCho + MatthiasSperber + AlexanderWaibel 366–373 W17-4736 10.18653/v1/W17-4736 Tilde’s Machine Translation Systems for <fixed-case>WMT</fixed-case> 2017 - Mārcis Pinnis - Rihards Krišlauks - Toms Miks - Daiga Deksne - Valters Šics + MārcisPinnis + RihardsKrišlauks + TomsMiks + DaigaDeksne + ValtersŠics 374–381 W17-4737 10.18653/v1/W17-4737 C-3<fixed-case>MA</fixed-case>: Tartu-<fixed-case>R</fixed-case>iga-<fixed-case>Z</fixed-case>urich Translation Systems for <fixed-case>WMT</fixed-case>17 - Matīss Rikters - Chantal Amrhein - Maksym Del - Mark Fishel + MatīssRikters + ChantalAmrhein + MaksymDel + MarkFishel 382–388 W17-4738 10.18653/v1/W17-4738 The University of <fixed-case>E</fixed-case>dinburgh’s Neural <fixed-case>MT</fixed-case> Systems for <fixed-case>WMT</fixed-case>17 - Rico Sennrich - Alexandra Birch - Anna Currey - Ulrich Germann - Barry Haddow - Kenneth Heafield - Antonio Valerio Miceli Barone - Philip Williams + RicoSennrich + AlexandraBirch + AnnaCurrey + UlrichGermann + BarryHaddow + KennethHeafield + Antonio ValerioMiceli Barone + PhilipWilliams 389–399 W17-4739 10.18653/v1/W17-4739 <fixed-case>XMU</fixed-case> Neural Machine Translation Systems for <fixed-case>WMT</fixed-case> 17 - Zhixing Tan - Boli Wang - Jinming Hu - Yidong Chen - Xiaodong Shi + ZhixingTan + BoliWang + JinmingHu + YidongChen + XiaodongShi 400–404 W17-4740 10.18653/v1/W17-4740 @@ -7991,32 +7991,32 @@ is able to handle phenomena related to scope by means of an higher-order type th The <fixed-case>JAIST</fixed-case> Machine Translation Systems for <fixed-case>WMT</fixed-case> 17 - Hai-Long Trieu - Trung-Tin Pham - Le-Minh Nguyen + Hai-LongTrieu + Trung-TinPham + Le-MinhNguyen 405–409 W17-4741 10.18653/v1/W17-4741 Sogou Neural Machine Translation Systems for <fixed-case>WMT</fixed-case>17 - Yuguang Wang - Shanbo Cheng - Liyang Jiang - Jiajun Yang - Wei Chen - Muze Li - Lin Shi - Yanfeng Wang - Hongtao Yang + YuguangWang + ShanboCheng + LiyangJiang + JiajunYang + WeiChen + MuzeLi + LinShi + YanfengWang + HongtaoYang 410–415 W17-4742 10.18653/v1/W17-4742 <fixed-case>PJIIT</fixed-case>’s systems for <fixed-case>WMT</fixed-case> 2017 Conference - Krzysztof Wolk - Krzysztof Marasek + KrzysztofWolk + KrzysztofMarasek 416–421 W17-4743 10.18653/v1/W17-4743 @@ -8024,296 +8024,296 @@ is able to handle phenomena related to scope by means of an higher-order type th Hunter <fixed-case>MT</fixed-case>: A Course for Young Researchers in <fixed-case>WMT</fixed-case>17 - Jia Xu - Yi Zong Kuang - Shondell Baijoo - Jacob Hyun Lee - Uman Shahzad - Mir Ahmed - Meredith Lancaster - Chris Carlan + JiaXu + Yi ZongKuang + ShondellBaijoo + Jacob HyunLee + UmanShahzad + MirAhmed + MeredithLancaster + ChrisCarlan 422–427 W17-4744 10.18653/v1/W17-4744 <fixed-case>CASICT</fixed-case>-<fixed-case>DCU</fixed-case> Neural Machine Translation Systems for <fixed-case>WMT</fixed-case>17 - Jinchao Zhang - Peerachet Porkaew - Jiawei Hu - Qiuye Zhao - Qun Liu + JinchaoZhang + PeerachetPorkaew + JiaweiHu + QiuyeZhao + QunLiu 428–431 W17-4745 10.18653/v1/W17-4745 <fixed-case>LIUM</fixed-case>-<fixed-case>CVC</fixed-case> Submissions for <fixed-case>WMT</fixed-case>17 Multimodal Translation Task - Ozan Caglayan - Walid Aransa - Adrien Bardet - Mercedes García-Martínez - Fethi Bougares - Loïc Barrault - Marc Masana - Luis Herranz - Joost van de Weijer + OzanCaglayan + WalidAransa + AdrienBardet + MercedesGarcía-Martínez + FethiBougares + LoïcBarrault + MarcMasana + LuisHerranz + Joostvan de Weijer 432–439 W17-4746 10.18653/v1/W17-4746 <fixed-case>DCU</fixed-case> System Report on the <fixed-case>WMT</fixed-case> 2017 Multi-modal Machine Translation Task - Iacer Calixto - Koel Dutta Chowdhury - Qun Liu + IacerCalixto + KoelDutta Chowdhury + QunLiu 440–444 W17-4747 10.18653/v1/W17-4747 The <fixed-case>AFRL</fixed-case>-<fixed-case>OSU</fixed-case> <fixed-case>WMT</fixed-case>17 Multimodal Translation System: An Image Processing Approach - John Duselis - Michael Hutt - Jeremy Gwinnup - James Davis - Joshua Sandvick + JohnDuselis + MichaelHutt + JeremyGwinnup + JamesDavis + JoshuaSandvick 445–449 W17-4748 10.18653/v1/W17-4748 <fixed-case>CUNI</fixed-case> System for the <fixed-case>WMT</fixed-case>17 Multimodal Translation Task - Jindřich Helcl - Jindřich Libovický + JindřichHelcl + JindřichLibovický 450–457 W17-4749 10.18653/v1/W17-4749 Generating Image Descriptions using Multilingual Data - Alan Jaffe + AlanJaffe 458–464 W17-4750 10.18653/v1/W17-4750 <fixed-case>OSU</fixed-case> Multimodal Machine Translation System Report - Mingbo Ma - Dapeng Li - Kai Zhao - Liang Huang + MingboMa + DapengLi + KaiZhao + LiangHuang 465–469 W17-4751 10.18653/v1/W17-4751 <fixed-case>S</fixed-case>heffield <fixed-case>M</fixed-case>ulti<fixed-case>MT</fixed-case>: Using Object Posterior Predictions for Multimodal Machine Translation - Pranava Swaroop Madhyastha - Josiah Wang - Lucia Specia + Pranava SwaroopMadhyastha + JosiahWang + LuciaSpecia 470–476 W17-4752 10.18653/v1/W17-4752 <fixed-case>NICT</fixed-case>-<fixed-case>NAIST</fixed-case> System for <fixed-case>WMT</fixed-case>17 Multimodal Translation Task - Jingyi Zhang - Masao Utiyama - Eiichro Sumita - Graham Neubig - Satoshi Nakamura + JingyiZhang + MasaoUtiyama + EiichroSumita + GrahamNeubig + SatoshiNakamura 477–482 W17-4753 10.18653/v1/W17-4753 Automatic Threshold Detection for Data Selection in Machine Translation - Mirela-Stefania Duma - Wolfgang Menzel + Mirela-StefaniaDuma + WolfgangMenzel 483–488 W17-4754 10.18653/v1/W17-4754 Results of the <fixed-case>WMT</fixed-case>17 Metrics Shared Task - Ondřej Bojar - Yvette Graham - Amir Kamran + OndřejBojar + YvetteGraham + AmirKamran 489–513 W17-4755 10.18653/v1/W17-4755 A Shared Task on Bandit Learning for Machine Translation - Artem Sokolov - Julia Kreutzer - Kellen Sunderland - Pavel Danchenko - Witold Szymaniak - Hagen Fürstenau - Stefan Riezler + ArtemSokolov + JuliaKreutzer + KellenSunderland + PavelDanchenko + WitoldSzymaniak + HagenFürstenau + StefanRiezler 514–524 W17-4756 10.18653/v1/W17-4756 Results of the <fixed-case>WMT</fixed-case>17 Neural <fixed-case>MT</fixed-case> Training Task - Ondřej Bojar - Jindřich Helcl - Tom Kocmi - Jindřich Libovický - Tomáš Musil + OndřejBojar + JindřichHelcl + TomKocmi + JindřichLibovický + TomášMusil 525–533 W17-4757 10.18653/v1/W17-4757 Sentence-level quality estimation by predicting <fixed-case>HTER</fixed-case> as a multi-component metric - Eleftherios Avramidis + EleftheriosAvramidis 534–539 W17-4758 10.18653/v1/W17-4758 Predicting Translation Performance with Referential Translation Machines - Ergun Biçici + ErgunBiçici 540–544 W17-4759 10.18653/v1/W17-4759 Bilexical Embeddings for Quality Estimation - Frédéric Blain - Carolina Scarton - Lucia Specia + FrédéricBlain + CarolinaScarton + LuciaSpecia 545–550 W17-4760 10.18653/v1/W17-4760 Improving Machine Translation Quality Estimation with Neural Network Features - Zhiming Chen - Yiming Tan - Chenlin Zhang - Qingyu Xiang - Lilin Zhang - Maoxi Li - Mingwen Wang + ZhimingChen + YimingTan + ChenlinZhang + QingyuXiang + LilinZhang + MaoxiLi + MingwenWang 551–555 W17-4761 10.18653/v1/W17-4761 <fixed-case>UHH</fixed-case> Submission to the <fixed-case>WMT</fixed-case>17 Quality Estimation Shared Task - Melania Duma - Wolfgang Menzel + MelaniaDuma + WolfgangMenzel 556–561 W17-4762 10.18653/v1/W17-4762 Predictor-Estimator using Multilevel Task Learning with Stack Propagation for Neural Quality Estimation - Hyun Kim - Jong-Hyeok Lee - Seung-Hoon Na + HyunKim + Jong-HyeokLee + Seung-HoonNa 562–568 W17-4763 10.18653/v1/W17-4763 Unbabel’s Participation in the <fixed-case>WMT</fixed-case>17 Translation Quality Estimation Shared Task - André F. T. Martins - Fabio Kepler - José Monteiro + André F. T.Martins + FabioKepler + JoséMonteiro 569–574 W17-4764 10.18653/v1/W17-4764 Feature-Enriched Character-Level Convolutions for Text Regression - Gustavo Paetzold - Lucia Specia + GustavoPaetzold + LuciaSpecia 575–581 W17-4765 10.18653/v1/W17-4765 <fixed-case>UHH</fixed-case> Submission to the <fixed-case>WMT</fixed-case>17 Metrics Shared Task - Melania Duma - Wolfgang Menzel + MelaniaDuma + WolfgangMenzel 582–588 W17-4766 10.18653/v1/W17-4766 <fixed-case>MEANT</fixed-case> 2.0: Accurate semantic <fixed-case>MT</fixed-case> evaluation for any output language - Chi-kiu Lo + Chi-kiuLo 589–597 W17-4767 10.18653/v1/W17-4767 <fixed-case>B</fixed-case>lend: a Novel Combined <fixed-case>MT</fixed-case> Metric Based on Direct Assessment — <fixed-case>CASICT</fixed-case>-<fixed-case>DCU</fixed-case> submission to <fixed-case>WMT</fixed-case>17 Metrics Task - Qingsong Ma - Yvette Graham - Shugen Wang - Qun Liu + QingsongMa + YvetteGraham + ShugenWang + QunLiu 598–603 W17-4768 10.18653/v1/W17-4768 <fixed-case>CUNI</fixed-case> Experiments for <fixed-case>WMT</fixed-case>17 Metrics Task - David Mareček - Ondřej Bojar - Ondřej Hübsch - Rudolf Rosa - Dušan Variš + DavidMareček + OndřejBojar + OndřejHübsch + RudolfRosa + DušanVariš 604–611 W17-4769 10.18653/v1/W17-4769 chr<fixed-case>F</fixed-case>++: words helping character n-grams - Maja Popović + MajaPopović 612–618 W17-4770 10.18653/v1/W17-4770 bleu2vec: the Painfully Familiar Metric on Continuous Vector Space Steroids - Andre Tättar - Mark Fishel + AndreTättar + MarkFishel 619–622 W17-4771 10.18653/v1/W17-4771 <fixed-case>LIG</fixed-case>-<fixed-case>CRIS</fixed-case>t<fixed-case>AL</fixed-case> Submission for the <fixed-case>WMT</fixed-case> 2017 Automatic Post-Editing Task - Alexandre Bérard - Laurent Besacier - Olivier Pietquin + AlexandreBérard + LaurentBesacier + OlivierPietquin 623–629 W17-4772 10.18653/v1/W17-4772 Multi-source Neural Automatic Post-Editing: <fixed-case>FBK</fixed-case>’s participation in the <fixed-case>WMT</fixed-case> 2017 <fixed-case>APE</fixed-case> shared task - Rajen Chatterjee - M. Amin Farajian - Matteo Negri - Marco Turchi - Ankit Srivastava - Santanu Pal + RajenChatterjee + M. AminFarajian + MatteoNegri + MarcoTurchi + AnkitSrivastava + SantanuPal 630–638 W17-4773 10.18653/v1/W17-4773 The <fixed-case>AMU</fixed-case>-<fixed-case>UE</fixed-case>din Submission to the <fixed-case>WMT</fixed-case> 2017 Shared Task on Automatic Post-Editing - Marcin Junczys-Dowmunt - Marcin Junczys-Dowmunt + MarcinJunczys-Dowmunt + MarcinJunczys-Dowmunt 639–646 W17-4774 10.18653/v1/W17-4774 @@ -8321,63 +8321,63 @@ is able to handle phenomena related to scope by means of an higher-order type th Ensembling Factored Neural Machine Translation Models for Automatic Post-Editing and Quality Estimation - Chris Hokamp + ChrisHokamp 647–654 W17-4775 10.18653/v1/W17-4775 Neural Post-Editing Based on Quality Estimation - Yiming Tan - Zhiming Chen - Liu Huang - Lilin Zhang - Maoxi Li - Mingwen Wang + YimingTan + ZhimingChen + LiuHuang + LilinZhang + MaoxiLi + MingwenWang 655–660 W17-4776 10.18653/v1/W17-4776 <fixed-case>CUNI</fixed-case> System for <fixed-case>WMT</fixed-case>17 Automatic Post-Editing Task - Dušan Variš - Ondřej Bojar + DušanVariš + OndřejBojar 661–666 W17-4777 10.18653/v1/W17-4777 The <fixed-case>UMD</fixed-case> Neural Machine Translation Systems at <fixed-case>WMT</fixed-case>17 Bandit Learning Task - Amr Sharaf - Shi Feng - Khanh Nguyen - Kianté Brantley - Hal Daumé III + AmrSharaf + ShiFeng + KhanhNguyen + KiantéBrantley + HalDaumé III 667–673 W17-4778 10.18653/v1/W17-4778 <fixed-case>LIMSI</fixed-case> Submission for <fixed-case>WMT</fixed-case>’17 Shared Task on Bandit Learning - Guillaume Wisniewski + GuillaumeWisniewski 674–679 W17-4779 10.18653/v1/W17-4779 Variable Mini-Batch Sizing and Pre-Trained Embeddings - Mostafa Abdou - Vladan Glončák - Ondřej Bojar + MostafaAbdou + VladanGlončák + OndřejBojar 680–686 W17-4780 10.18653/v1/W17-4780 The <fixed-case>AFRL</fixed-case> <fixed-case>WMT</fixed-case>17 Neural Machine Translation Training Task Submission - Jeremy Gwinnup - Grant Erdmann - Katherine Young + JeremyGwinnup + GrantErdmann + KatherineYoung 687–691 W17-4781 10.18653/v1/W17-4781 @@ -8401,13 +8401,13 @@ is able to handle phenomena related to scope by means of an higher-order type th Findings of the 2017 <fixed-case>D</fixed-case>isco<fixed-case>MT</fixed-case> Shared Task on Cross-lingual Pronoun Prediction - Sharid Loáiciga - Sara Stymne - Preslav Nakov - Christian Hardmeier - Jörg Tiedemann - Mauro Cettolo - Yannick Versley + SharidLoáiciga + SaraStymne + PreslavNakov + ChristianHardmeier + JörgTiedemann + MauroCettolo + YannickVersley 1–16 W17-4801 10.18653/v1/W17-4801 @@ -8416,8 +8416,8 @@ is able to handle phenomena related to scope by means of an higher-order type th Validation of an Automatic Metric for the Accuracy of Pronoun Translation (<fixed-case>APT</fixed-case>) - Lesly Miculicich Werlen - Andrei Popescu-Belis + LeslyMiculicich Werlen + AndreiPopescu-Belis 17–25 W17-4802 10.18653/v1/W17-4802 @@ -8425,9 +8425,9 @@ is able to handle phenomena related to scope by means of an higher-order type th Using a Graph-based Coherence Model in Document-Level Machine Translation - Leo Born - Mohsen Mesgar - Michael Strube + LeoBorn + MohsenMesgar + MichaelStrube 26–35 W17-4803 10.18653/v1/W17-4803 @@ -8435,7 +8435,7 @@ is able to handle phenomena related to scope by means of an higher-order type th Treatment of Markup in Statistical Machine Translation - Mathias Müller + MathiasMüller 36–46 W17-4804 10.18653/v1/W17-4804 @@ -8443,9 +8443,9 @@ is able to handle phenomena related to scope by means of an higher-order type th A <fixed-case>B</fixed-case>i<fixed-case>LSTM</fixed-case>-based System for Cross-lingual Pronoun Prediction - Sara Stymne - Sharid Loáiciga - Fabienne Cap + SaraStymne + SharidLoáiciga + FabienneCap 47–53 W17-4805 10.18653/v1/W17-4805 @@ -8453,10 +8453,10 @@ is able to handle phenomena related to scope by means of an higher-order type th Neural Machine Translation for Cross-Lingual Pronoun Prediction - Sebastien Jean - Stanislas Lauly - Orhan Firat - Kyunghyun Cho + SebastienJean + StanislasLauly + OrhanFirat + KyunghyunCho 54–57 W17-4806 10.18653/v1/W17-4806 @@ -8464,7 +8464,7 @@ is able to handle phenomena related to scope by means of an higher-order type th Predicting Pronouns with a Convolutional Network and an N-gram Model - Christian Hardmeier + ChristianHardmeier 58–62 W17-4807 10.18653/v1/W17-4807 @@ -8472,9 +8472,9 @@ is able to handle phenomena related to scope by means of an higher-order type th Cross-Lingual Pronoun Prediction with Deep Recurrent Neural Networks v2.0 - Juhani Luotolahti - Jenna Kanerva - Filip Ginter + JuhaniLuotolahti + JennaKanerva + FilipGinter 63–66 W17-4808 10.18653/v1/W17-4808 @@ -8482,7 +8482,7 @@ is able to handle phenomena related to scope by means of an higher-order type th Combining the output of two coreference resolution systems for two source languages to improve annotation projection - Yulia Grishina + YuliaGrishina 67–72 W17-4809 10.18653/v1/W17-4809 @@ -8490,8 +8490,8 @@ is able to handle phenomena related to scope by means of an higher-order type th Discovery of Discourse-Related Language Contrasts through Alignment Discrepancies in <fixed-case>E</fixed-case>nglish-<fixed-case>G</fixed-case>erman Translation - Ekaterina Lapshinova-Koltunski - Christian Hardmeier + EkaterinaLapshinova-Koltunski + ChristianHardmeier 73–81 W17-4810 10.18653/v1/W17-4810 @@ -8499,8 +8499,8 @@ is able to handle phenomena related to scope by means of an higher-order type th Neural Machine Translation with Extended Context - Jörg Tiedemann - Yves Scherrer + JörgTiedemann + YvesScherrer 82–92 W17-4811 10.18653/v1/W17-4811 @@ -8508,9 +8508,9 @@ is able to handle phenomena related to scope by means of an higher-order type th Translating Implicit Discourse Connectives Based on Cross-lingual Annotation and Alignment - Hongzheng Li - Philippe Langlais - Yaohong Jin + HongzhengLi + PhilippeLanglais + YaohongJin 93–98 W17-4812 10.18653/v1/W17-4812 @@ -8518,7 +8518,7 @@ is able to handle phenomena related to scope by means of an higher-order type th Lexical Chains meet Word Embeddings in Document-level Statistical Machine Translation - Laura Mascarell + LauraMascarell 99–109 W17-4813 10.18653/v1/W17-4813 @@ -8526,7 +8526,7 @@ is able to handle phenomena related to scope by means of an higher-order type th On Integrating Discourse in Machine Translation - Karin Sim Smith + KarinSim Smith 110–121 W17-4814 10.18653/v1/W17-4814 @@ -8551,7 +8551,7 @@ is able to handle phenomena related to scope by means of an higher-order type th From Shakespeare to Twitter: What are Language Styles all about? - Wei Xu + WeiXu 1–9 W17-4901 10.18653/v1/W17-4901 @@ -8559,10 +8559,10 @@ is able to handle phenomena related to scope by means of an higher-order type th Shakespearizing Modern Language Using Copy-Enriched Sequence to Sequence Models - Harsh Jhamtani - Varun Gangal - Eduard Hovy - Eric Nyberg + HarshJhamtani + VarunGangal + EduardHovy + EricNyberg 10–19 W17-4902 10.18653/v1/W17-4902 @@ -8571,8 +8571,8 @@ is able to handle phenomena related to scope by means of an higher-order type th Discovering Stylistic Variations in Distributional Vector Space Models via Lexical Paraphrases - Xing Niu - Marine Carpuat + XingNiu + MarineCarpuat 20–27 W17-4903 10.18653/v1/W17-4903 @@ -8580,9 +8580,9 @@ is able to handle phenomena related to scope by means of an higher-order type th Harvesting Creative Templates for Generating Stylistically Varied Restaurant Reviews - Shereen Oraby - Sheideh Homayon - Marilyn Walker + ShereenOraby + SheidehHomayon + MarilynWalker 28–36 W17-4904 10.18653/v1/W17-4904 @@ -8590,8 +8590,8 @@ is able to handle phenomena related to scope by means of an higher-order type th Is writing style predictive of scientific fraud? - Chloé Braud - Anders Søgaard + ChloéBraud + AndersSøgaard 37–42 W17-4905 10.18653/v1/W17-4905 @@ -8599,9 +8599,9 @@ is able to handle phenomena related to scope by means of an higher-order type th “Deep” Learning : Detecting Metaphoricity in Adjective-Noun Pairs - Yuri Bizzoni - Stergios Chatzikyriakidis - Mehdi Ghanimifard + YuriBizzoni + StergiosChatzikyriakidis + MehdiGhanimifard 43–52 W17-4906 10.18653/v1/W17-4906 @@ -8609,9 +8609,9 @@ is able to handle phenomena related to scope by means of an higher-order type th Authorship Attribution with Convolutional Neural Networks and <fixed-case>POS</fixed-case>-Eliding - Julian Hitschler - Esther van den Berg - Ines Rehbein + JulianHitschler + Esthervan den Berg + InesRehbein 53–58 W17-4907 10.18653/v1/W17-4907 @@ -8619,9 +8619,9 @@ is able to handle phenomena related to scope by means of an higher-order type th Topic and audience effects on distinctively <fixed-case>S</fixed-case>cottish vocabulary usage in Twitter data - Philippa Shoemark - James Kirby - Sharon Goldwater + PhilippaShoemark + JamesKirby + SharonGoldwater 59–68 W17-4908 10.18653/v1/W17-4908 @@ -8629,10 +8629,10 @@ is able to handle phenomena related to scope by means of an higher-order type th Differences in type-token ratio and part-of-speech frequencies in male and female <fixed-case>R</fixed-case>ussian written texts - Tatiana Litvinova - Pavel Seredin - Olga Litvinova - Olga Zagorovskaya + TatianaLitvinova + PavelSeredin + OlgaLitvinova + OlgaZagorovskaya 69–73 W17-4909 10.18653/v1/W17-4909 @@ -8640,8 +8640,8 @@ is able to handle phenomena related to scope by means of an higher-order type th Modeling Communicative Purpose with Functional Style: Corpus and Features for <fixed-case>G</fixed-case>erman Genre and Register Analysis - Thomas Haider - Alexis Palmer + ThomasHaider + AlexisPalmer 74–84 W17-4910 10.18653/v1/W17-4910 @@ -8649,8 +8649,8 @@ is able to handle phenomena related to scope by means of an higher-order type th Stylistic Variation in Television Dialogue for Natural Language Generation - Grace Lin - Marilyn Walker + GraceLin + MarilynWalker 85–93 W17-4911 10.18653/v1/W17-4911 @@ -8658,8 +8658,8 @@ is able to handle phenomena related to scope by means of an higher-order type th Controlling Linguistic Style Aspects in Neural Language Generation - Jessica Ficler - Yoav Goldberg + JessicaFicler + YoavGoldberg 94–104 W17-4912 10.18653/v1/W17-4912 @@ -8667,8 +8667,8 @@ is able to handle phenomena related to scope by means of an higher-order type th Approximating Style by N-gram-based Annotation - Melanie Andresen - Heike Zinsmeister + MelanieAndresen + HeikeZinsmeister 105–115 W17-4913 10.18653/v1/W17-4913 @@ -8676,10 +8676,10 @@ is able to handle phenomena related to scope by means of an higher-order type th Assessing the Stylistic Properties of Neurally Generated Text in Authorship Attribution - Enrique Manjavacas - Jeroen De Gussem - Walter Daelemans - Mike Kestemont + EnriqueManjavacas + JeroenDe Gussem + WalterDaelemans + MikeKestemont 116–125 W17-4914 10.18653/v1/W17-4914 @@ -8705,7 +8705,7 @@ is able to handle phenomena related to scope by means of an higher-order type th Question Difficulty – How to Estimate Without Norming, How to Use for Automated Grading - Ulrike Padó + UlrikePadó 1–10 W17-5001 10.18653/v1/W17-5001 @@ -8713,11 +8713,11 @@ is able to handle phenomena related to scope by means of an higher-order type th Combining <fixed-case>CNN</fixed-case>s and Pattern Matching for Question Interpretation in a Virtual Patient Dialogue System - Lifeng Jin - Michael White - Evan Jaffe - Laura Zimmerman - Douglas Danforth + LifengJin + MichaelWhite + EvanJaffe + LauraZimmerman + DouglasDanforth 11–21 W17-5002 10.18653/v1/W17-5002 @@ -8725,10 +8725,10 @@ is able to handle phenomena related to scope by means of an higher-order type th Continuous fluency tracking and the challenges of varying text complexity - Beata Beigman Klebanov - Anastassia Loukina - John Sabatini - Tenaha O’Reilly + BeataBeigman Klebanov + AnastassiaLoukina + JohnSabatini + TenahaO’Reilly 22–32 W17-5003 10.18653/v1/W17-5003 @@ -8736,8 +8736,8 @@ is able to handle phenomena related to scope by means of an higher-order type th Auxiliary Objectives for Neural Error Detection Models - Marek Rei - Helen Yannakoudakis + MarekRei + HelenYannakoudakis 33–43 W17-5004 10.18653/v1/W17-5004 @@ -8745,10 +8745,10 @@ is able to handle phenomena related to scope by means of an higher-order type th Linked Data for Language-Learning Applications - Robyn Loughnane - Kate McCurdy - Peter Kolb - Stefan Selent + RobynLoughnane + KateMcCurdy + PeterKolb + StefanSelent 44–51 W17-5005 10.18653/v1/W17-5005 @@ -8756,8 +8756,8 @@ is able to handle phenomena related to scope by means of an higher-order type th Predicting Specificity in Classroom Discussion - Luca Lugini - Diane Litman + LucaLugini + DianeLitman 52–61 W17-5006 10.18653/v1/W17-5006 @@ -8765,14 +8765,14 @@ is able to handle phenomena related to scope by means of an higher-order type th A Report on the 2017 Native Language Identification Shared Task - Shervin Malmasi - Keelan Evanini - Aoife Cahill - Joel Tetreault - Robert Pugh - Christopher Hamill - Diane Napolitano - Yao Qian + ShervinMalmasi + KeelanEvanini + AoifeCahill + JoelTetreault + RobertPugh + ChristopherHamill + DianeNapolitano + YaoQian 62–75 W17-5007 10.18653/v1/W17-5007 @@ -8780,8 +8780,8 @@ is able to handle phenomena related to scope by means of an higher-order type th Evaluation of Automatically Generated Pronoun Reference Questions - Arief Yudha Satria - Takenobu Tokunaga + Arief YudhaSatria + TakenobuTokunaga 76–85 W17-5008 10.18653/v1/W17-5008 @@ -8789,8 +8789,8 @@ is able to handle phenomena related to scope by means of an higher-order type th Predicting Audience’s Laughter During Presentations Using Convolutional Neural Network - Lei Chen - Chong Min Lee + LeiChen + Chong MinLee 86–90 W17-5009 10.18653/v1/W17-5009 @@ -8798,9 +8798,9 @@ is able to handle phenomena related to scope by means of an higher-order type th Collecting fluency corrections for spoken learner <fixed-case>E</fixed-case>nglish - Andrew Caines - Emma Flint - Paula Buttery + AndrewCaines + EmmaFlint + PaulaButtery 91–100 W17-5010 10.18653/v1/W17-5010 @@ -8808,10 +8808,10 @@ is able to handle phenomena related to scope by means of an higher-order type th Exploring Relationships Between Writing & Broader Outcomes With Automated Writing Evaluation - Jill Burstein - Dan McCaffrey - Beata Beigman Klebanov - Guangming Ling + JillBurstein + DanMcCaffrey + BeataBeigman Klebanov + GuangmingLing 101–108 W17-5011 10.18653/v1/W17-5011 @@ -8819,10 +8819,10 @@ is able to handle phenomena related to scope by means of an higher-order type th An Investigation into the Pedagogical Features of Documents - Emily Sheng - Prem Natarajan - Jonathan Gordon - Gully Burns + EmilySheng + PremNatarajan + JonathanGordon + GullyBurns 109–120 W17-5012 10.18653/v1/W17-5012 @@ -8830,10 +8830,10 @@ is able to handle phenomena related to scope by means of an higher-order type th Combining Multiple Corpora for Readability Assessment for People with Cognitive Disabilities - Victoria Yaneva - Constantin Orăsan - Richard Evans - Omid Rohanian + VictoriaYaneva + ConstantinOrăsan + RichardEvans + OmidRohanian 121–132 W17-5013 10.18653/v1/W17-5013 @@ -8841,8 +8841,8 @@ is able to handle phenomena related to scope by means of an higher-order type th Automatic Extraction of High-Quality Example Sentences for Word Learning Using a Determinantal Point Process - Arseny Tolmachev - Sadao Kurohashi + ArsenyTolmachev + SadaoKurohashi 133–142 W17-5014 10.18653/v1/W17-5014 @@ -8851,8 +8851,8 @@ is able to handle phenomena related to scope by means of an higher-order type th Distractor Generation for <fixed-case>C</fixed-case>hinese Fill-in-the-blank Items - Shu Jiang - John Lee + ShuJiang + JohnLee 143–148 W17-5015 10.18653/v1/W17-5015 @@ -8860,9 +8860,9 @@ is able to handle phenomena related to scope by means of an higher-order type th An Error-Oriented Approach to Word Embedding Pre-Training - Youmna Farag - Marek Rei - Ted Briscoe + YoumnaFarag + MarekRei + TedBriscoe 149–158 W17-5016 10.18653/v1/W17-5016 @@ -8870,11 +8870,11 @@ is able to handle phenomena related to scope by means of an higher-order type th Investigating neural architectures for short answer scoring - Brian Riordan - Andrea Horbach - Aoife Cahill - Torsten Zesch - Chong Min Lee + BrianRiordan + AndreaHorbach + AoifeCahill + TorstenZesch + Chong MinLee 159–168 W17-5017 10.18653/v1/W17-5017 @@ -8882,10 +8882,10 @@ is able to handle phenomena related to scope by means of an higher-order type th Human and Automated <fixed-case>CEFR</fixed-case>-based Grading of Short Answers - Anaïs Tack - Thomas François - Sophie Roekhaut - Cédrick Fairon + AnaïsTack + ThomasFrançois + SophieRoekhaut + CédrickFairon 169–179 W17-5018 10.18653/v1/W17-5018 @@ -8893,9 +8893,9 @@ is able to handle phenomena related to scope by means of an higher-order type th <fixed-case>GEC</fixed-case> into the future: Where are we going and how do we get there? - Keisuke Sakaguchi - Courtney Napoles - Joel Tetreault + KeisukeSakaguchi + CourtneyNapoles + JoelTetreault 180–187 W17-5019 10.18653/v1/W17-5019 @@ -8903,7 +8903,7 @@ is able to handle phenomena related to scope by means of an higher-order type th Detecting Off-topic Responses to Visual Prompts - Marek Rei + MarekRei 188–197 W17-5020 10.18653/v1/W17-5020 @@ -8911,11 +8911,11 @@ is able to handle phenomena related to scope by means of an higher-order type th Combining Textual and Speech Features in the <fixed-case>NLI</fixed-case> Task Using State-of-the-Art Machine Learning Techniques - Pavel Ircing - Jan Švec - Zbyněk Zajíc - Barbora Hladká - Martin Holub + PavelIrcing + JanŠvec + ZbyněkZajíc + BarboraHladká + MartinHolub 198–209 W17-5021 10.18653/v1/W17-5021 @@ -8923,9 +8923,9 @@ is able to handle phenomena related to scope by means of an higher-order type th Native Language Identification Using a Mixture of Character and Word N-grams - Elham Mohammadi - Hadi Veisi - Hessam Amini + ElhamMohammadi + HadiVeisi + HessamAmini 210–216 W17-5022 10.18653/v1/W17-5022 @@ -8934,11 +8934,11 @@ is able to handle phenomena related to scope by means of an higher-order type th Ensemble Methods for Native Language Identification - Sophia Chan - Maryam Honari Jahromi - Benjamin Benetti - Aazim Lakhani - Alona Fyshe + SophiaChan + MaryamHonari Jahromi + BenjaminBenetti + AazimLakhani + AlonaFyshe 217–223 W17-5023 10.18653/v1/W17-5023 @@ -8946,8 +8946,8 @@ is able to handle phenomena related to scope by means of an higher-order type th Can string kernels pass the test of time in Native Language Identification? - Radu Tudor Ionescu - Marius Popescu + Radu TudorIonescu + MariusPopescu 224–234 W17-5024 10.18653/v1/W17-5024 @@ -8955,10 +8955,10 @@ is able to handle phenomena related to scope by means of an higher-order type th Neural Networks and Spelling Features for Native Language Identification - Johannes Bjerva - Gintarė Grigonytė - Robert Östling - Barbara Plank + JohannesBjerva + GintarėGrigonytė + RobertÖstling + BarbaraPlank 235–239 W17-5025 10.18653/v1/W17-5025 @@ -8966,8 +8966,8 @@ is able to handle phenomena related to scope by means of an higher-order type th A study of N-gram and Embedding Representations for Native Language Identification - Sowmya Vajjala - Sagnik Banerjee + SowmyaVajjala + SagnikBanerjee 240–248 W17-5026 10.18653/v1/W17-5026 @@ -8975,9 +8975,9 @@ is able to handle phenomena related to scope by means of an higher-order type th A Shallow Neural Network for Native Language Identification with Character N-grams - Yunita Sari - Muhammad Rifqi Fatchurrahman - Meisyarah Dwiastuti + YunitaSari + MuhammadRifqi Fatchurrahman + MeisyarahDwiastuti 249–254 W17-5027 10.18653/v1/W17-5027 @@ -8985,8 +8985,8 @@ is able to handle phenomena related to scope by means of an higher-order type th Fewer features perform well at Native Language Identification task - Taraka Rama - Çağrı Çöltekin + TarakaRama + ÇağrıÇöltekin 255–260 W17-5028 10.18653/v1/W17-5028 @@ -8994,10 +8994,10 @@ is able to handle phenomena related to scope by means of an higher-order type th Structured Generation of Technical Reading Lists - Jonathan Gordon - Stephen Aguilar - Emily Sheng - Gully Burns + JonathanGordon + StephenAguilar + EmilySheng + GullyBurns 261–270 W17-5029 10.18653/v1/W17-5029 @@ -9005,10 +9005,10 @@ is able to handle phenomena related to scope by means of an higher-order type th Effects of Lexical Properties on Viewing Time per Word in Autistic and Neurotypical Readers - Sanja Štajner - Victoria Yaneva - Ruslan Mitkov - Simone Paolo Ponzetto + SanjaŠtajner + VictoriaYaneva + RuslanMitkov + Simone PaoloPonzetto 271–281 W17-5030 10.18653/v1/W17-5030 @@ -9016,8 +9016,8 @@ is able to handle phenomena related to scope by means of an higher-order type th Transparent text quality assessment with convolutional neural networks - Robert Östling - Gintare Grigonyte + RobertÖstling + GintareGrigonyte 282–286 W17-5031 10.18653/v1/W17-5031 @@ -9025,10 +9025,10 @@ is able to handle phenomena related to scope by means of an higher-order type th Artificial Error Generation with Machine Translation and Syntactic Patterns - Marek Rei - Mariano Felice - Zheng Yuan - Ted Briscoe + MarekRei + MarianoFelice + ZhengYuan + TedBriscoe 287–292 W17-5032 10.18653/v1/W17-5032 @@ -9036,8 +9036,8 @@ is able to handle phenomena related to scope by means of an higher-order type th Modelling semantic acquisition in second language learning - Ekaterina Kochmar - Ekaterina Shutova + EkaterinaKochmar + EkaterinaShutova 293–302 W17-5033 10.18653/v1/W17-5033 @@ -9045,8 +9045,8 @@ is able to handle phenomena related to scope by means of an higher-order type th Multiple Choice Question Generation Utilizing An Ontology - Katherine Stasaski - Marti A. Hearst + KatherineStasaski + Marti A.Hearst 303–312 W17-5034 10.18653/v1/W17-5034 @@ -9055,8 +9055,8 @@ is able to handle phenomena related to scope by means of an higher-order type th Simplifying metaphorical language for young readers: A corpus study on news text - Magdalena Wolska - Yulia Clausen + MagdalenaWolska + YuliaClausen 313–318 W17-5035 10.18653/v1/W17-5035 @@ -9064,8 +9064,8 @@ is able to handle phenomena related to scope by means of an higher-order type th Language Based Mapping of Science Assessment Items to Skills - Farah Nadeem - Mari Ostendorf + FarahNadeem + MariOstendorf 319–326 W17-5036 10.18653/v1/W17-5036 @@ -9074,8 +9074,8 @@ is able to handle phenomena related to scope by means of an higher-order type th Connecting the Dots: Towards Human-Level Grammatical Error Correction - Shamil Chollampatt - Hwee Tou Ng + ShamilChollampatt + Hwee TouNg 327–333 W17-5037 10.18653/v1/W17-5037 @@ -9083,8 +9083,8 @@ is able to handle phenomena related to scope by means of an higher-order type th Question Generation for Language Learning: From ensuring texts are read to supporting learning - Maria Chinkina - Detmar Meurers + MariaChinkina + DetmarMeurers 334–344 W17-5038 10.18653/v1/W17-5038 @@ -9092,8 +9092,8 @@ is able to handle phenomena related to scope by means of an higher-order type th Systematically Adapting Machine Translation for Grammatical Error Correction - Courtney Napoles - Chris Callison-Burch + CourtneyNapoles + ChrisCallison-Burch 345–356 W17-5039 10.18653/v1/W17-5039 @@ -9101,10 +9101,10 @@ is able to handle phenomena related to scope by means of an higher-order type th Fine-grained essay scoring of a complex writing task for native speakers - Andrea Horbach - Dirk Scholten-Akoun - Yuning Ding - Torsten Zesch + AndreaHorbach + DirkScholten-Akoun + YuningDing + TorstenZesch 357–366 W17-5040 10.18653/v1/W17-5040 @@ -9112,8 +9112,8 @@ is able to handle phenomena related to scope by means of an higher-order type th Exploring Optimal Voting in Native Language Identification - Cyril Goutte - Serge Léger + CyrilGoutte + SergeLéger 367–373 W17-5041 10.18653/v1/W17-5041 @@ -9121,10 +9121,10 @@ is able to handle phenomena related to scope by means of an higher-order type th <fixed-case>CIC</fixed-case>-<fixed-case>FBK</fixed-case> Approach to Native Language Identification - Ilia Markov - Lingzhen Chen - Carlo Strapparava - Grigori Sidorov + IliaMarkov + LingzhenChen + CarloStrapparava + GrigoriSidorov 374–381 W17-5042 10.18653/v1/W17-5042 @@ -9132,13 +9132,13 @@ is able to handle phenomena related to scope by means of an higher-order type th The Power of Character N-grams in Native Language Identification - Artur Kulmizev - Bo Blankers - Johannes Bjerva - Malvina Nissim - Gertjan van Noord - Barbara Plank - Martijn Wieling + ArturKulmizev + BoBlankers + JohannesBjerva + MalvinaNissim + Gertjanvan Noord + BarbaraPlank + MartijnWieling 382–389 W17-5043 10.18653/v1/W17-5043 @@ -9146,8 +9146,8 @@ is able to handle phenomena related to scope by means of an higher-order type th Classifier Stacking for Native Language Identification - Wen Li - Liang Zou + WenLi + LiangZou 390–397 W17-5044 10.18653/v1/W17-5044 @@ -9155,9 +9155,9 @@ is able to handle phenomena related to scope by means of an higher-order type th Native Language Identification on Text and Speech - Marcos Zampieri - Alina Maria Ciobanu - Liviu P. Dinu + MarcosZampieri + Alina MariaCiobanu + Liviu P.Dinu 398–404 W17-5045 10.18653/v1/W17-5045 @@ -9165,8 +9165,8 @@ is able to handle phenomena related to scope by means of an higher-order type th Native Language Identification using Phonetic Algorithms - Charese Smiley - Sandra Kübler + ChareseSmiley + SandraKübler 405–412 W17-5046 10.18653/v1/W17-5046 @@ -9174,12 +9174,12 @@ is able to handle phenomena related to scope by means of an higher-order type th A deep-learning based native-language classification by using a latent semantic analysis for the <fixed-case>NLI</fixed-case> Shared Task 2017 - Yoo Rhee Oh - Hyung-Bae Jeon - Hwa Jeon Song - Yun-Kyung Lee - Jeon-Gue Park - Yun-Keun Lee + Yoo RheeOh + Hyung-BaeJeon + Hwa JeonSong + Yun-KyungLee + Jeon-GuePark + Yun-KeunLee 413–422 W17-5047 10.18653/v1/W17-5047 @@ -9187,9 +9187,9 @@ is able to handle phenomena related to scope by means of an higher-order type th Fusion of Simple Models for Native Language Identification - Fabio Kepler - Ramon Astudillo - Alberto Abad + FabioKepler + RamonAstudillo + AlbertoAbad 423–429 W17-5048 10.18653/v1/W17-5048 @@ -9197,8 +9197,8 @@ is able to handle phenomena related to scope by means of an higher-order type th Stacked Sentence-Document Classifier Approach for Improving Native Language Identification - Andrea Cimino - Felice Dell’Orletta + AndreaCimino + FeliceDell’Orletta 430–437 W17-5049 10.18653/v1/W17-5049 @@ -9206,8 +9206,8 @@ is able to handle phenomena related to scope by means of an higher-order type th Using Gaze to Predict Text Readability - Ana Valeria González-Garduño - Anders Søgaard + Ana ValeriaGonzález-Garduño + AndersSøgaard 438–443 W17-5050 10.18653/v1/W17-5050 @@ -9215,11 +9215,11 @@ is able to handle phenomena related to scope by means of an higher-order type th Annotating Orthographic Target Hypotheses in a <fixed-case>G</fixed-case>erman <fixed-case>L</fixed-case>1 Learner Corpus - Ronja Laarmann-Quante - Katrin Ortmann - Anna Ehlert - Maurice Vogel - Stefanie Dipper + RonjaLaarmann-Quante + KatrinOrtmann + AnnaEhlert + MauriceVogel + StefanieDipper 444–456 W17-5051 10.18653/v1/W17-5051 @@ -9227,9 +9227,9 @@ is able to handle phenomena related to scope by means of an higher-order type th A Large Scale Quantitative Exploration of Modeling Strategies for Content Scoring - Nitin Madnani - Anastassia Loukina - Aoife Cahill + NitinMadnani + AnastassiaLoukina + AoifeCahill 457–467 W17-5052 10.18653/v1/W17-5052 @@ -9261,10 +9261,10 @@ is able to handle phenomena related to scope by means of an higher-order type th 200<fixed-case>K</fixed-case>+ Crowdsourced Political Arguments for a New <fixed-case>C</fixed-case>hilean Constitution - Constanza Fierro - Claudio Fuentes - Jorge Pérez - Mauricio Quezada + ConstanzaFierro + ClaudioFuentes + JorgePérez + MauricioQuezada 1–10 W17-5101 10.18653/v1/W17-5101 @@ -9272,11 +9272,11 @@ is able to handle phenomena related to scope by means of an higher-order type th Analyzing the Semantic Types of Claims and Premises in an Online Persuasive Forum - Christopher Hidey - Elena Musi - Alyssa Hwang - Smaranda Muresan - Kathy McKeown + ChristopherHidey + ElenaMusi + AlyssaHwang + SmarandaMuresan + KathyMcKeown 11–21 W17-5102 10.18653/v1/W17-5102 @@ -9284,9 +9284,9 @@ is able to handle phenomena related to scope by means of an higher-order type th Annotation of argument structure in <fixed-case>J</fixed-case>apanese legal documents - Hiroaki Yamada - Simone Teufel - Takenobu Tokunaga + HiroakiYamada + SimoneTeufel + TakenobuTokunaga 22–31 W17-5103 10.18653/v1/W17-5103 @@ -9294,10 +9294,10 @@ is able to handle phenomena related to scope by means of an higher-order type th Improving Claim Stance Classification with Lexical Knowledge Expansion and Context Utilization - Roy Bar-Haim - Lilach Edelstein - Charles Jochim - Noam Slonim + RoyBar-Haim + LilachEdelstein + CharlesJochim + NoamSlonim 32–38 W17-5104 10.18653/v1/W17-5104 @@ -9305,8 +9305,8 @@ is able to handle phenomena related to scope by means of an higher-order type th Mining Argumentative Structure from Natural Language text using Automatically Generated Premise-Conclusion Topic Models - John Lawrence - Chris Reed + JohnLawrence + ChrisReed 39–48 W17-5105 10.18653/v1/W17-5105 @@ -9314,16 +9314,16 @@ is able to handle phenomena related to scope by means of an higher-order type th Building an Argument Search Engine for the Web - Henning Wachsmuth - Martin Potthast - Khalid Al-Khatib - Yamen Ajjour - Jana Puschmann - Jiani Qu - Jonas Dorsch - Viorel Morari - Janek Bevendorff - Benno Stein + HenningWachsmuth + MartinPotthast + KhalidAl-Khatib + YamenAjjour + JanaPuschmann + JianiQu + JonasDorsch + ViorelMorari + JanekBevendorff + BennoStein 49–59 W17-5106 10.18653/v1/W17-5106 @@ -9331,8 +9331,8 @@ is able to handle phenomena related to scope by means of an higher-order type th Argument Relation Classification Using a Joint Inference Model - Yufang Hou - Charles Jochim + YufangHou + CharlesJochim 60–66 W17-5107 10.18653/v1/W17-5107 @@ -9340,8 +9340,8 @@ is able to handle phenomena related to scope by means of an higher-order type th Projection of Argumentative Corpora from Source to Target Languages - Ahmet Aker - Huangpan Zhang + AhmetAker + HuangpanZhang 67–72 W17-5108 10.18653/v1/W17-5108 @@ -9350,7 +9350,7 @@ is able to handle phenomena related to scope by means of an higher-order type th Manual Identification of Arguments with Implicit Conclusions Using Semantic Rules for Argument Mining - Nancy Green + NancyGreen 73–78 W17-5109 10.18653/v1/W17-5109 @@ -9358,12 +9358,12 @@ is able to handle phenomena related to scope by means of an higher-order type th Unsupervised corpus–wide claim detection - Ran Levy - Shai Gretz - Benjamin Sznajder - Shay Hummel - Ranit Aharonov - Noam Slonim + RanLevy + ShaiGretz + BenjaminSznajder + ShayHummel + RanitAharonov + NoamSlonim 79–84 W17-5110 10.18653/v1/W17-5110 @@ -9372,7 +9372,7 @@ is able to handle phenomena related to scope by means of an higher-order type th Using Question-Answering Techniques to Implement a Knowledge-Driven Argument Mining Approach - Patrick Saint-Dizier + PatrickSaint-Dizier 85–90 W17-5111 10.18653/v1/W17-5111 @@ -9380,13 +9380,13 @@ is able to handle phenomena related to scope by means of an higher-order type th What works and what does not: Classifier and feature analysis for argument mining - Ahmet Aker - Alfred Sliwa - Yuan Ma - Ruishen Lui - Niravkumar Borad - Seyedeh Ziyaei - Mina Ghobadi + AhmetAker + AlfredSliwa + YuanMa + RuishenLui + NiravkumarBorad + SeyedehZiyaei + MinaGhobadi 91–96 W17-5112 10.18653/v1/W17-5112 @@ -9394,9 +9394,9 @@ is able to handle phenomena related to scope by means of an higher-order type th Unsupervised Detection of Argumentative Units though Topic Modeling Techniques - Alfio Ferrara - Stefano Montanelli - Georgios Petasis + AlfioFerrara + StefanoMontanelli + GeorgiosPetasis 97–107 W17-5113 10.18653/v1/W17-5113 @@ -9404,8 +9404,8 @@ is able to handle phenomena related to scope by means of an higher-order type th Using Complex Argumentative Interactions to Reconstruct the Argumentative Structure of Large-Scale Debates - John Lawrence - Chris Reed + JohnLawrence + ChrisReed 108–117 W17-5114 10.18653/v1/W17-5114 @@ -9413,11 +9413,11 @@ is able to handle phenomena related to scope by means of an higher-order type th Unit Segmentation of Argumentative Texts - Yamen Ajjour - Wei-Fan Chen - Johannes Kiesel - Henning Wachsmuth - Benno Stein + YamenAjjour + Wei-FanChen + JohannesKiesel + HenningWachsmuth + BennoStein 118–128 W17-5115 10.18653/v1/W17-5115 @@ -9442,7 +9442,7 @@ is able to handle phenomena related to scope by means of an higher-order type th Detecting Sarcasm Using Different Forms Of Incongruity - Aditya Joshi + AdityaJoshi 1 W17-5201 10.18653/v1/W17-5201 @@ -9450,9 +9450,9 @@ is able to handle phenomena related to scope by means of an higher-order type th Assessing State-of-the-Art Sentiment Models on State-of-the-Art Sentiment Datasets - Jeremy Barnes - Roman Klinger - Sabine Schulte im Walde + JeremyBarnes + RomanKlinger + SabineSchulte im Walde 2–12 W17-5202 10.18653/v1/W17-5202 @@ -9467,11 +9467,11 @@ is able to handle phenomena related to scope by means of an higher-order type th Annotation, Modelling and Analysis of Fine-Grained Emotions on a Stance and Sentiment Detection Corpus - Hendrik Schuff - Jeremy Barnes - Julian Mohme - Sebastian Padó - Roman Klinger + HendrikSchuff + JeremyBarnes + JulianMohme + SebastianPadó + RomanKlinger 13–23 W17-5203 10.18653/v1/W17-5203 @@ -9482,10 +9482,10 @@ with emotion annotation. We (a) analyse annotation reliability and annotation me Ranking Right-Wing Extremist Social Media Profiles by Similarity to Democratic and Extremist Groups - Matthias Hartung - Roman Klinger - Franziska Schmidtke - Lars Vogel + MatthiasHartung + RomanKlinger + FranziskaSchmidtke + LarsVogel 24–33 W17-5204 10.18653/v1/W17-5204 @@ -9493,8 +9493,8 @@ with emotion annotation. We (a) analyse annotation reliability and annotation me <fixed-case>WASSA</fixed-case>-2017 Shared Task on Emotion Intensity - Saif Mohammad - Felipe Bravo-Marquez + SaifMohammad + FelipeBravo-Marquez 34–49 W17-5205 10.18653/v1/W17-5205 @@ -9502,9 +9502,9 @@ with emotion annotation. We (a) analyse annotation reliability and annotation me <fixed-case>IMS</fixed-case> at <fixed-case>E</fixed-case>mo<fixed-case>I</fixed-case>nt-2017: Emotion Intensity Prediction with Affective Norms, Automatically Extended Resources and Deep Learning - Maximilian Köper - Evgeny Kim - Roman Klinger + MaximilianKöper + EvgenyKim + RomanKlinger 50–57 W17-5206 10.18653/v1/W17-5206 @@ -9513,10 +9513,10 @@ with emotion annotation. We (a) analyse annotation reliability and annotation me Prayas at <fixed-case>E</fixed-case>mo<fixed-case>I</fixed-case>nt 2017: An Ensemble of Deep Neural Architectures for Emotion Intensity Prediction in Tweets - Pranav Goel - Devang Kulshreshtha - Prayas Jain - Kaushal Kumar Shukla + PranavGoel + DevangKulshreshtha + PrayasJain + Kaushal KumarShukla 58–65 W17-5207 10.18653/v1/W17-5207 @@ -9524,7 +9524,7 @@ with emotion annotation. We (a) analyse annotation reliability and annotation me Latest News in Computational Argumentation: Surfing on the Deep Learning Wave, Scuba Diving in the Abyss of Fundamental Questions - Iryna Gurevych + IrynaGurevych 66 W17-5208 10.18653/v1/W17-5208 @@ -9532,10 +9532,10 @@ with emotion annotation. We (a) analyse annotation reliability and annotation me Towards Syntactic <fixed-case>I</fixed-case>berian Polarity Classification - David Vilares - Marcos Garcia - Miguel A. Alonso - Carlos Gómez-Rodríguez + DavidVilares + MarcosGarcia + Miguel A.Alonso + CarlosGómez-Rodríguez 67–73 W17-5209 10.18653/v1/W17-5209 @@ -9543,8 +9543,8 @@ with emotion annotation. We (a) analyse annotation reliability and annotation me Toward Stance Classification Based on Claim Microstructures - Filip Boltužić - Jan Šnajder + FilipBoltužić + JanŠnajder 74–80 W17-5210 10.18653/v1/W17-5210 @@ -9552,10 +9552,10 @@ with emotion annotation. We (a) analyse annotation reliability and annotation me Linguistic Reflexes of Well-Being and Happiness in Echo - Jiaqi Wu - Marilyn Walker - Pranav Anand - Steve Whittaker + JiaqiWu + MarilynWalker + PranavAnand + SteveWhittaker 81–91 W17-5211 10.18653/v1/W17-5211 @@ -9563,8 +9563,8 @@ with emotion annotation. We (a) analyse annotation reliability and annotation me Forecasting Consumer Spending from Purchase Intentions Expressed on Social Media - Viktor Pekar - Jane Binner + ViktorPekar + JaneBinner 92–101 W17-5212 10.18653/v1/W17-5212 @@ -9572,9 +9572,9 @@ with emotion annotation. We (a) analyse annotation reliability and annotation me Mining fine-grained opinions on closed captions of <fixed-case>Y</fixed-case>ou<fixed-case>T</fixed-case>ube videos with an attention-<fixed-case>RNN</fixed-case> - Edison Marrese-Taylor - Jorge Balazs - Yutaka Matsuo + EdisonMarrese-Taylor + JorgeBalazs + YutakaMatsuo 102–111 W17-5213 10.18653/v1/W17-5213 @@ -9582,7 +9582,7 @@ with emotion annotation. We (a) analyse annotation reliability and annotation me Understanding human values and their emotional effect - Alexandra Balahur + AlexandraBalahur 112 W17-5214 10.18653/v1/W17-5214 @@ -9590,8 +9590,8 @@ with emotion annotation. We (a) analyse annotation reliability and annotation me Did you ever read about Frogs drinking Coffee? Investigating the Compositionality of Multi-Emoji Expressions - Rebeca Padilla López - Fabienne Cap + RebecaPadilla López + FabienneCap 113–117 W17-5215 10.18653/v1/W17-5215 @@ -9599,8 +9599,8 @@ with emotion annotation. We (a) analyse annotation reliability and annotation me Investigating Redundancy in Emoji Use: Study on a Twitter Based Corpus - Giulia Donato - Patrizia Paggio + GiuliaDonato + PatriziaPaggio 118–126 W17-5216 10.18653/v1/W17-5216 @@ -9608,9 +9608,9 @@ with emotion annotation. We (a) analyse annotation reliability and annotation me Modeling Temporal Progression of Emotional Status in Mental Health Forum: A Recurrent Neural Net Approach - Kishaloy Halder - Lahari Poddar - Min-Yen Kan + KishaloyHalder + LahariPoddar + Min-YenKan 127–135 W17-5217 10.18653/v1/W17-5217 @@ -9618,11 +9618,11 @@ with emotion annotation. We (a) analyse annotation reliability and annotation me Towards an integrated pipeline for aspect-based sentiment analysis in various domains - Orphée De Clercq - Els Lefever - Gilles Jacobs - Tijl Carpels - Véronique Hoste + OrphéeDe Clercq + ElsLefever + GillesJacobs + TijlCarpels + VéroniqueHoste 136–142 W17-5218 10.18653/v1/W17-5218 @@ -9630,9 +9630,9 @@ with emotion annotation. We (a) analyse annotation reliability and annotation me Building a <fixed-case>S</fixed-case>enti<fixed-case>W</fixed-case>ord<fixed-case>N</fixed-case>et for <fixed-case>O</fixed-case>dia - Gaurav Mohanty - Abishek Kannan - Radhika Mamidi + GauravMohanty + AbishekKannan + RadhikaMamidi 143–148 W17-5219 10.18653/v1/W17-5219 @@ -9640,9 +9640,9 @@ with emotion annotation. We (a) analyse annotation reliability and annotation me Lexicon Integrated <fixed-case>CNN</fixed-case> Models with Attention for Sentiment Analysis - Bonggun Shin - Timothy Lee - Jinho D. Choi + BonggunShin + TimothyLee + Jinho D.Choi 149–158 W17-5220 10.18653/v1/W17-5220 @@ -9650,10 +9650,10 @@ with emotion annotation. We (a) analyse annotation reliability and annotation me Explaining Recurrent Neural Network Predictions in Sentiment Analysis - Leila Arras - Grégoire Montavon - Klaus-Robert Müller - Wojciech Samek + LeilaArras + GrégoireMontavon + Klaus-RobertMüller + WojciechSamek 159–168 W17-5221 10.18653/v1/W17-5221 @@ -9661,9 +9661,9 @@ with emotion annotation. We (a) analyse annotation reliability and annotation me <fixed-case>G</fixed-case>rad<fixed-case>A</fixed-case>scent at <fixed-case>E</fixed-case>mo<fixed-case>I</fixed-case>nt-2017: Character and Word Level Recurrent Neural Network Models for Tweet Emotion Intensity Detection - Egor Lakomkin - Chandrakant Bothe - Stefan Wermter + EgorLakomkin + ChandrakantBothe + StefanWermter 169–174 W17-5222 10.18653/v1/W17-5222 @@ -9671,9 +9671,9 @@ with emotion annotation. We (a) analyse annotation reliability and annotation me <fixed-case>NUIG</fixed-case> at <fixed-case>E</fixed-case>mo<fixed-case>I</fixed-case>nt-2017: <fixed-case>B</fixed-case>i<fixed-case>LSTM</fixed-case> and <fixed-case>SVR</fixed-case> Ensemble to Detect Emotion Intensity - Vladimir Andryushechkin - Ian Wood - James O’ Neill + VladimirAndryushechkin + IanWood + JamesO’ Neill 175–179 W17-5223 10.18653/v1/W17-5223 @@ -9681,10 +9681,10 @@ with emotion annotation. We (a) analyse annotation reliability and annotation me Unsupervised Aspect Term Extraction with B-<fixed-case>LSTM</fixed-case> & <fixed-case>CRF</fixed-case> using Automatically Labelled Datasets - Athanasios Giannakopoulos - Claudiu Musat - Andreea Hossmann - Michael Baeriswyl + AthanasiosGiannakopoulos + ClaudiuMusat + AndreeaHossmann + MichaelBaeriswyl 180–188 W17-5224 10.18653/v1/W17-5224 @@ -9692,8 +9692,8 @@ with emotion annotation. We (a) analyse annotation reliability and annotation me <fixed-case>PLN</fixed-case>-<fixed-case>PUCRS</fixed-case> at <fixed-case>E</fixed-case>mo<fixed-case>I</fixed-case>nt-2017: Psycholinguistic features for emotion intensity prediction in tweets - Henrique Santos - Renata Vieira + HenriqueSantos + RenataVieira 189–192 W17-5225 10.18653/v1/W17-5225 @@ -9701,10 +9701,10 @@ with emotion annotation. We (a) analyse annotation reliability and annotation me Textmining at <fixed-case>E</fixed-case>mo<fixed-case>I</fixed-case>nt-2017: A Deep Learning Approach to Sentiment Intensity Scoring of <fixed-case>E</fixed-case>nglish Tweets - Hardik Meisheri - Rupsa Saha - Priyanka Sinha - Lipika Dey + HardikMeisheri + RupsaSaha + PriyankaSinha + LipikaDey 193–199 W17-5226 10.18653/v1/W17-5226 @@ -9712,10 +9712,10 @@ with emotion annotation. We (a) analyse annotation reliability and annotation me <fixed-case>YNU</fixed-case>-<fixed-case>HPCC</fixed-case> at <fixed-case>E</fixed-case>mo<fixed-case>I</fixed-case>nt-2017: Using a <fixed-case>CNN</fixed-case>-<fixed-case>LSTM</fixed-case> Model for Sentiment Intensity Prediction - You Zhang - Hang Yuan - Jin Wang - Xuejie Zhang + YouZhang + HangYuan + JinWang + XuejieZhang 200–204 W17-5227 10.18653/v1/W17-5227 @@ -9723,8 +9723,8 @@ with emotion annotation. We (a) analyse annotation reliability and annotation me Seernet at <fixed-case>E</fixed-case>mo<fixed-case>I</fixed-case>nt-2017: Tweet Emotion Intensity Estimator - Venkatesh Duppada - Sushant Hiray + VenkateshDuppada + SushantHiray 205–211 W17-5228 10.18653/v1/W17-5228 @@ -9733,11 +9733,11 @@ with emotion annotation. We (a) analyse annotation reliability and annotation me <fixed-case>IITP</fixed-case> at <fixed-case>E</fixed-case>mo<fixed-case>I</fixed-case>nt-2017: Measuring Intensity of Emotions using Sentence Embeddings and Optimized Features - Md Shad Akhtar - Palaash Sawant - Asif Ekbal - Jyoti Pawar - Pushpak Bhattacharyya + Md ShadAkhtar + PalaashSawant + AsifEkbal + JyotiPawar + PushpakBhattacharyya 212–218 W17-5229 10.18653/v1/W17-5229 @@ -9745,8 +9745,8 @@ with emotion annotation. We (a) analyse annotation reliability and annotation me <fixed-case>NSE</fixed-case>mo at <fixed-case>E</fixed-case>mo<fixed-case>I</fixed-case>nt-2017: An Ensemble to Predict Emotion Intensity in Tweets - Sreekanth Madisetty - Maunendra Sankar Desarkar + SreekanthMadisetty + Maunendra SankarDesarkar 219–224 W17-5230 10.18653/v1/W17-5230 @@ -9754,7 +9754,7 @@ with emotion annotation. We (a) analyse annotation reliability and annotation me <fixed-case>T</fixed-case>ecnolengua <fixed-case>L</fixed-case>ingmotif at <fixed-case>E</fixed-case>mo<fixed-case>I</fixed-case>nt-2017: A lexicon-based approach - Antonio Moreno-Ortiz + AntonioMoreno-Ortiz 225–232 W17-5231 10.18653/v1/W17-5231 @@ -9762,8 +9762,8 @@ with emotion annotation. We (a) analyse annotation reliability and annotation me <fixed-case>E</fixed-case>mo<fixed-case>A</fixed-case>tt at <fixed-case>E</fixed-case>mo<fixed-case>I</fixed-case>nt-2017: Inner attention sentence embedding for Emotion Intensity - Edison Marrese-Taylor - Yutaka Matsuo + EdisonMarrese-Taylor + YutakaMatsuo 233–237 W17-5232 10.18653/v1/W17-5232 @@ -9771,10 +9771,10 @@ with emotion annotation. We (a) analyse annotation reliability and annotation me <fixed-case>YZU</fixed-case>-<fixed-case>NLP</fixed-case> at <fixed-case>E</fixed-case>mo<fixed-case>I</fixed-case>nt-2017: Determining Emotion Intensity Using a Bi-directional <fixed-case>LSTM</fixed-case>-<fixed-case>CNN</fixed-case> Model - Yuanye He - Liang-Chih Yu - K. Robert Lai - Weiyi Liu + YuanyeHe + Liang-ChihYu + K. RobertLai + WeiyiLiu 238–242 W17-5233 10.18653/v1/W17-5233 @@ -9782,8 +9782,8 @@ with emotion annotation. We (a) analyse annotation reliability and annotation me <fixed-case>DMG</fixed-case>roup at <fixed-case>E</fixed-case>mo<fixed-case>I</fixed-case>nt-2017: Emotion Intensity Using Ensemble Method - Song Jiang - Xiaotian Han + SongJiang + XiaotianHan 243–248 W17-5234 10.18653/v1/W17-5234 @@ -9791,8 +9791,8 @@ with emotion annotation. We (a) analyse annotation reliability and annotation me <fixed-case>UW</fixed-case>at-Emote at <fixed-case>E</fixed-case>mo<fixed-case>I</fixed-case>nt-2017: Emotion Intensity Detection using Affect Clues, Sentiment Polarity and Word Embeddings - Vineet John - Olga Vechtomova + VineetJohn + OlgaVechtomova 249–254 W17-5235 10.18653/v1/W17-5235 @@ -9800,8 +9800,8 @@ with emotion annotation. We (a) analyse annotation reliability and annotation me <fixed-case>LIPN</fixed-case>-<fixed-case>UAM</fixed-case> at <fixed-case>E</fixed-case>mo<fixed-case>I</fixed-case>nt-2017:Combination of Lexicon-based features and Sentence-level Vector Representations for Emotion Intensity Determination - Davide Buscaldi - Belem Priego + DavideBuscaldi + BelemPriego 255–258 W17-5236 10.18653/v1/W17-5236 @@ -9809,11 +9809,11 @@ with emotion annotation. We (a) analyse annotation reliability and annotation me deep<fixed-case>C</fixed-case>yb<fixed-case>E</fixed-case>r<fixed-case>N</fixed-case>et at <fixed-case>E</fixed-case>mo<fixed-case>I</fixed-case>nt-2017: Deep Emotion Intensities in Tweets - Vinayakumar R - Premjith B - Sachin Kumar S - Soman KP - Prabaharan Poornachandran + VinayakumarR + PremjithB + Sachin KumarS + SomanKP + PrabaharanPoornachandran 259–263 W17-5237 10.18653/v1/W17-5237 @@ -9842,10 +9842,10 @@ with emotion annotation. We (a) analyse annotation reliability and annotation me The <fixed-case>R</fixed-case>ep<fixed-case>E</fixed-case>val 2017 Shared Task: Multi-Genre Natural Language Inference with Sentence Representations - Nikita Nangia - Adina Williams - Angeliki Lazaridou - Samuel Bowman + NikitaNangia + AdinaWilliams + AngelikiLazaridou + SamuelBowman 1–10 W17-5301 10.18653/v1/W17-5301 @@ -9853,11 +9853,11 @@ with emotion annotation. We (a) analyse annotation reliability and annotation me Traversal-Free Word Vector Evaluation in Analogy Space - Xiaoyin Che - Nico Ring - Willi Raschkowski - Haojin Yang - Christoph Meinel + XiaoyinChe + NicoRing + WilliRaschkowski + HaojinYang + ChristophMeinel 11–15 W17-5302 10.18653/v1/W17-5302 @@ -9865,7 +9865,7 @@ with emotion annotation. We (a) analyse annotation reliability and annotation me Hypothesis Testing based Intrinsic Evaluation of Word Embeddings - Nishant Gurnani + NishantGurnani 16–20 W17-5303 10.18653/v1/W17-5303 @@ -9873,9 +9873,9 @@ with emotion annotation. We (a) analyse annotation reliability and annotation me Evaluation of word embeddings against cognitive processes: primed reaction times in lexical decision and naming tasks - Jeremy Auguste - Arnaud Rey - Benoit Favre + JeremyAuguste + ArnaudRey + BenoitFavre 21–26 W17-5304 10.18653/v1/W17-5304 @@ -9883,8 +9883,8 @@ with emotion annotation. We (a) analyse annotation reliability and annotation me Playing with Embeddings : Evaluating embeddings for Robot Language Learning through <fixed-case>MUD</fixed-case> Games - Anmol Gulati - Kumar Krishna Agrawal + AnmolGulati + Kumar KrishnaAgrawal 27–30 W17-5305 10.18653/v1/W17-5305 @@ -9892,7 +9892,7 @@ with emotion annotation. We (a) analyse annotation reliability and annotation me Recognizing Textual Entailment in Twitter Using Word Embeddings - Octavia-Maria Şulea + Octavia-MariaŞulea 31–35 W17-5306 10.18653/v1/W17-5306 @@ -9900,12 +9900,12 @@ with emotion annotation. We (a) analyse annotation reliability and annotation me Recurrent Neural Network-Based Sentence Encoder with Gated Attention for Natural Language Inference - Qian Chen - Xiaodan Zhu - Zhen-Hua Ling - Si Wei - Hui Jiang - Diana Inkpen + QianChen + XiaodanZhu + Zhen-HuaLing + SiWei + HuiJiang + DianaInkpen 36–40 W17-5307 10.18653/v1/W17-5307 @@ -9913,8 +9913,8 @@ with emotion annotation. We (a) analyse annotation reliability and annotation me Shortcut-Stacked Sentence Encoders for Multi-Domain Inference - Yixin Nie - Mohit Bansal + YixinNie + MohitBansal 41–45 W17-5308 10.18653/v1/W17-5308 @@ -9922,9 +9922,9 @@ with emotion annotation. We (a) analyse annotation reliability and annotation me Character-level Intra Attention Network for Natural Language Inference - Han Yang - Marta R. Costa-jussà - José A. R. Fonollosa + HanYang + Marta R.Costa-jussà + José A. R.Fonollosa 46–50 W17-5309 10.18653/v1/W17-5309 @@ -9932,10 +9932,10 @@ with emotion annotation. We (a) analyse annotation reliability and annotation me Refining Raw Sentence Representations for Textual Entailment Recognition via Attention - Jorge Balazs - Edison Marrese-Taylor - Pablo Loyola - Yutaka Matsuo + JorgeBalazs + EdisonMarrese-Taylor + PabloLoyola + YutakaMatsuo 51–55 W17-5310 10.18653/v1/W17-5310 @@ -9974,10 +9974,10 @@ with emotion annotation. We (a) analyse annotation reliability and annotation me Towards Linguistically Generalizable <fixed-case>NLP</fixed-case> Systems: A Workshop and Shared Task - Allyson Ettinger - Sudha Rao - Hal Daumé III - Emily M. Bender + AllysonEttinger + SudhaRao + HalDaumé III + Emily M.Bender 1–10 W17-5401 10.18653/v1/W17-5401 @@ -9985,11 +9985,11 @@ with emotion annotation. We (a) analyse annotation reliability and annotation me Analysing Errors of Open Information Extraction Systems - Rudolf Schneider - Tom Oberhauser - Tobias Klatt - Felix A. Gers - Alexander Löser + RudolfSchneider + TomOberhauser + TobiasKlatt + Felix A.Gers + AlexanderLöser 11–18 W17-5402 10.18653/v1/W17-5402 @@ -9997,9 +9997,9 @@ with emotion annotation. We (a) analyse annotation reliability and annotation me Massively Multilingual Neural Grapheme-to-Phoneme Conversion - Ben Peters - Jon Dehdari - Josef van Genabith + BenPeters + JonDehdari + Josefvan Genabith 19–26 W17-5403 10.18653/v1/W17-5403 @@ -10007,9 +10007,9 @@ with emotion annotation. We (a) analyse annotation reliability and annotation me <fixed-case>BIBI</fixed-case> System Description: Building with <fixed-case>CNN</fixed-case>s and Breaking with Deep Reinforcement Learning - Yitong Li - Trevor Cohn - Timothy Baldwin + YitongLi + TrevorCohn + TimothyBaldwin 27–32 W17-5404 10.18653/v1/W17-5404 @@ -10017,14 +10017,14 @@ with emotion annotation. We (a) analyse annotation reliability and annotation me Breaking <fixed-case>NLP</fixed-case>: Using Morphosyntax, Semantics, Pragmatics and World Knowledge to Fool Sentiment Analysis Systems - Taylor Mahler - Willy Cheung - Micha Elsner - David King - Marie-Catherine de Marneffe - Cory Shain - Symon Stevens-Guille - Michael White + TaylorMahler + WillyCheung + MichaElsner + DavidKing + Marie-Catherinede Marneffe + CoryShain + SymonStevens-Guille + MichaelWhite 33–39 W17-5405 10.18653/v1/W17-5405 @@ -10033,9 +10033,9 @@ with emotion annotation. We (a) analyse annotation reliability and annotation me An Adaptable Lexical Simplification Architecture for Major <fixed-case>I</fixed-case>bero-Romance Languages - Daniel Ferrés - Horacio Saggion - Xavier Gómez Guinovart + DanielFerrés + HoracioSaggion + XavierGómez Guinovart 40–47 W17-5406 10.18653/v1/W17-5406 @@ -10043,8 +10043,8 @@ with emotion annotation. We (a) analyse annotation reliability and annotation me Cross-genre Document Retrieval: Matching between Conversational and Formal Writings - Tomasz Jurczyk - Jinho D. Choi + TomaszJurczyk + Jinho D.Choi 48–53 W17-5407 10.18653/v1/W17-5407 @@ -10052,8 +10052,8 @@ with emotion annotation. We (a) analyse annotation reliability and annotation me <fixed-case>ACTSA</fixed-case>: Annotated Corpus for <fixed-case>T</fixed-case>elugu Sentiment Analysis - Sandeep Sricharan Mukku - Radhika Mamidi + Sandeep SricharanMukku + RadhikaMamidi 54–58 W17-5408 10.18653/v1/W17-5408 @@ -10061,7 +10061,7 @@ with emotion annotation. We (a) analyse annotation reliability and annotation me <fixed-case>S</fixed-case>trawman: An Ensemble of Deep Bag-of-Ngrams for Sentiment Analysis - Kyunghyun Cho + KyunghyunCho 59–60 W17-5409 10.18653/v1/W17-5409 @@ -10069,8 +10069,8 @@ with emotion annotation. We (a) analyse annotation reliability and annotation me Breaking Sentiment Analysis of Movie Reviews - Ieva Staliūnaitė - Ben Bonfil + IevaStaliūnaitė + BenBonfil 61–64 W17-5410 10.18653/v1/W17-5410 @@ -10096,8 +10096,8 @@ with emotion annotation. We (a) analyse annotation reliability and annotation me Automatic Mapping of <fixed-case>F</fixed-case>rench Discourse Connectives to <fixed-case>PDTB</fixed-case> Discourse Relations - Majid Laali - Leila Kosseim + MajidLaali + LeilaKosseim 1–6 W17-5501 10.18653/v1/W17-5501 @@ -10105,9 +10105,9 @@ with emotion annotation. We (a) analyse annotation reliability and annotation me Towards Full Text Shallow Discourse Relation Annotation: Experiments with Cross-Paragraph Implicit Relations in the <fixed-case>PDTB</fixed-case> - Rashmi Prasad - Katherine Forbes Riley - Alan Lee + RashmiPrasad + KatherineForbes Riley + AlanLee 7–16 W17-5502 10.18653/v1/W17-5502 @@ -10115,7 +10115,7 @@ with emotion annotation. We (a) analyse annotation reliability and annotation me User-initiated Sub-dialogues in State-of-the-art Dialogue Systems - Staffan Larsson + StaffanLarsson 17–22 W17-5503 10.18653/v1/W17-5503 @@ -10123,11 +10123,11 @@ with emotion annotation. We (a) analyse annotation reliability and annotation me A Multimodal Dialogue System for Medical Decision Support inside Virtual Reality - Alexander Prange - Margarita Chikobava - Peter Poller - Michael Barz - Daniel Sonntag + AlexanderPrange + MargaritaChikobava + PeterPoller + MichaelBarz + DanielSonntag 23–26 W17-5504 10.18653/v1/W17-5504 @@ -10135,10 +10135,10 @@ with emotion annotation. We (a) analyse annotation reliability and annotation me Generative Encoder-Decoder Models for Task-Oriented Spoken Dialog Systems with Chatting Capability - Tiancheng Zhao - Allen Lu - Kyusong Lee - Maxine Eskenazi + TianchengZhao + AllenLu + KyusongLee + MaxineEskenazi 27–36 W17-5505 10.18653/v1/W17-5505 @@ -10146,10 +10146,10 @@ with emotion annotation. We (a) analyse annotation reliability and annotation me Key-Value Retrieval Networks for Task-Oriented Dialogue - Mihail Eric - Lakshmi Krishnan - Francois Charette - Christopher D. Manning + MihailEric + LakshmiKrishnan + FrancoisCharette + Christopher D.Manning 37–49 W17-5506 10.18653/v1/W17-5506 @@ -10157,11 +10157,11 @@ with emotion annotation. We (a) analyse annotation reliability and annotation me Lexical Acquisition through Implicit Confirmations over Multiple Dialogues - Kohei Ono - Ryu Takeda - Eric Nichols - Mikio Nakano - Kazunori Komatani + KoheiOno + RyuTakeda + EricNichols + MikioNakano + KazunoriKomatani 50–59 W17-5507 10.18653/v1/W17-5507 @@ -10169,12 +10169,12 @@ with emotion annotation. We (a) analyse annotation reliability and annotation me Utterance Intent Classification of a Spoken Dialogue System with Efficiently Untied Recursive Autoencoders - Tsuneo Kato - Atsushi Nagai - Naoki Noda - Ryosuke Sumitomo - Jianming Wu - Seiichi Yamamoto + TsuneoKato + AtsushiNagai + NaokiNoda + RyosukeSumitomo + JianmingWu + SeiichiYamamoto 60–64 W17-5508 10.18653/v1/W17-5508 @@ -10182,15 +10182,15 @@ with emotion annotation. We (a) analyse annotation reliability and annotation me Reward-Balancing for Statistical Spoken Dialogue Systems using Multi-objective Reinforcement Learning - Stefan Ultes - Paweł Budzianowski - Iñigo Casanueva - Nikola Mrkšić - Lina M. Rojas-Barahona - Pei-Hao Su - Tsung-Hsien Wen - Milica Gašić - Steve Young + StefanUltes + PawełBudzianowski + IñigoCasanueva + NikolaMrkšić + Lina M.Rojas-Barahona + Pei-HaoSu + Tsung-HsienWen + MilicaGašić + SteveYoung 65–70 W17-5509 10.18653/v1/W17-5509 @@ -10198,9 +10198,9 @@ with emotion annotation. We (a) analyse annotation reliability and annotation me Automatic Measures to Characterise Verbal Alignment in Human-Agent Interaction - Guillaume Dubuisson Duplessis - Chloé Clavel - Frédéric Landragin + GuillaumeDubuisson Duplessis + ChloéClavel + FrédéricLandragin 71–81 W17-5510 10.18653/v1/W17-5510 @@ -10208,8 +10208,8 @@ with emotion annotation. We (a) analyse annotation reliability and annotation me Demonstration of interactive teaching for end-to-end dialog control with hybrid code networks - Jason D. Williams - Lars Liden + Jason D.Williams + LarsLiden 82–85 W17-5511 10.18653/v1/W17-5511 @@ -10217,14 +10217,14 @@ with emotion annotation. We (a) analyse annotation reliability and annotation me Sub-domain Modelling for Dialogue Management with Hierarchical Reinforcement Learning - Paweł Budzianowski - Stefan Ultes - Pei-Hao Su - Nikola Mrkšić - Tsung-Hsien Wen - Iñigo Casanueva - Lina M. Rojas-Barahona - Milica Gašić + PawełBudzianowski + StefanUltes + Pei-HaoSu + NikolaMrkšić + Tsung-HsienWen + IñigoCasanueva + Lina M.Rojas-Barahona + MilicaGašić 86–92 W17-5512 10.18653/v1/W17-5512 @@ -10232,9 +10232,9 @@ with emotion annotation. We (a) analyse annotation reliability and annotation me <fixed-case>MACA</fixed-case>: A Modular Architecture for Conversational Agents - Hoai Phuoc Truong - Prasanna Parthasarathi - Joelle Pineau + Hoai PhuocTruong + PrasannaParthasarathi + JoellePineau 93–102 W17-5513 10.18653/v1/W17-5513 @@ -10242,10 +10242,10 @@ with emotion annotation. We (a) analyse annotation reliability and annotation me Sequential Dialogue Context Modeling for Spoken Language Understanding - Ankur Bapna - Gokhan Tür - Dilek Hakkani-Tür - Larry Heck + AnkurBapna + GokhanTür + DilekHakkani-Tür + LarryHeck 103–114 W17-5514 10.18653/v1/W17-5514 @@ -10253,10 +10253,10 @@ with emotion annotation. We (a) analyse annotation reliability and annotation me Redundancy Localization for the Conversationalization of Unstructured Responses - Sebastian Krause - Mikhail Kozhevnikov - Eric Malmi - Daniele Pighin + SebastianKrause + MikhailKozhevnikov + EricMalmi + DanielePighin 115–126 W17-5515 10.18653/v1/W17-5515 @@ -10264,12 +10264,12 @@ with emotion annotation. We (a) analyse annotation reliability and annotation me Attentive listening system with backchanneling, response generation and flexible turn-taking - Divesh Lala - Pierrick Milhorat - Koji Inoue - Masanari Ishida - Katsuya Takanashi - Tatsuya Kawahara + DiveshLala + PierrickMilhorat + KojiInoue + MasanariIshida + KatsuyaTakanashi + TatsuyaKawahara 127–136 W17-5516 10.18653/v1/W17-5516 @@ -10277,8 +10277,8 @@ with emotion annotation. We (a) analyse annotation reliability and annotation me Natural Language Input for In-Car Spoken Dialog Systems: How Natural is Natural? - Patricia Braunger - Wolfgang Maier + PatriciaBraunger + WolfgangMaier 137–146 W17-5517 10.18653/v1/W17-5517 @@ -10286,11 +10286,11 @@ with emotion annotation. We (a) analyse annotation reliability and annotation me Sample-efficient Actor-Critic Reinforcement Learning with Supervised Data for Dialogue Management - Pei-Hao Su - Paweł Budzianowski - Stefan Ultes - Milica Gašić - Steve Young + Pei-HaoSu + PawełBudzianowski + StefanUltes + MilicaGašić + SteveYoung 147–157 W17-5518 10.18653/v1/W17-5518 @@ -10298,8 +10298,8 @@ with emotion annotation. We (a) analyse annotation reliability and annotation me A surprisingly effective out-of-the-box char2char model on the <fixed-case>E</fixed-case>2<fixed-case>E</fixed-case> <fixed-case>NLG</fixed-case> Challenge dataset - Shubham Agarwal - Marc Dymetman + ShubhamAgarwal + MarcDymetman 158–163 W17-5519 10.18653/v1/W17-5519 @@ -10307,9 +10307,9 @@ with emotion annotation. We (a) analyse annotation reliability and annotation me Interaction Quality Estimation Using Long Short-Term Memories - Niklas Rach - Wolfgang Minker - Stefan Ultes + NiklasRach + WolfgangMinker + StefanUltes 164–169 W17-5520 10.18653/v1/W17-5520 @@ -10317,18 +10317,18 @@ with emotion annotation. We (a) analyse annotation reliability and annotation me <fixed-case>D</fixed-case>ial<fixed-case>P</fixed-case>ort, Gone Live: An Update After A Year of Development - Kyusong Lee - Tiancheng Zhao - Yulun Du - Edward Cai - Allen Lu - Eli Pincus - David Traum - Stefan Ultes - Lina M. Rojas-Barahona - Milica Gasic - Steve Young - Maxine Eskenazi + KyusongLee + TianchengZhao + YulunDu + EdwardCai + AllenLu + EliPincus + DavidTraum + StefanUltes + Lina M.Rojas-Barahona + MilicaGasic + SteveYoung + MaxineEskenazi 170–173 W17-5521 10.18653/v1/W17-5521 @@ -10336,10 +10336,10 @@ with emotion annotation. We (a) analyse annotation reliability and annotation me Evaluating Natural Language Understanding Services for Conversational Question Answering Systems - Daniel Braun - Adrian Hernandez Mendez - Florian Matthes - Manfred Langen + DanielBraun + AdrianHernandez Mendez + FlorianMatthes + ManfredLangen 174–185 W17-5522 10.18653/v1/W17-5522 @@ -10347,9 +10347,9 @@ with emotion annotation. We (a) analyse annotation reliability and annotation me The Role of Conversation Context for Sarcasm Detection in Online Interactions - Debanjan Ghosh - Alexander Richard Fabbri - Smaranda Muresan + DebanjanGhosh + AlexanderRichard Fabbri + SmarandaMuresan 186–196 W17-5523 10.18653/v1/W17-5523 @@ -10357,9 +10357,9 @@ with emotion annotation. We (a) analyse annotation reliability and annotation me <fixed-case>VOILA</fixed-case>: An Optimised Dialogue System for Interactively Learning Visually-Grounded Word Meanings (Demonstration System) - Yanchao Yu - Arash Eshghi - Oliver Lemon + YanchaoYu + ArashEshghi + OliverLemon 197–200 W17-5524 10.18653/v1/W17-5524 @@ -10367,9 +10367,9 @@ with emotion annotation. We (a) analyse annotation reliability and annotation me The <fixed-case>E</fixed-case>2<fixed-case>E</fixed-case> Dataset: New Challenges For End-to-End Generation - Jekaterina Novikova - Ondřej Dušek - Verena Rieser + JekaterinaNovikova + OndřejDušek + VerenaRieser 201–206 W17-5525 10.18653/v1/W17-5525 @@ -10377,14 +10377,14 @@ with emotion annotation. We (a) analyse annotation reliability and annotation me <fixed-case>F</fixed-case>rames: a corpus for adding memory to goal-oriented dialogue systems - Layla El Asri - Hannes Schulz - Shikhar Sharma - Jeremie Zumer - Justin Harris - Emery Fine - Rahul Mehrotra - Kaheer Suleman + LaylaEl Asri + HannesSchulz + ShikharSharma + JeremieZumer + JustinHarris + EmeryFine + RahulMehrotra + KaheerSuleman 207–219 W17-5526 10.18653/v1/W17-5526 @@ -10393,7 +10393,7 @@ with emotion annotation. We (a) analyse annotation reliability and annotation me Towards a General, Continuous Model of Turn-taking in Spoken Dialogue using <fixed-case>LSTM</fixed-case> Recurrent Neural Networks - Gabriel Skantze + GabrielSkantze 220–230 W17-5527 10.18653/v1/W17-5527 @@ -10401,9 +10401,9 @@ with emotion annotation. We (a) analyse annotation reliability and annotation me Neural-based Natural Language Generation in Dialogue using <fixed-case>RNN</fixed-case> Encoder-Decoder with Semantic Aggregation - Van-Khanh Tran - Le-Minh Nguyen - Satoshi Tojo + Van-KhanhTran + Le-MinhNguyen + SatoshiTojo 231–240 W17-5528 10.18653/v1/W17-5528 @@ -10411,9 +10411,9 @@ with emotion annotation. We (a) analyse annotation reliability and annotation me Beyond On-hold Messages: Conversational Time-buying in Task-oriented Dialogue - Soledad López Gambino - Sina Zarrieß - David Schlangen + SoledadLópez Gambino + SinaZarrieß + DavidSchlangen 241–246 W17-5529 10.18653/v1/W17-5529 @@ -10421,8 +10421,8 @@ with emotion annotation. We (a) analyse annotation reliability and annotation me Neural-based Context Representation Learning for Dialog Act Classification - Daniel Ortega - Ngoc Thang Vu + DanielOrtega + Ngoc ThangVu 247–252 W17-5530 10.18653/v1/W17-5530 @@ -10430,9 +10430,9 @@ with emotion annotation. We (a) analyse annotation reliability and annotation me Predicting Success in Goal-Driven Human-Human Dialogues - Michael Noseworthy - Jackie Chi Kit Cheung - Joelle Pineau + MichaelNoseworthy + Jackie Chi KitCheung + JoellePineau 253–262 W17-5531 10.18653/v1/W17-5531 @@ -10440,10 +10440,10 @@ with emotion annotation. We (a) analyse annotation reliability and annotation me Generating and Evaluating Summaries for Partial Email Threads: Conversational <fixed-case>B</fixed-case>ayesian Surprise and Silver Standards - Jordon Johnson - Vaden Masrani - Giuseppe Carenini - Raymond Ng + JordonJohnson + VadenMasrani + GiuseppeCarenini + RaymondNg 263–272 W17-5532 10.18653/v1/W17-5532 @@ -10451,8 +10451,8 @@ with emotion annotation. We (a) analyse annotation reliability and annotation me Enabling robust and fluid spoken dialogue with cognitively impaired users - Ramin Yaghoubzadeh - Stefan Kopp + RaminYaghoubzadeh + StefanKopp 273–283 W17-5533 10.18653/v1/W17-5533 @@ -10460,8 +10460,8 @@ with emotion annotation. We (a) analyse annotation reliability and annotation me Adversarial evaluation for open-domain dialogue generation - Elia Bruni - Raquel Fernández + EliaBruni + RaquelFernández 284–288 W17-5534 10.18653/v1/W17-5534 @@ -10469,9 +10469,9 @@ with emotion annotation. We (a) analyse annotation reliability and annotation me Exploring Joint Neural Model for Sentence Level Discourse Parsing and Sentiment Analysis - Bita Nejat - Giuseppe Carenini - Raymond Ng + BitaNejat + GiuseppeCarenini + RaymondNg 289–298 W17-5535 10.18653/v1/W17-5535 @@ -10479,9 +10479,9 @@ with emotion annotation. We (a) analyse annotation reliability and annotation me Predicting Causes of Reformulation in Intelligent Assistants - Shumpei Sano - Nobuhiro Kaji - Manabu Sassano + ShumpeiSano + NobuhiroKaji + ManabuSassano 299–309 W17-5536 10.18653/v1/W17-5536 @@ -10489,11 +10489,11 @@ with emotion annotation. We (a) analyse annotation reliability and annotation me Are you serious?: Rhetorical Questions and Sarcasm in Social Media Dialog - Shereen Oraby - Vrindavan Harrison - Amita Misra - Ellen Riloff - Marilyn Walker + ShereenOraby + VrindavanHarrison + AmitaMisra + EllenRiloff + MarilynWalker 310–319 W17-5537 10.18653/v1/W17-5537 @@ -10502,10 +10502,10 @@ with emotion annotation. We (a) analyse annotation reliability and annotation me Finding Structure in Figurative Language: Metaphor Detection with Topic-based Frames - Hyeju Jang - Keith Maki - Eduard Hovy - Carolyn Rosé + HyejuJang + KeithMaki + EduardHovy + CarolynRosé 320–330 W17-5538 10.18653/v1/W17-5538 @@ -10513,9 +10513,9 @@ with emotion annotation. We (a) analyse annotation reliability and annotation me Using Reinforcement Learning to Model Incrementality in a Fast-Paced Dialogue Game - Ramesh Manuvinakurike - David DeVault - Kallirroi Georgila + RameshManuvinakurike + DavidDeVault + KallirroiGeorgila 331–341 W17-5539 10.18653/v1/W17-5539 @@ -10523,8 +10523,8 @@ with emotion annotation. We (a) analyse annotation reliability and annotation me Inferring Narrative Causality between Event Pairs in Films - Zhichao Hu - Marilyn Walker + ZhichaoHu + MarilynWalker 342–351 W17-5540 10.18653/v1/W17-5540 @@ -10533,8 +10533,8 @@ with emotion annotation. We (a) analyse annotation reliability and annotation me Lessons in Dialogue System Deployment - Anton Leuski - Ron Artstein + AntonLeuski + RonArtstein 352–355 W17-5541 10.18653/v1/W17-5541 @@ -10542,9 +10542,9 @@ with emotion annotation. We (a) analyse annotation reliability and annotation me Information Navigation System with Discovering User Interests - Koichiro Yoshino - Yu Suzuki - Satoshi Nakamura + KoichiroYoshino + YuSuzuki + SatoshiNakamura 356–359 W17-5542 10.18653/v1/W17-5542 @@ -10552,11 +10552,11 @@ with emotion annotation. We (a) analyse annotation reliability and annotation me Modelling Protagonist Goals and Desires in First-Person Narrative - Elahe Rahimtoroghi - Jiaqi Wu - Ruimin Wang - Pranav Anand - Marilyn Walker + ElaheRahimtoroghi + JiaqiWu + RuiminWang + PranavAnand + MarilynWalker 360–369 W17-5543 10.18653/v1/W17-5543 @@ -10565,14 +10565,14 @@ with emotion annotation. We (a) analyse annotation reliability and annotation me <fixed-case>SHIH</fixed-case>bot: A <fixed-case>F</fixed-case>acebook chatbot for Sexual Health Information on <fixed-case>HIV</fixed-case>/<fixed-case>AIDS</fixed-case> - Jacqueline Brixey - Rens Hoegen - Wei Lan - Joshua Rusow - Karan Singla - Xusen Yin - Ron Artstein - Anton Leuski + JacquelineBrixey + RensHoegen + WeiLan + JoshuaRusow + KaranSingla + XusenYin + RonArtstein + AntonLeuski 370–373 W17-5544 10.18653/v1/W17-5544 @@ -10580,12 +10580,12 @@ with emotion annotation. We (a) analyse annotation reliability and annotation me How Would You Say It? Eliciting Lexically Diverse Dialogue for Supervised Semantic Parsing - Abhilasha Ravichander - Thomas Manzini - Matthias Grabmair - Graham Neubig - Jonathan Francis - Eric Nyberg + AbhilashaRavichander + ThomasManzini + MatthiasGrabmair + GrahamNeubig + JonathanFrancis + EricNyberg 374–383 W17-5545 10.18653/v1/W17-5545 @@ -10593,8 +10593,8 @@ with emotion annotation. We (a) analyse annotation reliability and annotation me Not All Dialogues are Created Equal: Instance Weighting for Neural Conversational Models - Pierre Lison - Serge Bibauw + PierreLison + SergeBibauw 384–394 W17-5546 10.18653/v1/W17-5546 @@ -10602,7 +10602,7 @@ with emotion annotation. We (a) analyse annotation reliability and annotation me A data-driven model of explanations for a chatbot that helps to practice conversation in a foreign language - Sviatlana Höhn + SviatlanaHöhn 395–405 W17-5547 10.18653/v1/W17-5547 @@ -10625,29 +10625,29 @@ with emotion annotation. We (a) analyse annotation reliability and annotation me Building a Better Bitext for Structurally Different Languages through Self-training - Jungyeul Park - Loïc Dugast - Jeen-Pyo Hong - Chang-Uk Shin - Jeong-Won Cha + JungyeulPark + LoïcDugast + Jeen-PyoHong + Chang-UkShin + Jeong-WonCha 1–10 W17-5601 We propose a novel method to bootstrap the construction of parallel corpora for new pairs of structurally different languages. We do so by combining the use of a pivot language and self-training. A pivot language enables the use of existing translation models to bootstrap the alignment and a self-training procedure enables to achieve better alignment, both at the document and sentence level. We also propose several evaluation methods for the resulting alignment. <fixed-case>M</fixed-case>ulti<fixed-case>N</fixed-case>ews: A Web collection of an Aligned Multimodal and Multilingual Corpus - Haithem Afli - Pintu Lohar - Andy Way + HaithemAfli + PintuLohar + AndyWay 11–15 W17-5602 Integrating Natural Language Processing (NLP) and computer vision is a promising effort. However, the applicability of these methods directly depends on the availability of a specific multimodal data that includes images and texts. In this paper, we present a collection of a Multimodal corpus of comparable texts and their images in 9 languages from the web news articles of Euronews website. This corpus has found widespread use in the NLP community in Multilingual and multimodal tasks. Here, we focus on its acquisition of the images and text data and their multilingual alignment. Learning Phrase Embeddings from Paraphrases with <fixed-case>GRU</fixed-case>s - Zhihao Zhou - Lifu Huang - Heng Ji + ZhihaoZhou + LifuHuang + HengJi 16–23 W17-5603 Learning phrase representations has been widely explored in many Natural Language Processing tasks (e.g., Sentiment Analysis, Machine Translation) and has shown promising improvements. Previous studies either learn non-compositional phrase representations with general word embedding learning techniques or learn compositional phrase representations based on syntactic structures, which either require huge amounts of human annotations or cannot be easily generalized to all phrases. In this work, we propose to take advantage of large-scaled paraphrase database and present a pairwise-GRU framework to generate compositional phrase representations. Our framework can be re-used to generate representations for any phrases. Experimental results show that our framework achieves state-of-the-art results on several phrase similarity tasks. @@ -10669,24 +10669,24 @@ with emotion annotation. We (a) analyse annotation reliability and annotation me Overview of the 4th Workshop on <fixed-case>A</fixed-case>sian Translation - Toshiaki Nakazawa - Shohei Higashiyama - Chenchen Ding - Hideya Mino - Isao Goto - Hideto Kazawa - Yusuke Oda - Graham Neubig - Sadao Kurohashi + ToshiakiNakazawa + ShoheiHigashiyama + ChenchenDing + HideyaMino + IsaoGoto + HidetoKazawa + YusukeOda + GrahamNeubig + SadaoKurohashi 1–54 W17-5701 This paper presents the results of the shared tasks from the 4th workshop on Asian translation (WAT2017) including J↔E, J↔C scientific paper translation subtasks, C↔J, K↔J, E↔J patent translation subtasks, H↔E mixed domain subtasks, J↔E newswire subtasks and J↔E recipe subtasks. For the WAT2017, 12 institutions participated in the shared tasks. About 300 translation results have been submitted to the automatic evaluation server, and selected submissions were manually evaluated. Controlling Target Features in Neural Machine Translation via Prefix Constraints - Shunsuke Takeno - Masaaki Nagata - Kazuhide Yamamoto + ShunsukeTakeno + MasaakiNagata + KazuhideYamamoto 55–63 W17-5702 We propose prefix constraints, a novel method to enforce constraints on @@ -10705,141 +10705,141 @@ with emotion annotation. We (a) analyse annotation reliability and annotation me Improving <fixed-case>J</fixed-case>apanese-to-<fixed-case>E</fixed-case>nglish Neural Machine Translation by Paraphrasing the Target Language - Yuuki Sekizawa - Tomoyuki Kajiwara - Mamoru Komachi + YuukiSekizawa + TomoyukiKajiwara + MamoruKomachi 64–69 W17-5703 Neural machine translation (NMT) produces sentences that are more fluent than those produced by statistical machine translation (SMT). However, NMT has a very high computational cost because of the high dimensionality of the output layer. Generally, NMT restricts the size of vocabulary, which results in infrequent words being treated as out-of-vocabulary (OOV) and degrades the performance of the translation. In evaluation, we achieved a statistically significant BLEU score improvement of 0.55-0.77 over the baselines including the state-of-the-art method. Improving Low-Resource Neural Machine Translation with Filtered Pseudo-Parallel Corpus - Aizhan Imankulova - Takayuki Sato - Mamoru Komachi + AizhanImankulova + TakayukiSato + MamoruKomachi 70–78 W17-5704 Large-scale parallel corpora are indispensable to train highly accurate machine translators. However, manually constructed large-scale parallel corpora are not freely available in many language pairs. In previous studies, training data have been expanded using a pseudo-parallel corpus obtained using machine translation of the monolingual corpus in the target language. However, in low-resource language pairs in which only low-accuracy machine translation systems can be used, translation quality is reduces when a pseudo-parallel corpus is used naively. To improve machine translation performance with low-resource language pairs, we propose a method to expand the training data effectively via filtering the pseudo-parallel corpus using a quality estimation based on back-translation. As a result of experiments with three language pairs using small, medium, and large parallel corpora, language pairs with fewer training data filtered out more sentence pairs and improved BLEU scores more significantly. <fixed-case>J</fixed-case>apanese to <fixed-case>E</fixed-case>nglish/<fixed-case>C</fixed-case>hinese/<fixed-case>K</fixed-case>orean Datasets for Translation Quality Estimation and Automatic Post-Editing - Atsushi Fujita - Eiichiro Sumita + AtsushiFujita + EiichiroSumita 79–88 W17-5705 Aiming at facilitating the research on quality estimation (QE) and automatic post-editing (APE) of machine translation (MT) outputs, especially for those among Asian languages, we have created new datasets for Japanese to English, Chinese, and Korean translations. As the source text, actual utterances in Japanese were extracted from the log data of our speech translation service. MT outputs were then given by phrase-based statistical MT systems. Finally, human evaluators were employed to grade the quality of MT outputs and to post-edit them. This paper describes the characteristics of the created datasets and reports on our benchmarking experiments on word-level QE, sentence-level QE, and APE conducted using the created datasets. <fixed-case>NTT</fixed-case> Neural Machine Translation Systems at <fixed-case>WAT</fixed-case> 2017 - Makoto Morishita - Jun Suzuki - Masaaki Nagata + MakotoMorishita + JunSuzuki + MasaakiNagata 89–94 W17-5706 In this year, we participated in four translation subtasks at WAT 2017. Our model structure is quite simple but we used it with well-tuned hyper-parameters, leading to a significant improvement compared to the previous state-of-the-art system. We also tried to make use of the unreliable part of the provided parallel corpus by back-translating and making a synthetic corpus. Our submitted system achieved the new state-of-the-art performance in terms of the BLEU score, as well as human evaluation. <fixed-case>XMU</fixed-case> Neural Machine Translation Systems for <fixed-case>WAT</fixed-case> 2017 - Boli Wang - Zhixing Tan - Jinming Hu - Yidong Chen - Xiaodong Shi + BoliWang + ZhixingTan + JinmingHu + YidongChen + XiaodongShi 95–98 W17-5707 This paper describes the Neural Machine Translation systems of Xiamen University for the shared translation tasks of WAT 2017. Our systems are based on the Encoder-Decoder framework with attention. We participated in three subtasks. We experimented subword segmentation, synthetic training data and model ensembling. Experiments show that all these methods can give substantial improvements. A Bag of Useful Tricks for Practical Neural Machine Translation: Embedding Layer Initialization and Large Batch Size - Masato Neishi - Jin Sakuma - Satoshi Tohda - Shonosuke Ishiwatari - Naoki Yoshinaga - Masashi Toyoda + MasatoNeishi + JinSakuma + SatoshiTohda + ShonosukeIshiwatari + NaokiYoshinaga + MasashiToyoda 99–109 W17-5708 In this paper, we describe the team UT-IIS’s system and results for the WAT 2017 translation tasks. We further investigated several tricks including a novel technique for initializing embedding layers using only the parallel corpus, which increased the BLEU score by 1.28, found a practical large batch size of 256, and gained insights regarding hyperparameter settings. Ultimately, our system obtained a better result than the state-of-the-art system of WAT 2016. Our code is available on https://github.com/nem6ishi/wat17. Patent <fixed-case>NMT</fixed-case> integrated with Large Vocabulary Phrase Translation by <fixed-case>SMT</fixed-case> at <fixed-case>WAT</fixed-case> 2017 - Zi Long - Ryuichiro Kimura - Takehito Utsuro - Tomoharu Mitsuhashi - Mikio Yamamoto + ZiLong + RyuichiroKimura + TakehitoUtsuro + TomoharuMitsuhashi + MikioYamamoto 110–118 W17-5709 Neural machine translation (NMT) cannot handle a larger vocabulary because the training complexity and decoding complexity proportionally increase with the number of target words. This problem becomes even more serious when translating patent documents, which contain many technical terms that are observed infrequently. Long et al.(2017) proposed to select phrases that contain out-of-vocabulary words using the statistical approach of branching entropy. The selected phrases are then replaced with tokens during training and post-translated by the phrase translation table of SMT. In this paper, we apply the method proposed by Long et al. (2017) to the WAT 2017 Japanese-Chinese and Japanese-English patent datasets. Evaluation on Japanese-to-Chinese, Chinese-to-Japanese, Japanese-to-English and English-to-Japanese patent sentence translation proved the effectiveness of phrases selected with branching entropy, where the NMT model of Long et al.(2017) achieves a substantial improvement over a baseline NMT model without the technique proposed by Long et al.(2017). SMT reranked NMT - Terumasa Ehara + TerumasaEhara 119–126 W17-5710 System architecture, experimental settings and experimental results of the EHR team for the WAT2017 tasks are described. We participate in three tasks: JPCen-ja, JPCzh-ja and JPCko-ja. Although the basic architecture of our system is NMT, reranking technique is conducted using SMT results. One of the major drawback of NMT is under-translation and over-translation. On the other hand, SMT infrequently makes such translations. So, using reranking of n-best NMT outputs by the SMT output, discarding such translations can be expected. We can improve BLEU score from 46.03 to 47.08 by this technique in JPCzh-ja task. Ensemble and Reranking: Using Multiple Models in the <fixed-case>NICT</fixed-case>-2 Neural Machine Translation System at <fixed-case>WAT</fixed-case>2017 - Kenji Imamura - Eiichiro Sumita + KenjiImamura + EiichiroSumita 127–134 W17-5711 In this paper, we describe the NICT-2 neural machine translation system evaluated at WAT2017. This system uses multiple models as an ensemble and combines models with opposite decoding directions by reranking (called bi-directional reranking). In our experimental results on small data sets, the translation quality improved when the number of models was increased to 32 in total and did not saturate. In the experiments on large data sets, improvements of 1.59-3.32 BLEU points were achieved when six-model ensembles were combined by the bi-directional reranking. A Simple and Strong Baseline: <fixed-case>NAIST</fixed-case>-<fixed-case>NICT</fixed-case> Neural Machine Translation System for <fixed-case>WAT</fixed-case>2017 <fixed-case>E</fixed-case>nglish-<fixed-case>J</fixed-case>apanese Translation Task - Yusuke Oda - Katsuhito Sudoh - Satoshi Nakamura - Masao Utiyama - Eiichiro Sumita + YusukeOda + KatsuhitoSudoh + SatoshiNakamura + MasaoUtiyama + EiichiroSumita 135–139 W17-5712 This paper describes the details about the NAIST-NICT machine translation system for WAT2017 English-Japanese Scientific Paper Translation Task. The system consists of a language-independent tokenizer and an attentional encoder-decoder style neural machine translation model. According to the official results, our system achieves higher translation accuracy than any systems submitted previous campaigns despite simple model architecture. Comparison of <fixed-case>SMT</fixed-case> and <fixed-case>NMT</fixed-case> trained with large Patent Corpora: <fixed-case>J</fixed-case>apio at <fixed-case>WAT</fixed-case>2017 - Satoshi Kinoshita - Tadaaki Oshio - Tomoharu Mitsuhashi + SatoshiKinoshita + TadaakiOshio + TomoharuMitsuhashi 140–145 W17-5713 Japio participates in patent subtasks (JPC-EJ/JE/CJ/KJ) with phrase-based statistical machine translation (SMT) and neural machine translation (NMT) systems which are trained with its own patent corpora in addition to the subtask corpora provided by organizers of WAT2017. In EJ and CJ subtasks, SMT and NMT systems whose sizes of training corpora are about 50 million and 10 million sentence pairs respectively achieved comparable scores for automatic evaluations, but NMT systems were superior to SMT systems for both official and in-house human evaluations. <fixed-case>K</fixed-case>yoto University Participation to <fixed-case>WAT</fixed-case> 2017 - Fabien Cromieres - Raj Dabre - Toshiaki Nakazawa - Sadao Kurohashi + FabienCromieres + RajDabre + ToshiakiNakazawa + SadaoKurohashi 146–153 W17-5714 We describe here our approaches and results on the WAT 2017 shared translation tasks. Following our good results with Neural Machine Translation in the previous shared task, we continue this approach this year, with incremental improvements in models and training methods. We focused on the ASPEC dataset and could improve the state-of-the-art results for Chinese-to-Japanese and Japanese-to-Chinese translations. <fixed-case>CUNI</fixed-case> <fixed-case>NMT</fixed-case> System for <fixed-case>WAT</fixed-case> 2017 Translation Tasks - Tom Kocmi - Dušan Variš - Ondřej Bojar + TomKocmi + DušanVariš + OndřejBojar 154–159 W17-5715 The paper presents this year’s CUNI submissions to the WAT 2017 Translation Task focusing on the Japanese-English translation, namely Scientific papers subtask, Patents subtask and Newswire subtask. We compare two neural network architectures, the standard sequence-to-sequence with attention (Seq2Seq) and an architecture using convolutional sentence encoder (FBConv2Seq), both implemented in the NMT framework Neural Monkey that we currently participate in developing. We also compare various types of preprocessing of the source Japanese sentences and their impact on the overall results. Furthermore, we include the results of our experiments with out-of-domain data obtained by combining the corpora provided for each subtask. <fixed-case>T</fixed-case>okyo Metropolitan University Neural Machine Translation System for <fixed-case>WAT</fixed-case> 2017 - Yukio Matsumura - Mamoru Komachi + YukioMatsumura + MamoruKomachi 160–166 W17-5716 In this paper, we describe our neural machine translation (NMT) system, which is based on the attention-based NMT and uses long short-term memories (LSTM) as RNN. We implemented beam search and ensemble decoding in the NMT system. The system was tested on the 4th Workshop on Asian Translation (WAT 2017) shared tasks. In our experiments, we participated in the scientific paper subtasks and attempted Japanese-English, English-Japanese, and Japanese-Chinese translation tasks. The experimental results showed that implementation of beam search and ensemble decoding can effectively improve the translation quality. Comparing Recurrent and Convolutional Architectures for <fixed-case>E</fixed-case>nglish-<fixed-case>H</fixed-case>indi Neural Machine Translation - Sandhya Singh - Ritesh Panjwani - Anoop Kunchukuttan - Pushpak Bhattacharyya + SandhyaSingh + RiteshPanjwani + AnoopKunchukuttan + PushpakBhattacharyya 167–170 W17-5717 In this paper, we empirically compare the two encoder-decoder neural machine translation architectures: convolutional sequence to sequence model (ConvS2S) and recurrent sequence to sequence model (RNNS2S) for English-Hindi language pair as part of IIT Bombay’s submission to WAT2017 shared task. We report the results for both English-Hindi and Hindi-English direction of language pair. @@ -10862,100 +10862,100 @@ with emotion annotation. We (a) analyse annotation reliability and annotation me Automatic detection of stance towards vaccination in online discussion forums - Maria Skeppstedt - Andreas Kerren - Manfred Stede + MariaSkeppstedt + AndreasKerren + ManfredStede 1–8 W17-5801 A classifier for automatic detection of stance towards vaccination in online forums was trained and evaluated. Debate posts from six discussion threads on the British parental website Mumsnet were manually annotated for stance ‘against’ or ‘for’ vaccination, or as ‘undecided’. A support vector machine, trained to detect the three classes, achieved a macro F-score of 0.44, while a macro F-score of 0.62 was obtained by the same type of classifier on the binary classification task of distinguishing stance ‘against’ vaccination from stance ‘for’ vaccination. These results show that vaccine stance detection in online forums is a difficult task, at least for the type of model investigated and for the relatively small training corpus that was used. Future work will therefore include an expansion of the training data and an evaluation of other types of classifiers and features. Analysing the Causes of Depressed Mood from Depression Vulnerable Individuals - Noor Fazilla Abd Yusof - Chenghua Lin - Frank Guerin + Noor FazillaAbd Yusof + ChenghuaLin + FrankGuerin 9–17 W17-5802 We develop a computational model to discover the potential causes of depression by analysing the topics in a usergenerated text. We show the most prominent causes, and how these causes evolve over time. Also, we highlight the differences in causes between students with low and high neuroticism. Our studies demonstrate that the topics reveal valuable clues about the causes contributing to depressed mood. Identifying causes can have a significant impact on improving the quality of depression care; thereby providing greater insights into a patient’s state for pertinent treatment recommendations. Hence, this study significantly expands the ability to discover the potential factors that trigger depression, making it possible to increase the efficiency of depression treatment. Multivariate Linear Regression of Symptoms-related Tweets for Infectious Gastroenteritis Scale Estimation - Ryo Takeuchi - Hayate Iso - Kaoru Ito - Shoko Wakamiya - Eiji Aramaki + RyoTakeuchi + HayateIso + KaoruIto + ShokoWakamiya + EijiAramaki 18–25 W17-5803 To date, various Twitter-based event detection systems have been proposed. Most of their targets, however, share common characteristics. They are seasonal or global events such as earthquakes and flu pandemics. In contrast, this study targets unseasonal and local disease events. Our system investigates the frequencies of disease-related words such as “nausea”,“chill”,and “diarrhea” and estimates the number of patients using regression of these word frequencies. Experiments conducted using Japanese 47 areas from January 2017 to April 2017 revealed that the detection of small and unseasonal event is extremely difficult (overall performance: 0.13). However, we found that the event scale and the detection performance show high correlation in the specified cases (in the phase of patient increasing or decreasing). The results also suggest that when 150 and more patients appear in a high population area, we can expect that our social sensors detect this outbreak. Based on these results, we can infer that social sensors can reliably detect unseasonal and local disease events under certain conditions, just as they can for seasonal or global events. Incorporating Dependency Trees Improve Identification of Pregnant Women on Social Media Platforms - Yi-Jie Huang - Chu Hsien Su - Yi-Chun Chang - Tseng-Hsin Ting - Tzu-Yuan Fu - Rou-Min Wang - Hong-Jie Dai - Yung-Chun Chang - Jitendra Jonnagaddala - Wen-Lian Hsu + Yi-JieHuang + Chu HsienSu + Yi-ChunChang + Tseng-HsinTing + Tzu-YuanFu + Rou-MinWang + Hong-JieDai + Yung-ChunChang + JitendraJonnagaddala + Wen-LianHsu 26–32 W17-5804 The increasing popularity of social media lead users to share enormous information on the internet. This information has various application like, it can be used to develop models to understand or predict user behavior on social media platforms. For example, few online retailers have studied the shopping patterns to predict shopper’s pregnancy stage. Another interesting application is to use the social media platforms to analyze users’ health-related information. In this study, we developed a tree kernel-based model to classify tweets conveying pregnancy related information using this corpus. The developed pregnancy classification model achieved an accuracy of 0.847 and an F-score of 0.565. A new corpus from popular social media platform Twitter was developed for the purpose of this study. In future, we would like to improve this corpus by reducing noise such as retweets. Using a Recurrent Neural Network Model for Classification of Tweets Conveyed Influenza-related Information - Chen-Kai Wang - Onkar Singh - Zhao-Li Tang - Hong-Jie Dai + Chen-KaiWang + OnkarSingh + Zhao-LiTang + Hong-JieDai 33–38 W17-5805 Traditional disease surveillance systems depend on outpatient reporting and virological test results released by hospitals. These data have valid and accurate information about emerging outbreaks but it’s often not timely. In recent years the exponential growth of users getting connected to social media provides immense knowledge about epidemics by sharing related information. Social media can now flag more immediate concerns related to out-breaks in real time. In this paper we apply the long short-term memory recurrent neural net-work (RNN) architecture to classify tweets conveyed influenza-related information and compare its performance with baseline algorithms including support vector machine (SVM), decision tree, naive Bayes, simple logistics, and naive Bayes multinomial. The developed RNN model achieved an F-score of 0.845 on the MedWeb task test set, which outperforms the F-score of SVM without applying the synthetic minority oversampling technique by 0.08. The F-score of the RNN model is within 1% of the highest score achieved by SVM with oversampling technique. <fixed-case>Z</fixed-case>ika<fixed-case>H</fixed-case>ack 2016: A digital disease detection competition - Dillon C Adam - Jitendra Jonnagaddala - Daniel Han-Chen - Sean Batongbacal - Luan Almeida - Jing Z Zhu - Jenny J Yang - Jumail M Mundekkat - Steven Badman - Abrar Chughtai - C Raina MacIntyre + Dillon CAdam + JitendraJonnagaddala + DanielHan-Chen + SeanBatongbacal + LuanAlmeida + Jing ZZhu + Jenny JYang + Jumail MMundekkat + StevenBadman + AbrarChughtai + C RainaMacIntyre 39–46 W17-5806 Effective response to infectious diseases outbreaks relies on the rapid and early detection of those outbreaks. Invalidated, yet timely and openly available digital information can be used for the early detection of outbreaks. Public health surveillance authorities can exploit these early warnings to plan and co-ordinate rapid surveillance and emergency response programs. In 2016, a digital disease detection competition named ZikaHack was launched. The objective of the competition was for multidisciplinary teams to design, develop and demonstrate innovative digital disease detection solutions to retrospectively detect the 2015-16 Brazilian Zika virus outbreak earlier than traditional surveillance methods. In this paper, an overview of the ZikaHack competition is provided. The challenges and lessons learned in organizing this competition are also discussed for use by other researchers interested in organizing similar competitions. A Method to Generate a Machine-Labeled Data for Biomedical Named Entity Recognition with Various Sub-Domains - Juae Kim - Sunjae Kwon - Youngjoong Ko - Jungyun Seo + JuaeKim + SunjaeKwon + YoungjoongKo + JungyunSeo 47–51 W17-5807 Biomedical Named Entity (NE) recognition is a core technique for various works in the biomedical domain. In previous studies, using machine learning algorithm shows better performance than dictionary-based and rule-based approaches because there are too many terminological variations of biomedical NEs and new biomedical NEs are constantly generated. To achieve the high performance with a machine-learning algorithm, good-quality corpora are required. However, it is difficult to obtain the good-quality corpora because an-notating a biomedical corpus for ma-chine-learning is extremely time-consuming and costly. In addition, most previous corpora are insufficient for high-level tasks because they cannot cover various domains. Therefore, we propose a method for generating a large amount of machine-labeled data that covers various domains. To generate a large amount of machine-labeled data, firstly we generate an initial machine-labeled data by using a chunker and MetaMap. The chunker is developed to extract only biomedical NEs with manually annotated data. MetaMap is used to annotate the category of bio-medical NE. Then we apply the self-training approach to bootstrap the performance of initial machine-labeled data. In our experiments, the biomedical NE recognition system that is trained with our proposed machine-labeled data achieves much high performance. As a result, our system outperforms biomedical NE recognition system that using MetaMap only with 26.03%p improvements on F1-score. Enhancing Drug-Drug Interaction Classification with Corpus-level Feature and Classifier Ensemble - Jing Cyun Tu - Po-Ting Lai - Richard Tzong-Han Tsai + Jing CyunTu + Po-TingLai + Richard Tzong-HanTsai 52–56 W17-5808 The study of drug-drug interaction (DDI) is important in the drug discovering. Both PubMed and DrugBank are rich resources to retrieve DDI information which is usually represented in plain text. Automatically extracting DDI pairs from text improves the quality of drug discov-ering. In this paper, we presented a study that focuses on the DDI classification. We normalized the drug names, and developed both sentence-level and corpus-level features for DDI classification. A classifier ensemble approach is used for the unbalance DDI labels problem. Our approach achieved an F-score of 65.4% on SemEval 2013 DDI test set. The experimental results also show the effects of proposed corpus-level features in the DDI task. Chemical-Induced Disease Detection Using Invariance-based Pattern Learning Model - Neha Warikoo - Yung-Chun Chang - Wen-Lian Hsu + NehaWarikoo + Yung-ChunChang + Wen-LianHsu 57–64 W17-5809 In this work, we introduce a novel feature engineering approach named “algebraic invariance” to identify discriminative patterns for learning relation pair features for the chemical-disease relation (CDR) task of BioCreative V. Our method exploits the existing structural similarity of the key concepts of relation descriptions from the CDR corpus to generate robust linguistic patterns for SVM tree kernel-based learning. Preprocessing of the training data classifies the entity pairs as either related or unrelated to build instance types for both inter-sentential and intra-sentential scenarios. An invariant function is proposed to process and optimally cluster similar patterns for both positive and negative instances. The learning model for CDR pairs is based on the SVM tree kernel approach, which generates feature trees and vectors and is modeled on suitable invariance based patterns, bringing brevity, precision and context to the identifier features. Results demonstrate that our method outperformed other compared approaches, achieved a high recall rate of 85.08%, and averaged an F1-score of 54.34% without the use of any additional knowledge bases. @@ -10979,123 +10979,123 @@ with emotion annotation. We (a) analyse annotation reliability and annotation me <fixed-case>NTUCLE</fixed-case>: Developing a Corpus of Learner <fixed-case>E</fixed-case>nglish to Provide Writing Support for Engineering Students - Roger Vivek Placidus Winder - Joseph MacKinnon - Shu Yun Li - Benedict Christopher Tzer Liang Lin - Carmel Lee Hah Heah - Luís Morgado da Costa - Takayuki Kuribayashi - Francis Bond + Roger Vivek PlacidusWinder + JosephMacKinnon + Shu YunLi + Benedict Christopher Tzer LiangLin + Carmel Lee HahHeah + LuísMorgado da Costa + TakayukiKuribayashi + FrancisBond 1–11 W17-5901 This paper describes the creation of a new annotated learner corpus. The aim is to use this corpus to develop an automated system for corrective feedback on students’ writing. With this system, students will be able to receive timely feedback on language errors before they submit their assignments for grading. A corpus of assignments submitted by first year engineering students was compiled, and a new error tag set for the NTU Corpus of Learner English (NTUCLE) was developed based on that of the NUS Corpus of Learner English (NUCLE), as well as marking rubrics used at NTU. After a description of the corpus, error tag set and annotation process, the paper presents the results of the annotation exercise as well as follow up actions. The final error tag set, which is significantly larger than that for the NUCLE error categories, is then presented before a brief conclusion summarising our experience and future plans. Understanding Non-Native Writings: Can a Parser Help? - Jirka Hana - Barbora Hladká + JirkaHana + BarboraHladká 12–16 W17-5902 We present a pilot study on parsing non-native texts written by learners of Czech. We performed experiments that have shown that at least high-level syntactic functions, like subject, predicate, and object, can be assigned based on a parser trained on standard native language. Carrier Sentence Selection for Fill-in-the-blank Items - Shu Jiang - John Lee + ShuJiang + JohnLee 17–22 W17-5903 Fill-in-the-blank items are a common form of exercise in computer-assisted language learning systems. To automatically generate an effective item, the system must be able to select a high-quality carrier sentence that illustrates the usage of the target word. Previous approaches for carrier sentence selection have considered sentence length, vocabulary difficulty, the position of the target word and the presence of finite verbs. This paper investigates the utility of word co-occurrence statistics and lexical similarity as selection criteria. In an evaluation on generating fill-in-the-blank items for learning Chinese as a foreign language, we show that these two criteria can improve carrier sentence quality. <fixed-case>H</fixed-case>indi Shabdamitra: A <fixed-case>W</fixed-case>ordnet based E-Learning Tool for Language Learning and Teaching - Hanumant Redkar - Sandhya Singh - Meenakshi Somasundaram - Dhara Gorasia - Malhar Kulkarni - Pushpak Bhattacharyya + HanumantRedkar + SandhyaSingh + MeenakshiSomasundaram + DharaGorasia + MalharKulkarni + PushpakBhattacharyya 23–28 W17-5904 In today’s technology driven digital era, education domain is undergoing a transformation from traditional approaches to more learner controlled and flexible methods of learning. This transformation has opened the new avenues for interdisciplinary research in the field of educational technology and natural language processing in developing quality digital aids for learning and teaching. The tool presented here - Hindi Shabhadamitra, developed using Hindi Wordnet for Hindi language learning, is one such e-learning tool. It has been developed as a teaching and learning aid suitable for formal school based curriculum and informal setup for self learning users. Besides vocabulary, it also provides word based grammar along with images and pronunciation for better learning and retention. This aid demonstrates that how a rich lexical resource like wordnet can be systematically remodeled for practical usage in the educational domain. <fixed-case>NLPTEA</fixed-case> 2017 Shared Task – <fixed-case>C</fixed-case>hinese Spelling Check - Gabriel Fung - Maxime Debosschere - Dingmin Wang - Bo Li - Jia Zhu - Kam-Fai Wong + GabrielFung + MaximeDebosschere + DingminWang + BoLi + JiaZhu + Kam-FaiWong 29–34 W17-5905 This paper provides an overview along with our findings of the Chinese Spelling Check shared task at NLPTEA 2017. The goal of this task is to develop a computer-assisted system to automatically diagnose typing errors in traditional Chinese sentences written by students. We defined six types of errors which belong to two categories. Given a sentence, the system should detect where the errors are, and for each detected error determine its type and provide correction suggestions. We designed, constructed, and released a benchmark dataset for this task. <fixed-case>C</fixed-case>hinese Spelling Check based on N-gram and String Matching Algorithm - Jui-Feng Yeh - Li-Ting Chang - Chan-Yi Liu - Tsung-Wei Hsu + Jui-FengYeh + Li-TingChang + Chan-YiLiu + Tsung-WeiHsu 35–38 W17-5906 This paper presents a Chinese spelling check approach based on language models combined with string match algorithm to treat the problems resulted from the influence caused by Cantonese mother tone. N-grams first used to detecting the probability of sentence constructed by the writers, a string matching algorithm called Knuth-Morris-Pratt (KMP) Algorithm is used to detect and correct the error. According to the experimental results, the proposed approach can detect the error and provide the corresponding correction. N-gram Model for <fixed-case>C</fixed-case>hinese Grammatical Error Diagnosis - Jianbo Zhao - Hao Liu - Zuyi Bao - Xiaopeng Bai - Si Li - Zhiqing Lin + JianboZhao + HaoLiu + ZuyiBao + XiaopengBai + SiLi + ZhiqingLin 39–44 W17-5907 Detection and correction of Chinese grammatical errors have been two of major challenges for Chinese automatic grammatical error diagnosis.This paper presents an N-gram model for automatic detection and correction of Chinese grammatical errors in NLPTEA 2017 task. The experiment results show that the proposed method is good at correction of Chinese grammatical errors. The Influence of Spelling Errors on Content Scoring Performance - Andrea Horbach - Yuning Ding - Torsten Zesch + AndreaHorbach + YuningDing + TorstenZesch 45–53 W17-5908 Spelling errors occur frequently in educational settings, but their influence on automatic scoring is largely unknown. We therefore investigate the influence of spelling errors on content scoring performance using the example of the ASAP corpus. We conduct an annotation study on the nature of spelling errors in the ASAP dataset and utilize these finding in machine learning experiments that measure the influence of spelling errors on automatic content scoring. Our main finding is that scoring methods using both token and character n-gram features are robust against spelling errors up to the error frequency in ASAP. Analyzing the Impact of Spelling Errors on <fixed-case>POS</fixed-case>-Tagging and Chunking in Learner <fixed-case>E</fixed-case>nglish - Tomoya Mizumoto - Ryo Nagata + TomoyaMizumoto + RyoNagata 54–58 W17-5909 Part-of-speech (POS) tagging and chunking have been used in tasks targeting learner English; however, to the best our knowledge, few studies have evaluated their performance and no studies have revealed the causes of POS-tagging/chunking errors in detail. Therefore, we investigate performance and analyze the causes of failure. We focus on spelling errors that occur frequently in learner English. We demonstrate that spelling errors reduced POS-tagging performance by 0.23% owing to spelling errors, and that a spell checker is not necessary for POS-tagging/chunking of learner English. Complex Word Identification: Challenges in Data Annotation and System Performance - Marcos Zampieri - Shervin Malmasi - Gustavo Paetzold - Lucia Specia + MarcosZampieri + ShervinMalmasi + GustavoPaetzold + LuciaSpecia 59–63 W17-5910 This paper revisits the problem of complex word identification (CWI) following up the SemEval CWI shared task. We use ensemble classifiers to investigate how well computational methods can discriminate between complex and non-complex words. Furthermore, we analyze the classification performance to understand what makes lexical complexity challenging. Our findings show that most systems performed poorly on the SemEval CWI dataset, and one of the reasons for that is the way in which human annotation was performed. Suggesting Sentences for <fixed-case>ESL</fixed-case> using Kernel Embeddings - Kent Shioda - Mamoru Komachi - Rue Ikeya - Daichi Mochihashi + KentShioda + MamoruKomachi + RueIkeya + DaichiMochihashi 64–68 W17-5911 Sentence retrieval is an important NLP application for English as a Second Language (ESL) learners. ESL learners are familiar with web search engines, but generic web search results may not be adequate for composing documents in a specific domain. However, if we build our own search system specialized to a domain, it may be subject to the data sparseness problem. Recently proposed word2vec partially addresses the data sparseness problem, but fails to extract sentences relevant to queries owing to the modeling of the latent intent of the query. Thus, we propose a method of retrieving example sentences using kernel embeddings and N-gram windows. This method implicitly models latent intent of query and sentences, and alleviates the problem of noisy alignment. Our results show that our method achieved higher precision in sentence retrieval for ESL in the domain of a university press release corpus, as compared to a previous unsupervised method used for a semantic textual similarity task. Event Timeline Generation from History Textbooks - Harsimran Bedi - Sangameshwar Patil - Swapnil Hingmire - Girish Palshikar + HarsimranBedi + SangameshwarPatil + SwapnilHingmire + GirishPalshikar 69–77 W17-5912 Event timeline serves as the basic structure of history, and it is used as a disposition of key phenomena in studying history as a subject in secondary school. In order to enable a student to understand a historical phenomenon as a series of connected events, we present a system for automatic event timeline generation from history textbooks. Additionally, we propose Message Sequence Chart (MSC) and time-map based visualization techniques to visualize an event timeline. We also identify key computational challenges in developing natural language processing based applications for history textbooks. @@ -11117,52 +11117,52 @@ with emotion annotation. We (a) analyse annotation reliability and annotation me Group Linguistic Bias Aware Neural Response Generation - Jianan Wang - Xin Wang - Fang Li - Zhen Xu - Zhuoran Wang - Baoxun Wang + JiananWang + XinWang + FangLi + ZhenXu + ZhuoranWang + BaoxunWang 1–10 W17-6001 For practical chatbots, one of the essential factor for improving user experience is the capability of customizing the talking style of the agents, that is, to make chatbots provide responses meeting users’ preference on language styles, topics, etc. To address this issue, this paper proposes to incorporate linguistic biases, which implicitly involved in the conversation corpora generated by human groups in the Social Network Services (SNS), into the encoder-decoder based response generator. By attaching a specially designed neural component to dynamically control the impact of linguistic biases in response generation, a Group Linguistic Bias Aware Neural Response Generation (GLBA-NRG) model is eventually presented. The experimental results on the dataset from the Chinese SNS show that the proposed architecture outperforms the current response generating models by producing both meaningful and vivid responses with customized styles. Neural Regularized Domain Adaptation for <fixed-case>C</fixed-case>hinese Word Segmentation - Zuyi Bao - Si Li - Weiran Xu - Sheng Gao + ZuyiBao + SiLi + WeiranXu + ShengGao 11–20 W17-6002 For Chinese word segmentation, the large-scale annotated corpora mainly focus on newswire and only a handful of annotated data is available in other domains such as patents and literature. Considering the limited amount of annotated target domain data, it is a challenge for segmenters to learn domain-specific information while avoid getting over-fitted at the same time. In this paper, we propose a neural regularized domain adaptation method for Chinese word segmentation. The teacher networks trained in source domain are employed to regularize the training process of the student network by preserving the general knowledge. In the experiments, our neural regularized domain adaptation method achieves a better performance comparing to previous methods. The Sentimental Value of <fixed-case>C</fixed-case>hinese Sub-Character Components - Yassine Benajiba - Or Biran - Zhiliang Weng - Yong Zhang - Jin Sun + YassineBenajiba + OrBiran + ZhiliangWeng + YongZhang + JinSun 21–29 W17-6003 Sub-character components of Chinese characters carry important semantic information, and recent studies have shown that utilizing this information can improve performance on core semantic tasks. In this paper, we hypothesize that in addition to semantic information, sub-character components may also carry emotional information, and that utilizing it should improve performance on sentiment analysis tasks. We conduct a series of experiments on four Chinese sentiment data sets and show that we can significantly improve the performance in various tasks over that of a character-level embeddings baseline. We then focus on qualitatively assessing multiple examples and trying to explain how the sub-character components affect the results in each case. <fixed-case>C</fixed-case>hinese Answer Extraction Based on <fixed-case>POS</fixed-case> Tree and Genetic Algorithm - Shuihua Li - Xiaoming Zhang - Zhoujun Li + ShuihuaLi + XiaomingZhang + ZhoujunLi 30–36 W17-6004 Answer extraction is the most important part of a chinese web-based question answering system. In order to enhance the robustness and adaptability of answer extraction to new domains and eliminate the influence of the incomplete and noisy search snippets, we propose two new answer exraction methods. We utilize text patterns to generate Part-of-Speech (POS) patterns. In addition, a method is proposed to construct a POS tree by using these POS patterns. The POS tree is useful to candidate answer extraction of web-based question answering. To retrieve a efficient POS tree, the similarities between questions are used to select the question-answer pairs whose questions are similar to the unanswered question. Then, the POS tree is improved based on these question-answer pairs. In order to rank these candidate answers, the weights of the leaf nodes of the POS tree are calculated using a heuristic method. Moreover, the Genetic Algorithm (GA) is used to train the weights. The experimental results of 10-fold crossvalidation show that the weighted POS tree trained by GA can improve the accuracy of answer extraction. Learning from Parenthetical Sentences for Term Translation in Machine Translation - Guoping Huang - Jiajun Zhang - Yu Zhou - Chengqing Zong + GuopingHuang + JiajunZhang + YuZhou + ChengqingZong 37–45 W17-6005 Terms extensively exist in specific domains, and term translation plays a critical role in domain-specific machine translation (MT) tasks. However, it’s a challenging task to translate them correctly for the huge number of pre-existing terms and the endless new terms. To achieve better term translation quality, it is necessary to inject external term knowledge into the underlying MT system. Fortunately, there are plenty of term translation knowledge in parenthetical sentences on the Internet. In this paper, we propose a simple, straightforward and effective framework to improve term translation by learning from parenthetical sentences. This framework includes: (1) a focused web crawler; (2) a parenthetical sentence filter, acquiring parenthetical sentences including bilingual term pairs; (3) a term translation knowledge extractor, extracting bilingual term translation candidates; (4) a probability learner, generating the term translation table for MT decoders. The extensive experiments demonstrate that our proposed framework significantly improves the translation quality of terms and sentences. @@ -11184,108 +11184,108 @@ with emotion annotation. We (a) analyse annotation reliability and annotation me A Feature Structure Algebra for <fixed-case>FTAG</fixed-case> - Alexander Koller + AlexanderKoller 1–10 W17-6201 Parsing Minimalist Languages with Interpreted Regular Tree Grammars - Meaghan Fowlie - Alexander Koller + MeaghanFowlie + AlexanderKoller 11–20 W17-6202 Depictives in <fixed-case>E</fixed-case>nglish: An <fixed-case>LTAG</fixed-case> Approach - Benjamin Burkhardt - Timm Lichte - Laura Kallmeyer + BenjaminBurkhardt + TimmLichte + LauraKallmeyer 21–30 W17-6203 Reflexives and Reciprocals in Synchronous Tree Adjoining Grammar - Cristina Aggazzotti - Stuart M. Shieber + CristinaAggazzotti + Stuart M.Shieber 31–42 W17-6204 Coordination in <fixed-case>TAG</fixed-case> without the Conjoin Operation - Chung-hye Han - Anoop Sarkar + Chung-hyeHan + AnoopSarkar 43–52 W17-6205 Scope, Time, and Predicate Restriction in Blackfoot using <fixed-case>MC</fixed-case>-<fixed-case>STAG</fixed-case> - Dennis Ryan Storoshenko + Dennis RyanStoroshenko 53–60 W17-6206 Combining Predicate-Argument Structure and Operator Projection: Clause Structure in Role and Reference Grammar - Laura Kallmeyer - Rainer Osswald + LauraKallmeyer + RainerOsswald 61–70 W17-6207 Parsing with Dynamic Continuized <fixed-case>CCG</fixed-case> - Michael White - Simon Charlow - Jordan Needle - Dylan Bumford + MichaelWhite + SimonCharlow + JordanNeedle + DylanBumford 71–83 W17-6208 Multiword Expression-Aware <fixed-case>A</fixed-case>* <fixed-case>TAG</fixed-case> Parsing Revisited - Jakub Waszczuk - Agata Savary - Yannick Parmentier + JakubWaszczuk + AgataSavary + YannickParmentier 84–93 W17-6209 Single-Rooted <fixed-case>DAG</fixed-case>s in Regular <fixed-case>DAG</fixed-case> Languages: <fixed-case>P</fixed-case>arikh Image and Path Languages - Martin Berglund - Henrik Björklund - Frank Drewes + MartinBerglund + HenrikBjörklund + FrankDrewes 94–101 W17-6210 Contextual Hyperedge Replacement Grammars for Abstract Meaning Representations - Frank Drewes - Anna Jonsson + FrankDrewes + AnnaJonsson 102–111 W17-6211 Transforming Dependency Structures to <fixed-case>LTAG</fixed-case> Derivation Trees - Caio Corro - Joseph Le Roux + CaioCorro + JosephLe Roux 112–121 W17-6212 Linguistically Rich Vector Representations of Supertags for <fixed-case>TAG</fixed-case> Parsing - Dan Friedman - Jungo Kasai - R. Thomas McCoy - Robert Frank - Forrest Davis - Owen Rambow + DanFriedman + JungoKasai + R. ThomasMcCoy + RobertFrank + ForrestDavis + OwenRambow 122–131 W17-6213 <fixed-case>TAG</fixed-case> Parser Evaluation using Textual Entailments - Pauli Xu - Robert Frank - Jungo Kasai - Owen Rambow + PauliXu + RobertFrank + JungoKasai + OwenRambow 132–141 W17-6214 @@ -11306,121 +11306,121 @@ with emotion annotation. We (a) analyse annotation reliability and annotation me Automatically Acquired Lexical Knowledge Improves <fixed-case>J</fixed-case>apanese Joint Morphological and Dependency Analysis - Daisuke Kawahara - Yuta Hayashibe - Hajime Morita - Sadao Kurohashi + DaisukeKawahara + YutaHayashibe + HajimeMorita + SadaoKurohashi 1–10 W17-6301 This paper presents a joint model for morphological and dependency analysis based on automatically acquired lexical knowledge. This model takes advantage of rich lexical knowledge to simultaneously resolve word segmentation, POS, and dependency ambiguities. In our experiments on Japanese, we show the effectiveness of our joint model over conventional pipeline models. Dependency Language Models for Transition-based Dependency Parsing - Juntao Yu - Bernd Bohnet + JuntaoYu + BerndBohnet 11–17 W17-6302 In this paper, we present an approach to improve the accuracy of a strong transition-based dependency parser by exploiting dependency language models that are extracted from a large parsed corpus. We integrated a small number of features based on the dependency language models into the parser. To demonstrate the effectiveness of the proposed approach, we evaluate our parser on standard English and Chinese data where the base parser could achieve competitive accuracy scores. Our enhanced parser achieved state-of-the-art accuracy on Chinese data and competitive results on English data. We gained a large absolute improvement of one point (UAS) on Chinese and 0.5 points for English. Lexicalized vs. Delexicalized Parsing in Low-Resource Scenarios - Agnieszka Falenska - Özlem Çetinoğlu + AgnieszkaFalenska + ÖzlemÇetinoğlu 18–24 W17-6303 We present a systematic analysis of lexicalized vs. delexicalized parsing in low-resource scenarios, and propose a methodology to choose one method over another under certain conditions. We create a set of simulation experiments on 41 languages and apply our findings to 9 low-resource languages. Experimental results show that our methodology chooses the best approach in 8 out of 9 cases. Improving neural tagging with lexical information - Benoît Sagot - Héctor Martínez Alonso + BenoîtSagot + HéctorMartínez Alonso 25–31 W17-6304 Neural part-of-speech tagging has achieved competitive results with the incorporation of character-based and pre-trained word embeddings. In this paper, we show that a state-of-the-art bi-LSTM tagger can benefit from using information from morphosyntactic lexicons as additional input. The tagger, trained on several dozen languages, shows a consistent, average improvement when using lexical information, even when also using character-based embeddings, thus showing the complementarity of the different sources of lexical information. The improvements are particularly important for the smaller datasets. Prepositional Phrase Attachment over Word Embedding Products - Pranava Swaroop Madhyastha - Xavier Carreras - Ariadna Quattoni + Pranava SwaroopMadhyastha + XavierCarreras + AriadnaQuattoni 32–43 W17-6305 We present a low-rank multi-linear model for the task of solving prepositional phrase attachment ambiguity (PP task). Our model exploits tensor products of word embeddings, capturing all possible conjunctions of latent embeddings. Our results on a wide range of datasets and task settings show that tensor products are the best compositional operation and that a relatively simple multi-linear model that uses only word embeddings of lexical features can outperform more complex non-linear architectures that exploit the same information. Our proposed model gives the current best reported performance on an out-of-domain evaluation and performs competively on out-of-domain dependency parsing datasets. <fixed-case>L</fixed-case>1-<fixed-case>L</fixed-case>2 Parallel Dependency Treebank as Learner Corpus - John Lee - Keying Li - Herman Leung + JohnLee + KeyingLi + HermanLeung 44–49 W17-6306 This opinion paper proposes the use of parallel treebank as learner corpus. We show how an L1-L2 parallel treebank — i.e., parse trees of non-native sentences, aligned to the parse trees of their target hypotheses — can facilitate retrieval of sentences with specific learner errors. We argue for its benefits, in terms of corpus re-use and interoperability, over a conventional learner corpus annotated with error tags. As a proof of concept, we conduct a case study on word-order errors made by learners of Chinese as a foreign language. We report precision and recall in retrieving a range of word-order error categories from L1-L2 tree pairs annotated in the Universal Dependency framework. Splitting Complex <fixed-case>E</fixed-case>nglish Sentences - John Lee - J. Buddhika K. Pathirage Don + JohnLee + J. Buddhika K. PathirageDon 50–55 W17-6307 This paper applies parsing technology to the task of syntactic simplification of English sentences, focusing on the identification of text spans that can be removed from a complex sentence. We report the most comprehensive evaluation to-date on this task, using a dataset of sentences that exhibit simplification based on coordination, subordination, punctuation/parataxis, adjectival clauses, participial phrases, and appositive phrases. We train a decision tree with features derived from text span length, POS tags and dependency relations, and show that it significantly outperforms a parser-only baseline. Hierarchical Word Structure-based Parsing: A Feasibility Study on <fixed-case>UD</fixed-case>-style Dependency Parsing in <fixed-case>J</fixed-case>apanese - Takaaki Tanaka - Katsuhiko Hayashi - Masaaki Nagata + TakaakiTanaka + KatsuhikoHayashi + MasaakiNagata 56–60 W17-6308 In applying word-based dependency parsing such as Universal Dependencies (UD) to Japanese, the uncertainty of word segmentation emerges for defining a word unit of the dependencies. We introduce the following hierarchical word structures to dependency parsing in Japanese: morphological units (a short unit word, SUW) and syntactic units (a long unit word, LUW). An SUW can be used to segment a sentence consistently, while it is too short to represent syntactic construction. An LUW is a unit including functional multiwords and LUW-based analysis facilitates the capturing of syntactic structure and makes parsing results more precise than SUW-based analysis. This paper describes the results of a feasibility study on the ability and the effectiveness of parsing methods based on hierarchical word structure (LUW chunking+parsing) in comparison to single layer word structure (SUW parsing). We also show joint analysis of LUW-chunking and dependency parsing improves the performance of identifying predicate-argument structures, while there is not much difference between overall results of them. not much difference between overall results of them. Leveraging Newswire Treebanks for Parsing Conversational Data with Argument Scrambling - Riyaz A. Bhat - Irshad Bhat - Dipti Sharma + Riyaz A.Bhat + IrshadBhat + DiptiSharma 61–66 W17-6309 We investigate the problem of parsing conversational data of morphologically-rich languages such as Hindi where argument scrambling occurs frequently. We evaluate a state-of-the-art non-linear transition-based parsing system on a new dataset containing 506 dependency trees for sentences from Bollywood (Hindi) movie scripts and Twitter posts of Hindi monolingual speakers. We show that a dependency parser trained on a newswire treebank is strongly biased towards the canonical structures and degrades when applied to conversational data. Inspired by Transformational Generative Grammar (Chomsky, 1965), we mitigate the sampling bias by generating all theoretically possible alternative word orders of a clause from the existing (kernel) structures in the treebank. Training our parser on canonical and transformed structures improves performance on conversational data by around 9% LAS over the baseline newswire parser. Using hyperlinks to improve multilingual partial parsers - Anders Søgaard + AndersSøgaard 67–71 W17-6310 Syntactic annotation is costly and not available for the vast majority of the world’s languages. We show that sometimes we can do away with less labeled data by exploiting more readily available forms of mark-up. Specifically, we revisit an idea from Valentin Spitkovsky’s work (2010), namely that hyperlinks typically bracket syntactic constituents or chunks. We strengthen his results by showing that not only can hyperlinks help in low resource scenarios, exemplified here by Quechua, but learning from hyperlinks can also improve state-of-the-art NLP models for English newswire. We also present out-of-domain evaluation on English Ontonotes 4.0. Correcting prepositional phrase attachments using multimodal corpora - Sebastien Delecraz - Alexis Nasr - Frederic Bechet - Benoit Favre + SebastienDelecraz + AlexisNasr + FredericBechet + BenoitFavre 72–77 W17-6311 PP-attachments are an important source of errors in parsing natural language. We propose in this article to use data coming from a multimodal corpus, combining textual, visual and conceptual information, as well as a correction strategy, to propose alternative attachments in the output of a parser. Exploiting Structure in Parsing to 1-Endpoint-Crossing Graphs - Robin Kurtz - Marco Kuhlmann + RobinKurtz + MarcoKuhlmann 78–87 W17-6312 Deep dependency parsing can be cast as the search for maximum acyclic subgraphs in weighted digraphs. Because this search problem is intractable in the general case, we consider its restriction to the class of 1-endpoint-crossing (1ec) graphs, which has high coverage on standard data sets. Our main contribution is a characterization of 1ec graphs as a subclass of the graphs with pagenumber at most 3. Building on this we show how to extend an existing parsing algorithm for 1-endpoint-crossing trees to the full class. While the runtime complexity of the extended algorithm is polynomial in the length of the input sentence, it features a large constant, which poses a challenge for practical implementations. Effective Online Reordering with Arc-Eager Transitions - Ryosuke Kohita - Hiroshi Noji - Yuji Matsumoto + RyosukeKohita + HiroshiNoji + YujiMatsumoto 88–98 W17-6313 We present a new transition system with word reordering for unrestricted non-projective dependency parsing. Our system is based on decomposed arc-eager rather than arc-standard, which allows more flexible ambiguity resolution between a local projective and non-local crossing attachment. In our experiment on Universal Dependencies 2.0, we find our parser outperforms the ordinary swap-based parser particularly on languages with a large amount of non-projectivity. Arc-Hybrid Non-Projective Dependency Parsing with a Static-Dynamic Oracle - Miryam de Lhoneux - Sara Stymne - Joakim Nivre + Miryamde Lhoneux + SaraStymne + JoakimNivre 99–104 W17-6314 In this paper, we extend the arc-hybrid system for transition-based parsing with a swap transition that enables reordering of the words and construction of non-projective trees. Although this extension breaks the arc-decomposability of the transition system, we show how the existing dynamic oracle for this system can be modified and combined with a static oracle only for the swap transition. Experiments on 5 languages show that the new system gives competitive accuracy and is significantly better than a system trained with a purely static oracle. @@ -11428,33 +11428,33 @@ with emotion annotation. We (a) analyse annotation reliability and annotation me Encoder-Decoder Shift-Reduce Syntactic Parsing - Jiangming Liu - Yue Zhang + JiangmingLiu + YueZhang 105–114 W17-6315 Encoder-decoder neural networks have been used for many NLP tasks, such as neural machine translation. They have also been applied to constituent parsing by using bracketed tree structures as a target language, translating input sentences into syntactic trees. A more commonly used method to linearize syntactic trees is the shift-reduce system, which uses a sequence of transition-actions to build trees. We empirically investigate the effectiveness of applying the encoder-decoder network to transition-based parsing. On standard benchmarks, our system gives comparable results to the stack LSTM parser for dependency parsing, and significantly better results compared to the aforementioned parser for constituent parsing, which uses bracketed tree formats. Arc-Standard Spinal Parsing with Stack-<fixed-case>LSTM</fixed-case>s - Miguel Ballesteros - Xavier Carreras + MiguelBallesteros + XavierCarreras 115–121 W17-6316 We present a neural transition-based parser for spinal trees, a dependency representation of constituent trees. The parser uses Stack-LSTMs that compose constituent nodes with dependency-based derivations. In experiments, we show that this model adapts to different styles of dependency relations, but this choice has little effect for predicting constituent structure, suggesting that LSTMs induce useful states by themselves. Coarse-To-Fine Parsing for Expressive Grammar Formalisms - Christoph Teichmann - Alexander Koller - Jonas Groschwitz + ChristophTeichmann + AlexanderKoller + JonasGroschwitz 122–127 W17-6317 We generalize coarse-to-fine parsing to grammar formalisms that are more expressive than PCFGs and/or describe languages of trees or graphs. We evaluate our algorithm on PCFG, PTAG, and graph parsing. While we achieve the expected performance gains on PCFGs, coarse-to-fine does not help for PTAG and can even slow down parsing for graphs. We discuss the implications of this finding. Evaluating <fixed-case>LSTM</fixed-case> models for grammatical function labelling - Bich-Ngoc Do - Ines Rehbein + Bich-NgocDo + InesRehbein 128–133 W17-6318 To improve grammatical function labelling for German, we augment the labelling component of a neural dependency parser with a decision history. We present different ways to encode the history, using different LSTM architectures, and show that our models yield significant improvements, resulting in a LAS for German that is close to the best result from the SPMRL 2014 shared task (without the reranker). @@ -11476,236 +11476,236 @@ with emotion annotation. We (a) analyse annotation reliability and annotation me Capturing Dependency Syntax with “Deep” Sequential Models - Yoav Goldberg + YoavGoldberg 1 W17-6501 Syntax-Semantics Interface: A Plea for a Deep Dependency Sentence Structure - Eva Hajičová + EvaHajičová 2–3 W17-6502 The Benefit of Syntactic vs. Linear N-grams for Linguistic Description - Melanie Andresen - Heike Zinsmeister + MelanieAndresen + HeikeZinsmeister 4–14 W17-6503 On the Predicate-Argument Structure: Internal and Absorbing Scope - Igor Boguslavsky + IgorBoguslavsky 15–24 W17-6504 On the order of Words in <fixed-case>I</fixed-case>talian: a Study on Genre vs Complexity - Dominique Brunato - Felice Dell’Orletta + DominiqueBrunato + FeliceDell’Orletta 25–31 W17-6505 Revising the <fixed-case>METU</fixed-case>-Sabancı <fixed-case>T</fixed-case>urkish Treebank: An Exercise in Surface-Syntactic Annotation of Agglutinative Languages - Alicia Burga - Alp Öktem - Leo Wanner + AliciaBurga + AlpÖktem + LeoWanner 32–41 W17-6506 Enhanced <fixed-case>UD</fixed-case> Dependencies with Neutralized Diathesis Alternation - Marie Candito - Bruno Guillaume - Guy Perrier - Djamé Seddah + MarieCandito + BrunoGuillaume + GuyPerrier + DjaméSeddah 42–53 W17-6507 Classifying Languages by Dependency Structure. Typologies of Delexicalized Universal Dependency Treebanks - Xinying Chen - Kim Gerdes + XinyingChen + KimGerdes 54–63 W17-6508 A Dependency Treebank for <fixed-case>K</fixed-case>urmanji <fixed-case>K</fixed-case>urdish - Memduh Gökırmak - Francis M. Tyers + MemduhGökırmak + Francis M.Tyers 64–72 W17-6509 What are the limitations on the flux of syntactic dependencies? Evidence from <fixed-case>UD</fixed-case> treebanks - Sylvain Kahane - Chunxiao Yan - Marie-Amélie Botalla + SylvainKahane + ChunxiaoYan + Marie-AmélieBotalla 73–82 W17-6510 Fully Delexicalized Contexts for Syntax-Based Word Embeddings - Jenna Kanerva - Sampo Pyysalo - Filip Ginter + JennaKanerva + SampoPyysalo + FilipGinter 83–91 W17-6511 Universal Dependencies for Dargwa Mehweb - Alexandra Kozhukhar + AlexandraKozhukhar 92–99 W17-6512 Menzerath-Altmann Law in Syntactic Dependency Structure - Ján Mačutek - Radek Čech - Jiří Milička + JánMačutek + RadekČech + JiříMilička 100–107 W17-6513 Assessing the Annotation Consistency of the Universal Dependencies Corpora - Marie-Catherine de Marneffe - Matias Grioni - Jenna Kanerva - Filip Ginter + Marie-Catherinede Marneffe + MatiasGrioni + JennaKanerva + FilipGinter 108–115 W17-6514 To What Extent is Immediate Constituency Analysis Dependency-Based? A Survey of Foundational Texts - Nicolas Mazziotta - Sylvain Kahane + NicolasMazziotta + SylvainKahane 116–126 W17-6515 Dependency Structure of Binary Conjunctions(of the <fixed-case>IF</fixed-case>…, <fixed-case>THEN</fixed-case>… Type) - Igor Mel’čuk + IgorMel’čuk 127–134 W17-6516 Non-Projectivity in <fixed-case>S</fixed-case>erbian: Analysis of Formal and Linguistic Properties - Aleksandra Miletic - Assaf Urieli + AleksandraMiletic + AssafUrieli 135–144 W17-6517 Prices go Up, Surge, Jump, Spike, Skyrocket, Go through the Roof… Intensifier Collocations with Parametric Nouns of Type <fixed-case>PRICE</fixed-case> - Jasmina Milićević + JasminaMilićević 145–153 W17-6518 <fixed-case>C</fixed-case>hinese Descriptive and Resultative V-de Constructions. A Dependency-based Analysis - Ruochen Niu + RuochenNiu 154–164 W17-6519 The Component Unit. Introducing a Novel Unit of Syntactic Analysis - Timothy Osborne - Ruochen Niu + TimothyOsborne + RuochenNiu 165–175 W17-6520 Control vs. Raising in <fixed-case>E</fixed-case>nglish. A Dependency Grammar Account - Timothy Osborne - Matthew Reeve + TimothyOsborne + MatthewReeve 176–186 W17-6521 Segmentation Granularity in Dependency Representations for <fixed-case>K</fixed-case>orean - Jungyeul Park + JungyeulPark 187–196 W17-6522 Universal Dependencies for <fixed-case>P</fixed-case>ortuguese - Alexandre Rademaker - Fabricio Chalub - Livy Real - Cláudia Freitas - Eckhard Bick - Valeria de Paiva + AlexandreRademaker + FabricioChalub + LivyReal + CláudiaFreitas + EckhardBick + Valeriade Paiva 197–206 W17-6523 <fixed-case>UDL</fixed-case>ex: Towards Cross-language Subcategorization Lexicons - Giulia Rambelli - Alessandro Lenci - Thierry Poibeau + GiuliaRambelli + AlessandroLenci + ThierryPoibeau 207–2017 W17-6524 Universal Dependencies are Hard to Parse – or are They? - Ines Rehbein - Julius Steen - Bich-Ngoc Do - Anette Frank + InesRehbein + JuliusSteen + Bich-NgocDo + AnetteFrank 218–228 W17-6525 Annotating <fixed-case>I</fixed-case>talian Social Media Texts in Universal Dependencies - Manuela Sanguinetti - Cristina Bosco - Alessandro Mazzei - Alberto Lavelli - Fabio Tamburini + ManuelaSanguinetti + CristinaBosco + AlessandroMazzei + AlbertoLavelli + FabioTamburini 229–239 W17-6526 <fixed-case>H</fixed-case>ungarian Copula Constructions in Dependency Syntax and Parsing - Katalin Ilona Simkó - Veronika Vincze + Katalin IlonaSimkó + VeronikaVincze 240–247 W17-6527 Semgrex-Plus: a Tool for Automatic Dependency-Graph Rewriting - Fabio Tamburini + FabioTamburini 248–254 W17-6528 Unity in Diversity: A Unified Parsing Strategy for Major <fixed-case>I</fixed-case>ndian Languages - Juhi Tandon - Dipti Misra Sharma + JuhiTandon + Dipti MisraSharma 255–265 W17-6529 Quantitative Comparative Syntax on the <fixed-case>C</fixed-case>antonese-<fixed-case>M</fixed-case>andarin Parallel Dependency Treebank - Tak-sum Wong - Kim Gerdes - Herman Leung - John Lee + Tak-sumWong + KimGerdes + HermanLeung + JohnLee 266–275 W17-6530 Understanding Constraints on Non-Projectivity Using Novel Measures - Himanshu Yadav - Ashwini Vaidya - Samar Husain + HimanshuYadav + AshwiniVaidya + SamarHusain 276–286 W17-6531 Core Arguments in Universal Dependencies - Daniel Zeman + DanielZeman 287–296 W17-6532 @@ -12417,11 +12417,11 @@ with emotion annotation. We (a) analyse annotation reliability and annotation me Proceedings of Language, Ontology, Terminology and Knowledge Structures Workshop (LOTKS 2017) - Francesca Frontini - Larisa Grčić Simeunović - Špela Vintar - Anas Fahad Khan - Artemis Parvisi + FrancescaFrontini + LarisaGrčić Simeunović + ŠpelaVintar + Anas FahadKhan + ArtemisParvisi Association for Computational Linguistics
Montpellier, France
September @@ -12432,77 +12432,77 @@ with emotion annotation. We (a) analyse annotation reliability and annotation me Exploratory Analysis for Ontology Learning from Social Events on Social Media Streaming in <fixed-case>S</fixed-case>panish - Enrique Valeriano - Arturo Oncevay-Marcos + EnriqueValeriano + ArturoOncevay-Marcos W17-7001 Creating a gold standard corpus for terminological annotation from online forum data - Anna Hätty - Simon Tannert - Ulrich Heid + AnnaHätty + SimonTannert + UlrichHeid W17-7002 A conceptual ontology in the water domain of knowledge to bridge the lexical semantics of stratified discursive strata - Jean-Louis Janin - Henri Portine + Jean-LouisJanin + HenriPortine W17-7003 <fixed-case>G</fixed-case>eo<fixed-case>D</fixed-case>ict: an integrated gazetteer - Jacques Fize - Gaurav Shrivastava - Pierre André Ménard + JacquesFize + GauravShrivastava + Pierre AndréMénard W17-7004 Fine-grained domain classification of text using <fixed-case>TERMIUM</fixed-case> Plus - Gabriel Bernier-Colborne - Caroline Barrière - Pierre André Ménard + GabrielBernier-Colborne + CarolineBarrière + Pierre AndréMénard W17-7005 <fixed-case>TBX</fixed-case> in <fixed-case>ODD</fixed-case>: Schema-agnostic specification and documentation for <fixed-case>T</fixed-case>erm<fixed-case>B</fixed-case>ase e<fixed-case>X</fixed-case>change - Stefan Pernes - Laurent Romary + StefanPernes + LaurentRomary W17-7006 Enrichment of <fixed-case>F</fixed-case>rench Biomedical Ontologies with <fixed-case>UMLS</fixed-case> Concepts and Semantic Types for Biomedical Named Entity Recognition Though Ontological Semantic Annotation - Andon Tchechmedjiev - Clément Jonquet + AndonTchechmedjiev + ClémentJonquet W17-7007 Experiments in taxonomy induction in <fixed-case>S</fixed-case>panish and <fixed-case>F</fixed-case>rench - Irene Renau - Rogelio Nazar - Rafael Marín + IreneRenau + RogelioNazar + RafaelMarín W17-7008 A statistical model for morphology inspired by the Amis language - Isabelle Bril - Achraf Lassoued - Michel de Rougemont + IsabelleBril + AchrafLassoued + Michelde Rougemont W17-7009 Developing <fixed-case>L</fixed-case>ex<fixed-case>O</fixed-case>: a Collaborative Editor of Multilingual Lexica and Termino-Ontological Resources in the Humanities - Andrea Bellandi - Emiliano Giovannetti - Silvia Piccini - Anja Weingart + AndreaBellandi + EmilianoGiovannetti + SilviaPiccini + AnjaWeingart W17-7010 Designing an Ontology for the Study of Ritual in Ancient <fixed-case>G</fixed-case>reek Tragedy - Gloria Mugelli - Andrea Bellandi - Federico Boschetti - Anas Fahad Khan + GloriaMugelli + AndreaBellandi + FedericoBoschetti + Anas FahadKhan W17-7011
@@ -12608,56 +12608,56 @@ with emotion annotation. We (a) analyse annotation reliability and annotation me The Effect of Negative Sampling Strategy on Capturing Semantic Similarity in Document Embeddings - Marzieh Saeidi - Ritwik Kulkarni - Theodosia Togia - Michele Sama + MarziehSaeidi + RitwikKulkarni + TheodosiaTogia + MicheleSama 1–8 W17-7301 Building Graph Representations of Deep Vector Embeddings - Dario Garcia-Gasulla - Armand Vilalta - Ferran Parés - Jonathan Moreno - Eduard Ayguadé - Jesús Labarta - Ulises Cortés - Toyotaro Suzumura + DarioGarcia-Gasulla + ArmandVilalta + FerranParés + JonathanMoreno + EduardAyguadé + JesúsLabarta + UlisesCortés + ToyotaroSuzumura 9–15 W17-7302 Class Disjointness Constraints as Specific Objective Functions in Neural Network Classifiers - François Scharffe + FrançoisScharffe 16–23 W17-7303 Full-Network Embedding in a Multimodal Embedding Pipeline - Armand Vilalta - Dario Garcia-Gasulla - Ferran Parés - Jonathan Moreno - Eduard Ayguadé - Jesus Labarta - Ulises Cortés - Toyotaro Suzumura + ArmandVilalta + DarioGarcia-Gasulla + FerranParés + JonathanMoreno + EduardAyguadé + JesusLabarta + UlisesCortés + ToyotaroSuzumura 24–32 W17-7304 Extracting Tags from Large Raw Texts Using End-to-End Memory Networks - Feras Al Kassar - Frédéric Armetta + FerasAl Kassar + FrédéricArmetta 33–40 W17-7305 Dealing with Co-reference in Neural Semantic Parsing - Rik van Noord - Johan Bos + Rikvan Noord + JohanBos 41–49 W17-7306 @@ -13275,7 +13275,7 @@ with emotion annotation. We (a) analyse annotation reliability and annotation me Proceedings of the 16th International Workshop on Treebanks and Linguistic Theories W17-76 - Jan Hajič + JanHajič
Prague, Czech Republic
2017 @@ -13284,177 +13284,177 @@ with emotion annotation. We (a) analyse annotation reliability and annotation me Annotating and parsing to semantic frames: feedback from the <fixed-case>F</fixed-case>rench <fixed-case>F</fixed-case>rame<fixed-case>N</fixed-case>et project - Marie Candito + MarieCandito v W17-7601 Downstream use of syntactic analysis: does representation matter? - Lilja Øvrelid + LiljaØvrelid vi W17-7602 Distributional regularities of verbs and verbal adjectives: Treebank evidence and broader implications - Daniël de Kok - Patricia Fischer - Corina Dima - Erhard Hinrichs + Daniëlde Kok + PatriciaFischer + CorinaDima + ErhardHinrichs 1–9 W17-7603 <fixed-case>UD</fixed-case> Annotatrix: An annotation tool for Universal Dependencies - Francis M. Tyers - Mariya Sheyanova - Jonathan North Washington + Francis M.Tyers + MariyaSheyanova + Jonathan NorthWashington 10–17 W17-7604 The Treebanked Conspiracy. Actors and Actions in Bellum Catilinae - Marco Passarotti - Berta González Saavedra + MarcoPassarotti + BertaGonzález Saavedra 18–26 W17-7605 Universal Dependencies-based syntactic features in detecting human translation varieties - Maria Kunilovskaya - Andrey Kutuzov + MariaKunilovskaya + AndreyKutuzov 27–36 W17-7606 Graph Convolutional Networks for Named Entity Recognition - Alberto Cetoli - Stefano Bragaglia - Andrew O’Harney - Marc Sloan + AlbertoCetoli + StefanoBragaglia + AndrewO’Harney + MarcSloan 37–45 W17-7607 Extensions to the <fixed-case>G</fixed-case>r<fixed-case>ETEL</fixed-case> Treebank Query Application - Jan Odijk - Martijn van der Klis - Sheean Spoel + JanOdijk + Martijnvan der Klis + SheeanSpoel 46–55 W17-7608 The Relation of Form and Function in Linguistic Theory and in a Multilayer Treebank - Eduard Bejček - Eva Hajičová - Marie Mikulová - Jarmila Panevová + EduardBejček + EvaHajičová + MarieMikulová + JarmilaPanevová 56–63 W17-7609 Literal readings of multiword expressions: as scarce as hen’s teeth - Agata Savary - Silvio Ricardo Cordeiro + AgataSavary + Silvio RicardoCordeiro 64–72 W17-7610 Querying Multi-word Expressions Annotation with <fixed-case>CQL</fixed-case> - Natalia Klyueva - Anna Vernerová - Behrang Qasemizadeh + NataliaKlyueva + AnnaVernerová + BehrangQasemizadeh 73–79 W17-7611 <fixed-case>REALEC</fixed-case> learner treebank: annotation principles and evaluation of automatic parsing - Olga Lyashevskaya - Irina Panteleeva + OlgaLyashevskaya + IrinaPanteleeva 80–87 W17-7612 A semiautomatic lemmatisation procedure for treebanks. Old <fixed-case>E</fixed-case>nglish strong and weak verbs - Marta Tío Sáenz - Darío Metola Rodríguez + MartaTío Sáenz + DaríoMetola Rodríguez 88–94 W17-7613 Data point selection for genre-aware parsing - Ines Rehbein - Felix Bildhauer + InesRehbein + FelixBildhauer 95–105 W17-7614 Error Analysis of Cross-lingual Tagging and Parsing - Rudolf Rosa - Zdeněk Žabokrtský + RudolfRosa + ZdeněkŽabokrtský 106–118 W17-7615 A <fixed-case>T</fixed-case>elugu treebank based on a grammar book - Taraka Rama - Sowmya Vajjala + TarakaRama + SowmyaVajjala 119–128 W17-7616 Recent Developments within <fixed-case>B</fixed-case>ul<fixed-case>T</fixed-case>ree<fixed-case>B</fixed-case>ank - Petya Osenova - Kiril Simov + PetyaOsenova + KirilSimov 129–137 W17-7617 Towards a dependency-annotated treebank for <fixed-case>B</fixed-case>ambara - Ekaterina Aplonova - Francis M. Tyers + EkaterinaAplonova + Francis M.Tyers 138–145 W17-7618 Merging the Trees - Building a Morphological Treebank for <fixed-case>G</fixed-case>erman from Two Resources - Petra Steiner + PetraSteiner 146–160 W17-7619 What <fixed-case>I</fixed-case> think when <fixed-case>I</fixed-case> think about treebanks - Anders Søgaard + AndersSøgaard 161–166 W17-7620 Syntactic Semantic Correspondence in Dependency Grammar - Cătălina Mărănduc - Cătălin Mititelu - Victoria Bobicev + CătălinaMărănduc + CătălinMititelu + VictoriaBobicev 167–180 W17-7621 Multi-word annotation in syntactic treebanks - Propositions for Universal Dependencies - Sylvain Kahane - Marine Courtin - Kim Gerdes + SylvainKahane + MarineCourtin + KimGerdes 181–189 W17-7622 A Universal Dependencies Treebank for <fixed-case>M</fixed-case>arathi - Vinit Ravishankar + VinitRavishankar 190–200 W17-7623 Dangerous Relations in Dependency Treebanks - Chiara Alzetta - Felice Dell’Orletta - Simonetta Montemagni - Giulia Venturi + ChiaraAlzetta + FeliceDell’Orletta + SimonettaMontemagni + GiuliaVenturi 201–210 W17-7624 @@ -13472,8 +13472,8 @@ with emotion annotation. We (a) analyse annotation reliability and annotation me Proceedings of the 1st Workshop on Natural Language Processing and Information Retrieval associated with RANLP 2017 - Mireille Makary - Michael Oakes + MireilleMakary + MichaelOakes INCOMA Inc.
Varna, Bulgaria
September @@ -13484,9 +13484,9 @@ with emotion annotation. We (a) analyse annotation reliability and annotation me Deception Detection for the <fixed-case>R</fixed-case>ussian Language: Lexical and Syntactic Parameters - Dina Pisarevskaya - Tatiana Litvinova - Olga Litvinova + DinaPisarevskaya + TatianaLitvinova + OlgaLitvinova 1–10 10.26615/978-954-452-038-0_001 https://doi.org/10.26615/978-954-452-038-0_001 @@ -13494,10 +13494,10 @@ with emotion annotation. We (a) analyse annotation reliability and annotation me o<fixed-case>IQ</fixed-case>a: An Opinion Influence Oriented Question Answering Framework with Applications to Marketing Domain - Dumitru-Clementin Cercel - Cristian Onose - Stefan Trausan-Matu - Florin Pop + Dumitru-ClementinCercel + CristianOnose + StefanTrausan-Matu + FlorinPop 11–18 10.26615/978-954-452-038-0_002 https://doi.org/10.26615/978-954-452-038-0_002 @@ -13505,9 +13505,9 @@ with emotion annotation. We (a) analyse annotation reliability and annotation me Automatic Summarization of Online Debates - Nattapong Sanchan - Ahmet Aker - Kalina Bontcheva + NattapongSanchan + AhmetAker + KalinaBontcheva 19–27 10.26615/978-954-452-038-0_003 https://doi.org/10.26615/978-954-452-038-0_003 @@ -13515,10 +13515,10 @@ with emotion annotation. We (a) analyse annotation reliability and annotation me A Game with a Purpose for Automatic Detection of Children’s Speech Disabilities using Limited Speech Resources - Reem Salem - Mohamed Elmahdy - Slim Abdennadher - Injy Hamed + ReemSalem + MohamedElmahdy + SlimAbdennadher + InjyHamed 28–34 10.26615/978-954-452-038-0_004 https://doi.org/10.26615/978-954-452-038-0_004 @@ -13530,8 +13530,8 @@ with emotion annotation. We (a) analyse annotation reliability and annotation me Proceedings of the Workshop Knowledge Resources for the Socio-Economic Sciences and Humanities associated with RANLP 2017 KalliopiZervanou PetyaOsenova - Eveline Wandl-Vogt - Dan Cristea + EvelineWandl-Vogt + DanCristea INCOMA Inc.
Varna
September @@ -13542,8 +13542,8 @@ with emotion annotation. We (a) analyse annotation reliability and annotation me Connecting people digitally - a semantic web based approach to linking heterogeneous data sets - Katalin Lejtovicz - Amelie Dorn + KatalinLejtovicz + AmelieDorn 1–8 10.26615/978-954-452-040-3_001 https://doi.org/10.26615/978-954-452-040-3_001 @@ -13551,9 +13551,9 @@ with emotion annotation. We (a) analyse annotation reliability and annotation me A Multiform Balanced Dependency Treebank for <fixed-case>R</fixed-case>omanian - Mihaela Colhon - Cătălina Mărănduc - Cătălin Mititelu + MihaelaColhon + CătălinaMărănduc + CătălinMititelu 9–18 10.26615/978-954-452-040-3_002 https://doi.org/10.26615/978-954-452-040-3_002 @@ -13561,11 +13561,11 @@ with emotion annotation. We (a) analyse annotation reliability and annotation me <fixed-case>GR</fixed-case>a<fixed-case>SP</fixed-case>: Grounded Representation and Source Perspective - Antske Fokkens - Piek Vossen - Marco Rospocher - Rinke Hoekstra - Willem Robert van Hage + AntskeFokkens + PiekVossen + MarcoRospocher + RinkeHoekstra + Willem Robertvan Hage 19–25 10.26615/978-954-452-040-3_003 https://doi.org/10.26615/978-954-452-040-3_003 @@ -13573,7 +13573,7 @@ with emotion annotation. We (a) analyse annotation reliability and annotation me Educational Content Generation for Business and Administration <fixed-case>FL</fixed-case> Courses with the <fixed-case>NBU</fixed-case> <fixed-case>PLT</fixed-case> Platform - Maria Stambolieva + MariaStambolieva 26–30 10.26615/978-954-452-040-3_004 https://doi.org/10.26615/978-954-452-040-3_004 @@ -13581,15 +13581,15 @@ with emotion annotation. We (a) analyse annotation reliability and annotation me Machine Learning Models of Universal Grammar Parameter Dependencies - Dimitar Kazakov - Guido Cordoni - Andrea Ceolin - Monica-Alexandrina Irimia - Shin-Sook Kim - Dimitris Michelioudakis - Nina Radkevich - Cristina Guardiano - Giuseppe Longobardi + DimitarKazakov + GuidoCordoni + AndreaCeolin + Monica-AlexandrinaIrimia + Shin-SookKim + DimitrisMichelioudakis + NinaRadkevich + CristinaGuardiano + GiuseppeLongobardi 31–37 10.26615/978-954-452-040-3_005 https://doi.org/10.26615/978-954-452-040-3_005 @@ -13613,11 +13613,11 @@ with emotion annotation. We (a) analyse annotation reliability and annotation me Enhancing Machine Translation of Academic Course Catalogues with Terminological Resources - Randy Scansani - Silvia Bernardini - Adriano Ferraresi - Federico Gaspari - Marcello Soffritti + RandyScansani + SilviaBernardini + AdrianoFerraresi + FedericoGaspari + MarcelloSoffritti 1–10 10.26615/978-954-452-042-7_001 https://doi.org/10.26615/978-954-452-042-7_001 @@ -13625,9 +13625,9 @@ with emotion annotation. We (a) analyse annotation reliability and annotation me Experiments in Non-Coherent Post-editing - Cristina Toledo Báez - Moritz Schaeffer - Michael Carl + CristinaToledo Báez + MoritzSchaeffer + MichaelCarl 11–20 10.26615/978-954-452-042-7_002 https://doi.org/10.26615/978-954-452-042-7_002 @@ -13635,7 +13635,7 @@ with emotion annotation. We (a) analyse annotation reliability and annotation me Comparing Machine Translation and Human Translation: A Case Study - Lars Ahrenberg + LarsAhrenberg 21–28 10.26615/978-954-452-042-7_003 https://doi.org/10.26615/978-954-452-042-7_003 @@ -13643,8 +13643,8 @@ with emotion annotation. We (a) analyse annotation reliability and annotation me <fixed-case>T</fixed-case>rans<fixed-case>B</fixed-case>ank: Metadata as the Missing Link between <fixed-case>NLP</fixed-case> and Traditional Translation Studies - Michael Ustaszewski - Andy Stauder + MichaelUstaszewski + AndyStauder 29–35 10.26615/978-954-452-042-7_004 https://doi.org/10.26615/978-954-452-042-7_004 @@ -13652,11 +13652,11 @@ with emotion annotation. We (a) analyse annotation reliability and annotation me Interpreting Strategies Annotation in the <fixed-case>WAW</fixed-case> Corpus - Irina Temnikova - Ahmed Abdelali - Samy Hedaya - Stephan Vogel - Aishah Al Daher + IrinaTemnikova + AhmedAbdelali + SamyHedaya + StephanVogel + AishahAl Daher 36–43 10.26615/978-954-452-042-7_005 https://doi.org/10.26615/978-954-452-042-7_005 @@ -13664,8 +13664,8 @@ with emotion annotation. We (a) analyse annotation reliability and annotation me Translation Memory Systems Have a Long Way to Go - Andrea Silvestre Baquero - Ruslan Mitkov + AndreaSilvestre Baquero + RuslanMitkov 44–51 10.26615/978-954-452-042-7_006 https://doi.org/10.26615/978-954-452-042-7_006 @@ -13673,8 +13673,8 @@ with emotion annotation. We (a) analyse annotation reliability and annotation me Building Dialectal <fixed-case>A</fixed-case>rabic Corpora - Hani Elgabou - Dimitar Kazakov + HaniElgabou + DimitarKazakov 52–57 10.26615/978-954-452-042-7_007 https://doi.org/10.26615/978-954-452-042-7_007 @@ -13682,8 +13682,8 @@ with emotion annotation. We (a) analyse annotation reliability and annotation me Towards Producing Human-Validated Translation Resources for the <fixed-case>F</fixed-case>ula language through <fixed-case>W</fixed-case>ord<fixed-case>N</fixed-case>et Linking - Khalil Mrini - Martin Benjamin + KhalilMrini + MartinBenjamin 58–64 10.26615/978-954-452-042-7_008 https://doi.org/10.26615/978-954-452-042-7_008 @@ -13707,7 +13707,7 @@ with emotion annotation. We (a) analyse annotation reliability and annotation me Document retrieval and question answering in medical documents. A large-scale corpus challenge. - Curea Eric + CureaEric 1–7 10.26615/978-954-452-044-1_001 https://doi.org/10.26615/978-954-452-044-1_001 @@ -13715,8 +13715,8 @@ with emotion annotation. We (a) analyse annotation reliability and annotation me Adapting the <fixed-case>TTL</fixed-case> <fixed-case>R</fixed-case>omanian <fixed-case>POS</fixed-case> Tagger to the Biomedical Domain - Maria Mitrofan - Radu Ion + MariaMitrofan + RaduIon 8–14 10.26615/978-954-452-044-1_002 https://doi.org/10.26615/978-954-452-044-1_002 @@ -13724,9 +13724,9 @@ with emotion annotation. We (a) analyse annotation reliability and annotation me Discourse-Wide Extraction of Assay Frames from the Biological Literature - Dayne Freitag - Paul Kalmar - Eric Yeh + DayneFreitag + PaulKalmar + EricYeh 15–23 10.26615/978-954-452-044-1_003 https://doi.org/10.26615/978-954-452-044-1_003 @@ -13734,7 +13734,7 @@ with emotion annotation. We (a) analyse annotation reliability and annotation me Classification based extraction of numeric values from clinical narratives - Maximilian Zubke + MaximilianZubke 24–31 10.26615/978-954-452-044-1_004 https://doi.org/10.26615/978-954-452-044-1_004 @@ -13742,8 +13742,8 @@ with emotion annotation. We (a) analyse annotation reliability and annotation me Understanding of unknown medical words - Natalia Grabar - Thierry Hamon + NataliaGrabar + ThierryHamon 32–41 10.26615/978-954-452-044-1_005 https://doi.org/10.26615/978-954-452-044-1_005 @@ -13751,11 +13751,11 @@ with emotion annotation. We (a) analyse annotation reliability and annotation me Entity-Centric Information Access with Human in the Loop for the Biomedical Domain - Seid Muhie Yimam - Steffen Remus - Alexander Panchenko - Andreas Holzinger - Chris Biemann + Seid MuhieYimam + SteffenRemus + AlexanderPanchenko + AndreasHolzinger + ChrisBiemann 42–48 10.26615/978-954-452-044-1_006 https://doi.org/10.26615/978-954-452-044-1_006 @@ -13763,8 +13763,8 @@ with emotion annotation. We (a) analyse annotation reliability and annotation me One model per entity: using hundreds of machine learning models to recognize and normalize biomedical names in text - Victor Bellon - Raul Rodriguez-Esteban + VictorBellon + RaulRodriguez-Esteban 49–54 10.26615/978-954-452-044-1_007 https://doi.org/10.26615/978-954-452-044-1_007 @@ -13772,8 +13772,8 @@ with emotion annotation. We (a) analyse annotation reliability and annotation me Towards Confidence Estimation for Typed Protein-Protein Relation Extraction - Camilo Thorne - Roman Klinger + CamiloThorne + RomanKlinger 55–63 10.26615/978-954-452-044-1_008 https://doi.org/10.26615/978-954-452-044-1_008 @@ -13781,10 +13781,10 @@ with emotion annotation. We (a) analyse annotation reliability and annotation me Identification of Risk Factors in Clinical Texts through Association Rules - Svetla Boytcheva - Ivelina Nikolova - Galia Angelova - Zhivko Angelov + SvetlaBoytcheva + IvelinaNikolova + GaliaAngelova + ZhivkoAngelov 64–72 10.26615/978-954-452-044-1_009 https://doi.org/10.26615/978-954-452-044-1_009 @@ -13792,11 +13792,11 @@ with emotion annotation. We (a) analyse annotation reliability and annotation me <fixed-case>POMELO</fixed-case>: <fixed-case>M</fixed-case>edline corpus with manually annotated food-drug interactions - Thierry Hamon - Vincent Tabanou - Fleur Mougin - Natalia Grabar - Frantz Thiessard + ThierryHamon + VincentTabanou + FleurMougin + NataliaGrabar + FrantzThiessard 73–80 10.26615/978-954-452-044-1_010 https://doi.org/10.26615/978-954-452-044-1_010 @@ -13804,10 +13804,10 @@ with emotion annotation. We (a) analyse annotation reliability and annotation me Annotation of Clinical Narratives in <fixed-case>B</fixed-case>ulgarian language - Ivajlo Radev - Kiril Simov - Galia Angelova - Svetla Boytcheva + IvajloRadev + KirilSimov + GaliaAngelova + SvetlaBoytcheva 81–87 10.26615/978-954-452-044-1_011 https://doi.org/10.26615/978-954-452-044-1_011 @@ -13830,9 +13830,9 @@ with emotion annotation. We (a) analyse annotation reliability and annotation me A Diachronic Corpus for <fixed-case>R</fixed-case>omanian (<fixed-case>R</fixed-case>o<fixed-case>D</fixed-case>ia) - Ludmila Malahov - Cătălina Mărănduc - Alexandru Colesnicov + LudmilaMalahov + CătălinaMărănduc + AlexandruColesnicov 1–9 0.26615/978-954-452-046-5_001 http://doi.org/10.26615/978-954-452-046-5_001 @@ -13840,9 +13840,9 @@ with emotion annotation. We (a) analyse annotation reliability and annotation me Tools for Building a Corpus to Study the Historical and Geographical Variation of the <fixed-case>R</fixed-case>omanian Language - Victoria Bobicev - Cătălina Mărănduc - Cenel Augusto Perez + VictoriaBobicev + CătălinaMărănduc + Cenel AugustoPerez 10–19 0.26615/978-954-452-046-5_002 http://doi.org/10.26615/978-954-452-046-5_002 @@ -13850,12 +13850,12 @@ with emotion annotation. We (a) analyse annotation reliability and annotation me Multilingual Ontologies for the Representation and Processing of Folktales - Thierry Declerck - Anastasija Aman - Martin Banzer - Dominik Macháček - Lisa Schäfer - Natalia Skachkova + ThierryDeclerck + AnastasijaAman + MartinBanzer + DominikMacháček + LisaSchäfer + NataliaSkachkova 20–23 0.26615/978-954-452-046-5_003 http://doi.org/10.26615/978-954-452-046-5_003 @@ -13863,9 +13863,9 @@ with emotion annotation. We (a) analyse annotation reliability and annotation me On the annotation of vague expressions: a case study on <fixed-case>R</fixed-case>omanian historical texts - Anca Dinu - Walther von Hahn - Cristina Vertan + AncaDinu + Walthervon Hahn + CristinaVertan 24–31 0.26615/978-954-452-046-5_004 http://doi.org/10.26615/978-954-452-046-5_004 @@ -13873,11 +13873,11 @@ with emotion annotation. We (a) analyse annotation reliability and annotation me Language Technologies in Teaching Bugarian at Primary and Secondary School Level: the <fixed-case>NBU</fixed-case> Platform of Language Teaching (<fixed-case>PLT</fixed-case>) - Maria Stambolieva - Valentina Ivanova - Mariana Raykova - Milka Hadjikoteva - Mariya Neykova + MariaStambolieva + ValentinaIvanova + MarianaRaykova + MilkaHadjikoteva + MariyaNeykova 32–38 0.26615/978-954-452-046-5_005 http://doi.org/10.26615/978-954-452-046-5_005 @@ -13885,7 +13885,7 @@ with emotion annotation. We (a) analyse annotation reliability and annotation me Natural Language Processing in Political Campaigns - Cristina Moise + CristinaMoise 39–43 0.26615/978-954-452-046-5_006 http://doi.org/10.26615/978-954-452-046-5_006 diff --git a/data/xml/W19.xml b/data/xml/W19.xml index 02495714ef..85c327331c 100644 --- a/data/xml/W19.xml +++ b/data/xml/W19.xml @@ -16525,4 +16525,369 @@ In this tutorial on MT and post-editing we would like to continue sharing the la 10.18653/v1/W19-8673
+ + + Proceedings of the Human-Informed Translation and Interpreting Technology Workshop (HiT-IT 2019) + W19-87 + Incoma Ltd., Shoumen, Bulgaria +
Varna, Bulgaria
+ September + 2019 + + + W19-8700 + + + Comparison between Automatic and Human Subtitling: A Case Study with Game of Thrones + SabrinaBaldo de Brébisson + 1–10 + In this submission, I would like to share my experiences with the software DeepL and the comparison analysis I have made with human subtitling offered by the DVD version of the corpus I have chosen as the topic of my study – the eight Seasons of Game of Thrones. The idea is to study if the version proposed by an automatic translation program could be used as a first draft for the professional subtitler. It is expected that the latter would work on the form of the subtitles, that is to say mainly on their length, in a second step. + W19-8701 + 10.26615/issn.2683-0078.2019_001 + + + Parallel Corpus of <fixed-case>C</fixed-case>roatian-<fixed-case>I</fixed-case>talian Administrative Texts + MarijaBrkic Bakaric + IvanaLalli Pacelat + 11–18 + Parallel corpora constitute a unique re-source for providing assistance to human translators. The selection and preparation of the parallel corpora also conditions the quality of the resulting MT engine. Since Croatian is a national language and Italian is officially recognized as a minority lan-guage in seven cities and twelve munici-palities of Istria County, a large amount of parallel texts is produced on a daily basis. However, there have been no attempts in using these texts for compiling a parallel corpus. A domain-specific sentence-aligned parallel Croatian-Italian corpus of administrative texts would be of high value in creating different language tools and resources. The aim of this paper is, therefore, to explore the value of parallel documents which are publicly available mostly in pdf format and to investigate the use of automatically-built dictionaries in corpus compilation. The effects that a document format and, consequently sentence splitting, and the dictionary input have on the sentence alignment process are manually evaluated. + W19-8702 + 10.26615/issn.2683-0078.2019_002 + + + What Influences the Features of Post-editese? A Preliminary Study + SheilaCastilho + NatáliaResende + RuslanMitkov + 19–27 + While a number of studies have shown evidence of translationese phenomena, that is, statistical differences between original texts and translated texts (Gellerstam, 1986), results of studies searching for translationese features in postedited texts (what has been called ”posteditese” (Daems et al., 2017)) have presented mixed results. This paper reports a preliminary study aimed at identifying the presence of post-editese features in machine-translated post-edited texts and at understanding how they differ from translationese features. We test the influence of factors such as post-editing (PE) levels (full vs. light), translation proficiency (professionals vs. students) and text domain (news vs. literary). Results show evidence of post-editese features, especially in light PE texts and in certain domains. + W19-8703 + 10.26615/issn.2683-0078.2019_003 + + + Designing a Frame-Semantic Machine Translation Evaluation Metric + OliverCzulo + TiagoTorrent + ElyMatos + Alexandre Diniz daCosta + DebanjanaKar + 28–35 + We propose a metric for machine translation evaluation based on frame semantics which does not require the use of reference translations or human corrections, but is aimed at comparing original and translated output directly. The metrics is described on the basis of an existing manual frame-semantic annotation of a parallel corpus with an English original and a Brazilian Portuguese and a German translation. We discuss implications of our metrics design, including the potential of scaling it for multiple languages. + W19-8704 + 10.26615/issn.2683-0078.2019_004 + + + Human Evaluation of Neural Machine Translation: The Case of Deep Learning + MarieEscribe + 36–46 + Recent advances in artificial neural networks now have a great impact on translation technology. A considerable achievement was reached in this field with the publication of L’Apprentissage Profond. This book, originally written in English (Deep Learning), was entirely machine-translated into French and post-edited by several experts. In this context, it appears essential to have a clear vision of the performance of MT tools. Providing an evaluation of NMT is precisely the aim of the present research paper. To accomplish this objective, a framework for error categorisation was built and a comparative analysis of the raw translation output and the post-edited version was performed with the purpose of identifying recurring patterns of errors. The findings showed that even though some grammatical errors were spotted, the output was generally correct from a linguistic point of view. The most recurring errors are linked to the specialised terminology employed in this book. Further errors include parts of text that were not translated as well as edits based on stylistic preferences. The major part of the output was not acceptable as such and required several edits per segment, but some sentences were of publishable quality and were therefore left untouched in the final version. + W19-8705 + 10.26615/issn.2683-0078.2019_005 + + + Translationese Features as Indicators of Quality in <fixed-case>E</fixed-case>nglish-<fixed-case>R</fixed-case>ussian Human Translation + MariaKunilovskaya + EkaterinaLapshinova-Koltunski + 47–56 + We use a range of morpho-syntactic features inspired by research in register studies (e.g. Biber, 1995; Neumann, 2013) and translation studies (e.g. Ilisei et al., 2010; Zanettin, 2013; Kunilovskaya and Kutuzov, 2018) to reveal the association between translationese and human translation quality. Translationese is understood as any statistical deviations of translations from non-translations (Baker, 1993) and is assumed to affect the fluency of translations, rendering them foreign-sounding and clumsy of wording and structure. This connection is often posited or implied in the studies of translationese or translational varieties (De Sutter et al., 2017), but is rarely directly tested. Our 45 features include frequencies of selected morphological forms and categories, some types of syntactic structures and relations, as well as several overall text measures extracted from Universal Dependencies annotation. The research corpora include English-to-Russian professional and student translations of informational or argumentative newspaper texts and a comparable corpus of non-translated Russian. Our results indicate lack of direct association between translationese and quality in our data: while our features distinguish translations and non-translations with the near perfect accuracy, the performance of the same algorithm on the quality classes barely exceeds the chance level. + W19-8706 + 10.26615/issn.2683-0078.2019_006 + + + The Punster’s Amanuensis: The Proper Place of Humans and Machines in the Translation of Wordplay + TristanMiller + 57–65 + The translation of wordplay is one of the most extensively researched problems in translation studies, but it has attracted little attention in the fields of natural language processing and machine translation. This is because today’s language technologies treat anomalies and ambiguities in the input as things that must be resolved in favour of a single “correct” interpretation, rather than preserved and interpreted in their own right. But if computers cannot yet process such creative language on their own, can they at least provide specialized support to translation professionals? In this paper, I survey the state of the art relevant to computational processing of humorous wordplay and put forth a vision of how existing theories, resources, and technologies could be adapted and extended to support interactive, computer-assisted translation. + W19-8707 + 10.26615/issn.2683-0078.2019_007 + + + Comparing a Hand-crafted to an Automatically Generated Feature Set for Deep Learning: Pairwise Translation Evaluation + DespoinaMouratidis + Katia LidaKermanidis + 66–74 + The automatic evaluation of machine translation (MT) has proven to be a very significant research topic. Most automatic evaluation methods focus on the evaluation of the output of MT as they compute similarity scores that represent translation quality. This work targets on the performance of MT evaluation. We present a general scheme for learning to classify parallel translations, using linguistic information, of two MT model outputs and one human (reference) translation. We present three experiments to this scheme using neural networks (NN). One using string based hand-crafted features (Exp1), the second using automatically trained embeddings from the reference and the two MT outputs (one from a statistical machine translation (SMT) model and the other from a neural ma-chine translation (NMT) model), which are learned using NN (Exp2), and the third experiment (Exp3) that combines information from the other two experiments. The languages involved are English (EN), Greek (GR) and Italian (IT) segments are educational in domain. The proposed language-independent learning scheme which combines information from the two experiments (experiment 3) achieves higher classification accuracy compared with models using BLEU score information as well as other classification approaches, such as Random Forest (RF) and Support Vector Machine (SVM). + W19-8708 + 10.26615/issn.2683-0078.2019_008 + + + Differences between <fixed-case>SMT</fixed-case> and <fixed-case>NMT</fixed-case> Output - a Translators’ Point of View + JonathanMutal + LiseVolkart + PierretteBouillon + SabrinaGirletti + PaulaEstrella + 75–81 + In this study, we compare the output quality of two MT systems, a statistical (SMT) and a neural (NMT) engine, customised for Swiss Post’s Language Service using the same training data. We focus on the point of view of professional translators and investigate how they perceive the differences between the MT output and a human reference (namely deletions, substitutions, insertions and word order). Our findings show that translators more frequently consider these differences to be errors in SMT than NMT, and that deletions are the most serious errors in both architectures. We also observe lower agreement on differences to be corrected in NMT than in SMT, suggesting that errors are easier to identify in SMT. These findings confirm the ability of NMT to produce correct paraphrases, which could also explain why BLEU is often considered as an inadequate metric to evaluate the performance of NMT systems. + W19-8709 + 10.26615/issn.2683-0078.2019_009 + + + The <fixed-case>C</fixed-case>hinese/<fixed-case>E</fixed-case>nglish Political Interpreting Corpus (<fixed-case>CEPIC</fixed-case>): A New Electronic Resource for Translators and Interpreters + JunPan + 82–88 + The Chinese/English Political Interpreting Corpus (CEPIC) is a new electronic and open access resource developed for translators and interpreters, especially those working with political text types. Over 6 million word tokens in size, the online corpus consists of transcripts of Chinese (Cantonese & Putonghua) / English political speeches and their translated and interpreted texts. It includes rich meta-data and is POS-tagged and annotated with prosodic and paralinguistic features that are of concern to spoken language and interpreting. The online platform of the CEPIC features main functions including Keyword Search, Word Collocation and Expanded Keyword in Context, which are illustrated in the paper. The CEPIC can shed light on online translation and interpreting corpora development in the future. + W19-8710 + 10.26615/issn.2683-0078.2019_010 + + + Translation Quality Assessment Tools and Processes in Relation to <fixed-case>CAT</fixed-case> Tools + ViktoriyaPetrova + 89–97 + Modern translation QA tools are the latest attempt to overcome the inevitable subjective component of human revisers. This paper analyzes the current situation in the translation industry in respect to those tools and their relationship with CAT tools. The adoption of international standards has set the basic frame that defines “quality”. Because of the clear impossibility to develop a universal QA tool, all of the existing ones have in common a wide variety of settings for the user to choose from. A brief comparison is made between most popular standalone QA tools. In order to verify their results in practice, QA outputs from two of those tools have been compared. Polls that cover a period of 12 years have been collected. Their participants explained what practices they adopted in order to guarantee quality. + W19-8711 + 10.26615/issn.2683-0078.2019_011 + + + Corpus Linguistics, Translation and Error Analysis + MariaStambolieva + 98–104 + The paper presents a study of the French Imparfait and its functional equivalents in Bulgarian and English in view of applications in machine translation and error analysis. The aims of the study are: 1/ based on the analysis of a corpus of text, to validate/revise earlier research on the values of the French Imparfait, 2/ to define the contextual factors pointing to the realisation of one or another value of the forms, 3/ based on the analysis of aligned translations, to identify the translation equivalents of these values, 4/ to formulate translation rules, 5/ based on the analysis of the translation rules, to refine the annotation modules of the environment used – the NBU e-Platform for language teaching and research. + W19-8712 + 10.26615/issn.2683-0078.2019_012 + + + Human-Informed Speakers and Interpreters Analysis in the <fixed-case>WAW</fixed-case> Corpus and an Automatic Method for Calculating Interpreters’ Décalage + IrinaTemnikova + AhmedAbdelali + SouhilaDjabri + SamyHedaya + 105–115 + This article presents a multi-faceted analysis of a subset of interpreted conference speeches from the WAW corpus for the English-Arabic language pair. We analyze several speakers and interpreters variables via manual annotation and automatic methods. We propose a new automatic method for calculating interpreters’ décalage based on Automatic Speech Recognition (ASR) and automatic alignment of named entities and content words between speaker and interpreter. The method is evaluated by two human annotators who have expertise in interpreting and Interpreting Studies and shows highly satisfactory results, accompanied with a high inter-annotator agreement. We provide insights about the relations of speakers’ variables, interpreters’ variables and décalage and discuss them from Interpreting Studies and interpreting practice point of view. We had interesting findings about interpreters behavior which need to be extended to a large number of conference sessions in our future research. + W19-8713 + 10.26615/issn.2683-0078.2019_013 + + + Towards a Proactive <fixed-case>MWE</fixed-case> Terminological Platform for Cross-Lingual Mediation in the Age of Big Data + Benjamin K.Tsou + KapoChow + JUNRUNie + YuanYuan + 116–121 + The emergence of China as a global economic power in the 21st Century has brought about surging needs for cross-lingual and cross-cultural mediation, typically performed by translators. Advances in Artificial Intelligence and Language Engineering have been bolstered by Machine learning and suitable Big Data cultivation. They have helped to meet some of the translator’s needs, though the technical specialists have not kept pace with the practical and expanding requirements in language mediation. One major technical and linguistic hurdle involves words outside the vocabulary of the translator or the lexical database he/she consults, especially Multi- Word Expressions (Compound Words) in technical subjects. A further problem is in the multiplicity of renditions of a term in the target language. This paper discusses a proactive approach following the successful extraction and application of sizable bilingual Multi-Word Expressions (Compound Words) for language mediation in technical subjects, which do not fall within the expertise of typical translators, who have inadequate appreciation of the range of new technical tools available to help him/her. Our approach draws on the personal reflections of translators and teachers of translation and is based on the prior R&D efforts relating to 300,000 comparable Chinese-English patents. The subsequent protocol we have developed aims to be proactive in meeting four identified practical challenges in technical translation (e.g. patents). It has broader economic implication in the Age of Big Data (Tsou et al, 2015) and Trade War, as the workload, if not, the challenges, increasingly cannot be met by currently available front-line translators. We shall demonstrate how new tools can be harnessed to spearhead the application of language technology not only in language mediation but also in the “teaching” and “learning” of translation. It shows how a better appreciation of their needs may enhance the contributions of the technical specialists, and thus enhance the resultant synergetic benefits. + W19-8714 + 10.26615/issn.2683-0078.2019_014 + + + Exploring Adequacy Errors in Neural Machine Translation with the Help of Cross-Language Aligned Word Embeddings + MichaelUstaszewski + 122–128 + Neural machine translation (NMT) was shown to produce more fluent output than phrase-based statistical (PBMT) and rule-based machine translation (RBMT). However, improved fluency makes it more difficult for post editors to identify and correct adequacy errors, because unlike RBMT and SMT, in NMT adequacy errors are frequently not anticipated by fluency errors. Omissions and additions of content in otherwise flawlessly fluent NMT output are the most prominent types of such adequacy errors, which can only be detected with reference to source texts. This contribution explores the degree of semantic similarity between source texts, NMT output and post edited output. In this way, computational semantic similarity scores (cosine similarity) are related to human quality judgments. The analyses are based on publicly available NMT post editing data annotated for errors in three language pairs (EN-DE, EN-LV, EN-HR) with the Multidimensional Quality Metrics (MQM). Methodologically, this contribution tests whether cross-language aligned word embeddings as the sole source of semantic information mirror human error annotation. + W19-8715 + 10.26615/issn.2683-0078.2019_015 + + + The Success Story of Mitra Translations + MinaIlieva + MariyaKancheva + 129–133 + Technologies and their constant updates and innovative nature drastically and irreversibly transformed this small business into a leading brand on the translation market, along with just few other LSPs integrating translation software solutions. Now, we are constantly following the new developments in software updates and online platforms and we are successfully keeping up with any new trend in the field of translation, localization, transcreation, revision, post-editing, etc. Ultimately, we are positive that proper implementation of technology (with focus on quality, cost and time) and hard work are the stepping stones in the way to become a trusted translation services provider. + W19-8716 + 10.26615/issn.2683-0078.2019_016 + + + The Four Stages of Machine Translation Acceptance in a Freelancer’s Life + MariaSgourou + 134–135 + Technology is a big challenge and raises many questions and issues when it comes to its application in the translation process, but translation’s biggest problem is not technology; it is rather how technology is perceived by translators. MT developers and researchers should take into account this perception and move towards a more democratized approach to include the base of the translation industry and perhaps its more valuable asset, the translators. + W19-8717 + 10.26615/issn.2683-0078.2019_017 + + + Optimising the Machine Translation Post-editing Workflow + AnnaZaretskaya + 136–139 + In this article, we describe how machine translation is used for post-editing at TransPerfect and the ways in which we optimise the workflow. This includes MT evaluation, MT engine customisation, leveraging MT suggestions compared to TM matches, and the lessons learnt from implementing MT at a large scale. + W19-8718 + 10.26615/issn.2683-0078.2019_018 + +
+ + + Proceedings of the Workshop MultiLing 2019: Summarization Across Languages, Genres and Sources + W19-89 + GeorgeGiannakopoulos + INCOMA Ltd. +
Varna, Bulgaria
+ September + 2019 + + + W19-8900 + + + <fixed-case>RANLP</fixed-case> 2019 Multilingual Headline Generation Task Overview + MarinaLitvak + John M.Conroy + Peter A.Rankel + 1–5 + The objective of the 2019 RANLP Multilingual Headline Generation (HG) Task is to explore some of the challenges highlighted by current state of the art approaches on creating informative headlines to news articles: non-descriptive headlines, out-of-domain training data, generating headlines from long documents which are not well represented by the head heuristic, and dealing with multilingual domain. This tasks makes available a large set of training data for headline generation and provides an evaluation methods for the task. Our data sets are drawn from Wikinews as well as Wikipedia. Participants were required to generate headlines for at least 3 languages, which were evaluated via automatic methods. A key aspect of the task is multilinguality. The task measures the performance of multilingual headline generation systems using the Wikipedia and Wikinews articles in multiple languages. The objective is to assess the performance of automatic head- line generation techniques on text documents covering a diverse range of languages and topics outside the news domain. + W19-8901 + 10.26615/978-954-452-058-8_001 + + + <fixed-case>M</fixed-case>ulti<fixed-case>L</fixed-case>ing 2019: Financial Narrative Summarisation + MahmoudEl-Haj + 6–10 + The Financial Narrative Summarisation task at MultiLing 2019 aims to demonstrate the value and challenges of applying automatic text summarisation to financial text written in English, usually referred to as financial narrative disclosures. The task dataset has been extracted from UK annual reports published in PDF file format. The participants were asked to provide structured summaries, based on real-world, publicly available financial an- nual reports of UK firms by extracting information from different key sections. Partici- pants were asked to generate summaries that reflects the analysis and assessment of the financial trend of the business over the past year, as provided by annual reports. The evaluation of the summaries was performed using AutoSummENG and Rouge automatic metrics. This paper focuses mainly on the data creation process. + W19-8902 + 10.26615/978-954-452-058-8_002 + + + The Summary Evaluation Task in the <fixed-case>M</fixed-case>ulti<fixed-case>L</fixed-case>ing - <fixed-case>RANLP</fixed-case> 2019 Workshop + GeorgeGiannakopoulos + NikiforosPittaras + 11–16 + This report covers the summarization evaluation task, proposed to the summarization community via the MultiLing 2019 Workshop of the RANLP 2019 conference. The task aims to encourage the development of automatic summarization evaluation methods closely aligned with manual, human-authored summary grades and judgements. A multilingual setting is adopted, building upon a corpus of Wikinews articles across 6 languages (English, Arabic, Romanian, Greek, Spanish and Czech). The evaluation utilizes human (golden) and machine-generated (peer) summaries, which have been assigned human evaluation scores from previous MultiLing tasks. Using these resources, the original corpus is augmented with synthetic data, combining summary texts under three different strategies (reorder, merge and replace), each engineered to introduce noise in the summary in a controlled and quantifiable way. We estimate that the utilization of such data can extract and highlight useful attributes of summary quality estimation, aiding the creation of data-driven automatic methods with an increased correlation to human summary evaluations across domains and languages. This paper provides a brief description of the summary evaluation task, the data generation protocol and the resources made available by the MultiLing community, towards improving automatic summarization evaluation. + W19-8903 + 10.26615/978-954-452-058-8_003 + + + Multi-lingual <fixed-case>W</fixed-case>ikipedia Summarization and Title Generation On Low Resource Corpus + WeiLiu + LeiLi + ZuyingHuang + YinanLiu + 17–25 + MultiLing 2019 Headline Generation Task on Wikipedia Corpus raised a critical and practical problem: multilingual task on low resource corpus. In this paper we proposed QDAS extractive summarization model enhanced by sentence2vec and try to apply transfer learning based on large multilingual pre-trained language model for Wikipedia Headline Generation task. We treat it as sequence labeling task and develop two schemes to handle with it. Experimental results have shown that large pre-trained model can effectively utilize learned knowledge to extract certain phrase using low resource supervised data. + W19-8904 + 10.26615/978-954-452-058-8_004 + + + A topic-based sentence representation for extractive text summarization + NikolaosGialitsis + NikiforosPittaras + PanagiotisStamatopoulos + 26–34 + In this study, we examine the effect of probabilistic topic model-based word representations, on sentence-based extractive summarization. We formulate the task of summary extraction as a binary classification problem, and we test a variety of machine learning algorithms, exploring a range of different settings. An wide experimental evaluation on the MultiLing 2015 MSS dataset illustrates that topic-based representations can prove beneficial to the extractive summarization process in terms of F1, ROUGE-L and ROUGE-W scores, compared to a TF-IDF baseline, with QDA-based analysis providing the best results. + W19-8905 + 10.26615/978-954-452-058-8_005 + + + A Study on Game Review Summarization + GeorgePanagiotopoulos + GeorgeGiannakopoulos + AntoniosLiapis + 35–43 + Game reviews have constituted a unique means of interaction between players and companies for many years. The dynamics appearing through online publishing have significantly grown the number of comments per game, giving rise to very interesting communities. The growth has, in turn, led to a difficulty in dealing with the volume and varying quality of the comments as a source of information. This work studies whether and how game reviews can be summarized, based on the notions pre-existing in aspect-based summarization and sentiment analysis. The work provides suggested pipeline of analysis, also offering preliminary findings on whether aspects detected in a set of comments can be consistently evaluated by human users. + W19-8906 + 10.26615/978-954-452-058-8_006 + + + Social Web Observatory: An entity-driven, holistic information summarization platform across sources + LeonidasTsekouras + GeorgiosPetasis + ArisKosmopoulos + 44–52 + The Social Web Observatory is an entity-driven, sentiment-aware, event summarization web platform, combining various methods and tools to overview trends across social media and news sources in Greek. SWO crawls, clusters and summarizes information following an entity-centric view of text streams, allowing to monitor the public sentiment towards a specific person, organization or other entity. In this paper, we overview the platform, outline the analysis pipeline and describe a user study aimed to quantify the usefulness of the system and especially the meaningfulness and coherence of discovered events. + W19-8907 + 10.26615/978-954-452-058-8_007 + + + <fixed-case>EASY</fixed-case>-M: Evaluation System for Multilingual Summarizers + 53–62 + Automatic text summarization aims at producing a shorter version of a document (or a document set). Evaluation of summarization quality is a challenging task. Because human evaluations are expensive and evaluators often disagree between themselves, many researchers prefer to evaluate their systems automatically, with help of software tools. Such a tool usually requires a point of reference in the form of one or more human-written summaries for each text in the corpus. Then, a system-generated summary is compared to one or more human-written summaries, according to selected metrics. However, a single metric cannot reflect all quality-related aspects of a summary. In this paper we present the EvAluation SYstem for Multilingual Summarization (EASY-M), which enables the evaluation of system-generated summaries in 17 different languages with several quality measures, based on comparison with their human-generated counterparts. The system also provides comparative results with two built-in baselines. The source code and both online and offline versions of EASY-M is freely available for the NLP community. + W19-8908 + 10.26615/978-954-452-058-8_008 + + + A study of semantic augmentation of word embeddings for extractive summarization + NikiforosPittaras + VangelisKarkaletsis + 63–72 + In this study we examine the effect of semantic augmentation approaches on extractive text summarization. Wordnet hypernym relations are used to extract term-frequency concept information, subsequently concatenated to sentence-level representations produced by aggregated deep neural word embeddings. Multiple dimensionality reduction techniques and combination strategies are examined via feature transformation and clustering methods. An experimental evaluation on the MultiLing 2015 MSS dataset illustrates that semantic information can introduce benefits to the extractive summarization process in terms of F1, ROUGE-1 and ROUGE-2 scores, with LSA-based post-processing introducing the largest improvements. + W19-8909 + 10.26615/978-954-452-058-8_009 + + + <fixed-case>HE</fixed-case>v<fixed-case>AS</fixed-case>: Headline Evaluation and Analysis System + MarinaLitvak + NataliaVanetik + ItzhakEretz Kdosha + 73–80 + Automatic headline generation is a subtask of one-line summarization with many reported applications. Evaluation of systems generating headlines is a very challenging and undeveloped area. We introduce the Headline Evaluation and Analysis System (HEvAS) that performs automatic evaluation of systems in terms of a quality of the generated headlines. HEvAS provides two types of metrics– one which measures the informativeness of a headline, and another that measures its readability. The results of evaluation can be compared to the results of baseline methods which are implemented in HEvAS. The system also performs the statistical analysis of the evaluation results and provides different visualization charts. This paper describes all evaluation metrics, baselines, analysis, and architecture, utilized by our system. + W19-8910 + 10.26615/978-954-452-058-8_010 + +
+ + + Proceedings of the Workshop on Language Technology for Digital Historical Archives + W19-90 + University of HamburgCristina Vertan + Bulgarian cdemy of SciencesPetya Osenova + St. Kliment Ohridski Universityof Sofia + St. Kliment Ohridski University of SofiaDimitar Iliev + INCOMA Ltd. +
Varna, Bulgaria
+ September + 2019 + + + W19-9000 + + + Graphemic ambiguous queries on <fixed-case>A</fixed-case>rabic-scripted historical corpora + AliciaGonzález Martínez + 1–2 + + W19-9001 + 10.26615/978-954-452-059-5_001 + + + Word Clustering for Historical Newspapers Analysis + LidiaPivovarova + ElaineZosa + JaniMarjanen + 3–10 + This paper is a part of a collaboration between computer scientists and historians aimed at development of novel tools and methods to improve analysis of historical newspapers. We present a case study of ideological terms ending with -ism suffix in nineteenth century Finnish newspapers. We propose a two-step procedure to trace differences in word usages over time: training of diachronic embeddings on sev- eral time slices and when clustering embeddings of selected words together with their neighbours to obtain historical context. The obtained clusters turn out to be useful for historical studies. The paper also discuss specific difficulties related to development historian-oriented tools. + W19-9002 + 10.26615/978-954-452-059-5_002 + + + Geotagging a Diachronic Corpus of Alpine Texts: Comparing Distinct Approaches to Toponym Recognition + TannonKew + AnastassiaShaitarova + IsabelMeraner + JanisGoldzycher + SimonClematide + MartinVolk + 11–18 + Geotagging historic and cultural texts provides valuable access to heritage data, enabling location-based searching and new geographically related discoveries. In this paper, we describe two distinct approaches to geotagging a variety of fine-grained toponyms in a diachronic corpus of alpine texts. By applying a traditional gazetteer-based approach, aided by a few simple heuristics, we attain strong high-precision annotations. Using the output of this earlier system, we adopt a state-of-the-art neural approach in order to facilitate the detection of new toponyms on the basis of context. Additionally, we present the results of preliminary experiments on integrating a small amount of crowdsourced annotations to improve overall performance of toponym recognition in our heritage corpus. + W19-9003 + 10.26615/978-954-452-059-5_003 + + + Controlled Semi-automatic Annotation of Classical Ethiopic + CristinaVertan + 19–23 + Preservation of the cultural heritage by means of digital methods became ex-tremely popular during last years. After intensive digitization campaigns the fo-cus moves slowly from the genuine preservation (i.e digital archiving togeth-er with standard search mechanisms) to research-oriented usage of materials available electronically. This usage is in-tended to go far beyond simple reading of digitized materials; researchers should be able to gain new insigts in materials, discover new facts by means of tools rely-ing on innovative algorithms.In this arti-cle we will describe the workflow neces-sary for the annotation of a dichronic corpus of classical Ethiopic, language of essential importance for the study of Ear-ly Christianity + W19-9004 + 10.26615/978-954-452-059-5_004 + + + Implementing an archival, multilingual and Semantic Web-compliant taxonomy by means of <fixed-case>SKOS</fixed-case> (Simple Knowledge Organization System) + FrancescoGelati + 24–27 + The paper shows how a multilingual hierarchical thesaurus, or taxonomy, can be created and implemented in compliance with Semantic Web requirements by means of the data model SKOS (Simple Knowledge Organization System). It takes the EHRI (European Holocaust Research Infrastructure) portal as an example, and shows how open-source software like SKOS Play! can facilitate the task. + W19-9005 + 10.26615/978-954-452-059-5_005 + + + <fixed-case>EU</fixed-case> 4 U: An educational platform for the cultural heritage of the <fixed-case>EU</fixed-case> + MariaStambolieva + 28–33 + The paper presents an ongoing project of the NBU Laboratory for Language Technology aiming to create a multilingual, CEFR-graded electronic didactic resource for online learning, centered on the history and cultural heritage of the EU (e-EULearn). The resource is developed within the e-Platform of the NBU Laboratory for Language Technology and re-uses the rich corpus of educational material created at the Laboratory for the needs of NBU program modules, distance and blended learning language courses and other projects. Focus being not just on foreign language tuition, but above all on people, places and events in the history and culture of the EU member states, the annotation modules of the e-Platform have been accordingly extended. Current and upcoming activities are directed at: 1/ enriching the English corpus of didactic materials on EU history and culture, 2/ translating the texts into (the) other official EU languages and aligning the translations with the English texts; 3/ developing new test modules. In the process of developing this resource, a database on important people, places, objects and events in the cultural history of the EU will be created. + W19-9006 + 10.26615/978-954-452-059-5_006 + + + Modelling linguistic vagueness and uncertainty in historical texts + CristinaVertan + 34–38 + Many applications in Digital Humanities (DH) rely on annotations of the raw mate-rial. These annotations (inferred automat-ically or done manually) assume that la-belled facts are either true or false, thus all inferences started on such annotations us boolean logic. This contradicts her-meneutic principles used by humanites in which most part of the knowledge has a degree of truth which varies depending on the experience and the world knowledge of the interpreter. In this pa-per we will show how uncertainty and vagueness, two main features of any his-torical text can be encoded in annota-tions and thus be considered by DH ap-plications. + W19-9007 + 10.26615/978-954-452-059-5_007 + +
diff --git a/data/yaml/venues_joint_map.yaml b/data/yaml/venues_joint_map.yaml index 372d457120..bcea414e19 100644 --- a/data/yaml/venues_joint_map.yaml +++ b/data/yaml/venues_joint_map.yaml @@ -727,3 +727,6 @@ D19-65: [WS,DiscoMT] D19-66: WS D19-67: WS D19-68: WS +W19-87: RANLP +W19-89: RANLP +W19-90: RANLP