diff --git a/data/xml/2024.aiwolfdial.xml b/data/xml/2024.aiwolfdial.xml new file mode 100644 index 0000000000..59d1c8234b --- /dev/null +++ b/data/xml/2024.aiwolfdial.xml @@ -0,0 +1,104 @@ + + + + + Proceedings of the 2nd International AIWolfDial Workshop + YoshinobuKano + Association for Computational Linguistics +
Tokyo, Japan
+ September + 2024 + 2024.aiwolfdial-1 + aiwolfdial + ws + + + 2024.aiwolfdial-1.0 + aiwolfdial-2024-1 + + + <fixed-case>AIW</fixed-case>olf<fixed-case>D</fixed-case>ial 2024: Summary of Natural Language Division of 6th International <fixed-case>AIW</fixed-case>olf Contest + YoshinobuKano + YutoSahashi + NeoWatanabe + KaitoKagaminuma + ClausAranha + DaisukeKatagami + KeiHarada + MichimasaInaba + TakeshiIto + HirotakaOsawa + TakashiOtsuki + FujioToriumi + 1–12 + We held our 6th annual AIWolf international contest to automatically play the Werewolf game “Mafia”, where players try finding liars via conversations, aiming at promoting developments in creating agents of more natural conversations in higher level, such as longer contexts, personal relationships, semantics, pragmatics, and logics, revealing the capabilities and limits of the generative AIs. In our Natural Language Division of the contest, we had eight Japanese speaking agent teams, and five English speaking agents, to mutually run games. By using the game logs, we performed human subjective evaluations, win rates, and detailed log analysis. We found that the entire system performance has largely improved over the previous year, due to the recent advantages of the LLMs. There are several new ideas to improve the way using LLMs such as the summarization, characterization, and the logics outside LLMs, etc. However, it is not perfect at all yet; the generated talks are sometimes inconsistent with the game actions. Our future work includes to reveal the capability of the LLMs, whether they can make the duality of the “liar”, in other words, holding a “true” and a “false” circumstances of the agent at the same time, even holding what these circumstances look like from other agents. + 2024.aiwolfdial-1.1 + kano-etal-2024-aiwolfdial + + + Text Generation Indistinguishable from Target Person by Prompting Few Examples Using <fixed-case>LLM</fixed-case> + YukaTsubota + YoshinobuKano + 13–20 + To achieve smooth and natural communication between a dialogue system and a human, it is necessary for the dialogue system to behave more human-like. Recreating the personality of an actual person can be an effective way for this purpose. This study proposes a method to recreate a personality by a large language model (generative AI) without training, but with prompt technique to make the creation cost as low as possible. Collecting a large amount of dialogue data from a specific person is not easy and requires a significant amount of time for training. Therefore, we aim to recreate the personality of a specific individual without using dialogue data. The personality referred to in this paper denotes the image of a person that can be determined solely from the input and output of text dialogues. As a result of the experiments, it was revealed that by using prompts combining profile information, responses to few questions, and extracted speaking characteristics from those responses, it is possible to improve the reproducibility of a specific individual’s personality. + 2024.aiwolfdial-1.2 + tsubota-kano-2024-text + + + Werewolf Game Agent by Generative <fixed-case>AI</fixed-case> Incorporating Logical Information Between Players + NeoWatanabe + YoshinobuKano + 21–29 + In recent years, AI models based on GPT have advanced rapidly. These models are capable of generating text, translating between different languages, and answering questions with high accuracy. However, the process behind their outputs remains a black box, making it difficult to ascertain the data influencing their responses. These AI models do not always produce accurate outputs and are known for generating incorrect information, known as hallucinations, whose causes are hard to pinpoint. Moreover, they still face challenges in solving complex problems that require step-by-step reasoning, despite various improvements like the Chain-of-Thought approach. There’s no guarantee that these models can independently perform logical reasoning from scratch, raising doubts about the reliability and accuracy of their inferences. To address these concerns, this study proposes the incorporation of an explicit logical structure into the AI’s text generation process. As a validation experiment, a text-based agent capable of playing the Werewolf game, which requires deductive reasoning, was developed using GPT-4. By comparing the model combined with an external explicit logical structure and a baseline that lacks such a structure, the proposed method demonstrated superior reasoning capabilities in subjective evaluations, suggesting the effectiveness of adding an explicit logical framework to the conventional AI models. + 2024.aiwolfdial-1.3 + watanabe-kano-2024-werewolf + + + Enhancing Dialogue Generation in Werewolf Game Through Situation Analysis and Persuasion Strategies + ZhiyangQi + MichimasaInaba + 30–39 + Recent advancements in natural language processing, particularly with large language models (LLMs) like GPT-4, have significantly enhanced dialogue systems, enabling them to generate more natural and fluent conversations. Despite these improvements, challenges persist, such as managing continuous dialogues, memory retention, and minimizing hallucinations. The AIWolfDial2024 addresses these challenges by employing the Werewolf Game, an incomplete information game, to test the capabilities of LLMs in complex interactive environments. This paper introduces a LLM-based Werewolf Game AI, where each role is supported by situation analysis to aid response generation. Additionally, for the werewolf role, various persuasion strategies, including logical appeal, credibility appeal, and emotional appeal, are employed to effectively persuade other players to align with its actions. + 2024.aiwolfdial-1.4 + qi-inaba-2024-enhancing + + + Verification of Reasoning Ability using <fixed-case>BDI</fixed-case> Logic and Large Language Model in <fixed-case>AIW</fixed-case>olf + HirakuGondo + HirokiSakaji + ItsukiNoda + 40–47 + We attempt to improve the reasoning capability of LLMs in werewolf game by combining BDI logic with LLMs. While LLMs such as ChatGPT has been developed and used for various tasks, there remain several weakness of the LLMs. Logical reasoning is one of such weakness. Therefore, we try to introduce BDI logic-based prompts to verify the logical reasoning ability of LLMs in dialogue of werewofl game. Experiments and evaluations were conducted using “AI-Werewolf,” a communication game for AI with incomplete information. From the results of the game played by five agents, we compare the logical reasoning ability of LLMs by using the win rate and the vote rate against werewolf. + 2024.aiwolfdial-1.5 + 2024.aiwolfdial-1.5.Supplementary_Attachment.zip + gondo-etal-2024-verification + + + Enhancing Consistency of Werewolf <fixed-case>AI</fixed-case> through Dialogue Summarization and Persona Information + YoshikiTanaka + TakumasaKaneko + HirokiOnozeki + NatsumiEzure + RyuichiUehara + ZhiyangQi + TomoyaHiguchi + RyutaroAsahara + MichimasaInaba + 48–57 + The Werewolf Game is a communication game where players’ reasoning and discussion skills are essential. In this study, we present a Werewolf AI agent developed for the AIWolfDial 2024 shared task, co-hosted with the 17th INLG. In recent years, large language models like ChatGPT have garnered attention for their exceptional response generation and reasoning capabilities. We thus develop the LLM-based agents for the Werewolf Game. This study aims to enhance the consistency of the agent’s utterances by utilizing dialogue summaries generated by LLMs and manually designed personas and utterance examples. By analyzing self-match game logs, we demonstrate that the agent’s utterances are contextually consistent and that the character, including tone, is maintained throughout the game. + 2024.aiwolfdial-1.6 + tanaka-etal-2024-enhancing + + + An Implementation of Werewolf Agent That does not Truly Trust <fixed-case>LLM</fixed-case>s + TakehiroSato + ShintaroOzaki + DaisakuYokoyama + 58–67 + Werewolf is an incomplete information game, which has several challenges when creating a computer agent as a player given the lack of understanding of the situation and individuality of utterance (e.g., computer agents are not capable of characterful utterance or situational lying). We propose a werewolf agent that solves some of those difficulties by combining a Large Language Model (LLM) and a rule-based algorithm. In particular, our agent uses a rule-based algorithm to select an output either from an LLM or a template prepared beforehand based on the results of analyzing conversation history using an LLM. It allows the agent to refute in specific situations, identify when to end the conversation, and behave with persona. This approach mitigated conversational inconsistencies and facilitated logical utterance as a result. We also conducted a qualitative evaluation, which resulted in our agent being perceived as more human-like compared to an unmodified LLM. The agent is freely available for contributing to advance the research in the field of Werewolf game. + 2024.aiwolfdial-1.7 + 2024.aiwolfdial-1.7.Supplementary_Attachment.pdf + sato-etal-2024-implementation + +
+
diff --git a/data/xml/2024.inlg.xml b/data/xml/2024.inlg.xml new file mode 100644 index 0000000000..54c83e5151 --- /dev/null +++ b/data/xml/2024.inlg.xml @@ -0,0 +1,734 @@ + + + + + Proceedings of the 17th International Natural Language Generation Conference + SaadMahamood + Nguyen LeMinh + DaphneIppolito + Association for Computational Linguistics +
Tokyo, Japan
+ September + 2024 + 2024.inlg-main + inlg + + + 2024.inlg-main.0 + inlg-2024-main + + + <fixed-case>A</fixed-case>uto<fixed-case>T</fixed-case>emplate: A Simple Recipe for Lexically Constrained Text Generation + HayateIso + 1–12 + Lexically constrained text generation is one of the constrained text generation tasks, which aims to generate text that covers all the given constraint lexicons. While the existing approaches tackle this problem using a lexically constrained beam search algorithm or dedicated model using non-autoregressive decoding, there is a trade-off between the generated text quality and the hard constraint satisfaction. We introduce AutoTemplate, a simple yet effective lexically constrained text generation framework divided into template generation and lexicalization tasks. The template generation is to generate the text with the placeholders, and lexicalization replaces them into the constraint lexicons to perform lexically constrained text generation. We conducted the experiments on two tasks: keywords-to-sentence generations and entity-guided summarization. Experimental results show that the AutoTemplate outperforms the competitive baselines on both tasks while satisfying the hard lexical constraints. The code is available at https://github.com/megagonlabs/autotemplate + 2024.inlg-main.1 + 2024.inlg-main.1.Supplementary_Attachment.pdf + iso-2024-autotemplate-simple + + + Noisy Pairing and Partial Supervision for Stylized Opinion Summarization + HayateIso + XiaolanWang + YoshiSuhara + 13–23 + Opinion summarization research has primarily focused on generating summaries reflecting important opinions from customer reviews without paying much attention to the writing style. In this paper, we propose the stylized opinion summarization task, which aims to generate a summary of customer reviews in the desired (e.g., professional) writing style. To tackle the difficulty in collecting customer and professional review pairs, we develop a non-parallel training framework, Noisy Pairing and Partial Supervision (NAPA), which trains a stylized opinion summarization system from non-parallel customer and professional review sets. We create a benchmark ProSum by collecting customer and professional reviews from Yelp and Michelin. Experimental results on ProSum and FewSum demonstrate that our non-parallel training framework consistently improves both automatic and human evaluations, successfully building a stylized opinion summarization model that can generate professionally-written summaries from customer reviews. The code is available at https://github.com/megagonlabs/napa + 2024.inlg-main.2 + 2024.inlg-main.2.Supplementary_Attachment.pdf + iso-etal-2024-noisy-pairing + + + <fixed-case>LLM</fixed-case> Neologism: Emergence of Mutated Characters due to Byte Encoding + RanIwamoto + HiroshiKanayama + 24–29 + The process of language generation, which selects the most probable tokens one by one, may intrinsically result in output strings that humans never utter. We name this phenomenon “LLM neologism” and investigate it focusing on Japanese, Chinese, and Korean languages, where tokens can be smaller than characters. Our findings show that LLM neologism occurs through the combination of two high-frequency words with common tokens. We also clarify the cause of LLM neologism in the tokenization process with limited vocabularies. The results of this study provides important clues for better encoding of multibyte characters, aiming to prevent catastrophic results in AI-generated documents. + 2024.inlg-main.3 + iwamoto-kanayama-2024-llm-neologism + + + Communicating Uncertainty in Explanations of the Outcomes of Machine Learning Models + IngridZukerman + SameenMaruf + 30–46 + We consider two types of numeric representations for conveying the uncertainty of predictions made by Machine Learning (ML) models: confidence-based (e.g., “the AI is 90% confident”) and frequency-based (e.g., “the AI was correct in 180 (90%) out of 200 cases”). We conducted a user study to determine which factors influence users’ acceptance of predictions made by ML models, and how the two types of uncertainty representations affect users’ views about explanations. Our results show that users’ acceptance of ML model predictions depends mainly on the models’ confidence, and that explanations that include uncertainty information are deemed better in several respects than explanations that omit it, with frequency-based representations being deemed better than confidence-based representations. + 2024.inlg-main.4 + zukerman-maruf-2024-communicating-uncertainty + + + Entity-aware Multi-task Training Helps Rare Word Machine Translation + MatissRikters + MakotoMiwa + 47–54 + Named entities (NE) are integral for preserving context and conveying accurate information in the machine translation (MT) task. Challenges often lie in handling NE diversity, ambiguity, rarity, and ensuring alignment and consistency. In this paper, we explore the effect of NE-aware model fine-tuning to improve handling of NEs in MT. We generate data for NE recognition (NER) and NE-aware MT using common NER tools from Spacy, and align entities in parallel data. Experiments with fine-tuning variations of pre-trained T5 models on NE-related generation tasks between English and German show promising results with increasing amounts of NEs in the output and BLEU score improvements compared to the non-tuned baselines. + 2024.inlg-main.5 + rikters-miwa-2024-entity-aware + + + <fixed-case>CE</fixed-case>val: A Benchmark for Evaluating Counterfactual Text Generation + Van BachNguyen + ChristinSeifert + JörgSchlötterer + 55–69 + Counterfactual text generation aims to minimally change a text, such that it is classified differently. Assessing progress in method development for counterfactual text generation is hindered by a non-uniform usage of data sets and metrics in related work. We propose CEval, a benchmark for comparing counterfactual text generation methods. CEval unifies counterfactual and text quality metrics, includes common counterfactual datasets with human annotations, standard baselines (MICE, GDBA, CREST) and the open-source language model LLAMA-2. Our experiments found no perfect method for generating counterfactual text. Methods that excel at counterfactual metrics often produce lower-quality text while LLMs with simple prompts generate high-quality text but struggle with counterfactual criteria. By making CEval available as an open-source Python library, we encourage the community to contribute additional methods and maintain consistent evaluation in future work. + 2024.inlg-main.6 + 2024.inlg-main.6.Supplementary_Attachment.pdf + nguyen-etal-2024-ceval-benchmark + + + Generating from <fixed-case>AMR</fixed-case>s into High and Low-Resource Languages using Phylogenetic Knowledge and Hierarchical <fixed-case>QL</fixed-case>o<fixed-case>RA</fixed-case> Training (<fixed-case>HQL</fixed-case>) + William EduardoSoto Martinez + YannickParmentier + ClaireGardent + 70–81 + Multilingual generation from Abstract Meaning Representations (AMRs) verbalises AMRs into multiple languages. Previous work has focused on high- and medium-resource languages relying on large amounts of training data. In this work, we consider both high- and low-resource languages capping training data size at the lower bound set by our low-resource languages i.e. 31K. We propose a straightforward technique to enhance results on low-resource while preserving performance on high-resource languages. We iteratively refine a multilingua model to a set of monolingual models using Low-Rank Adaptation with a training curriculum based on a tree structure; this permits investigating how the languages used at each iteration impact generation performance on high and low-resource languages. We show an improvement over both mono and multilingual approaches. Comparing different ways of grouping languages at each iteration step we find two working configurations: grouping related languages which promotes transfer, or grouping distant languages which facilitates regularisation + 2024.inlg-main.7 + 2024.inlg-main.7.Supplementary_Attachment.pdf + soto-martinez-etal-2024-generating-amrs + + + <fixed-case>AMERICANO</fixed-case>: Argument Generation with Discourse-driven Decomposition and Agent Interaction + ZheHu + Hou PongChan + YuYin + 82–102 + Argument generation is a challenging task in natural language processing, which requires rigorous reasoning and proper content organization. Inspired by recent chain-of-thought prompting that breaks down a complex task into intermediate steps, we propose Americano, a novel framework with agent interaction for argument generation. Our approach decomposes the generation process into sequential actions grounded on argumentation theory, which first executes actions sequentially to generate argumentative discourse components, and then produces a final argument conditioned on the components. To further mimic the human writing process and improve the left-to-right generation paradigm of current autoregressive language models, we introduce an argument refinement module that automatically evaluates and refines argument drafts based on feedback received. We evaluate our framework on the task of counterargument generation using a subset of Reddit/CMV dataset. The results show that our method outperforms both end-to-end and chain-of-thought prompting methods and can generate more coherent and persuasive arguments with diverse and rich contents. + 2024.inlg-main.8 + hu-etal-2024-americano-argument + + + Generating Simple, Conservative and Unifying Explanations for Logistic Regression Models + SameenMaruf + IngridZukerman + XuelinSitu + CecileParis + GholamrezaHaffari + 103–120 + In this paper, we generate and compare three types of explanations of Machine Learning (ML) predictions: simple, conservative and unifying. Simple explanations are concise, conservative explanations address the surprisingness of a prediction, and unifying explanations convey the extent to which an ML model’s predictions are applicable. The results of our user study show that (1) conservative and unifying explanations are liked equally and considered largely equivalent in terms of completeness, helpfulness for understanding the AI, and enticement to act, and both are deemed better than simple explanations; and (2)users’ views about explanations are influenced by the (dis)agreement between the ML model’s predictions and users’ estimations of these predictions, and by the inclusion/omission of features users expect to see in explanations. + 2024.inlg-main.9 + maruf-etal-2024-generating-simple + + + Extractive Summarization via Fine-grained Semantic Tuple Extraction + YubinGe + SullamJeoung + JanaDiesner + 121–133 + Traditional extractive summarization treats the task as sentence-level classification and requires a fixed number of sentences for extraction. However, this rigid constraint on the number of sentences to extract may hinder model generalization due to varied summary lengths across datasets. In this work, we leverage the interrelation between information extraction (IE) and text summarization, and introduce a fine-grained autoregressive method for extractive summarization through semantic tuple extraction. Specifically, we represent each sentence as a set of semantic tuples, where tuples are predicate-argument structures derived from conducting IE. Then we adopt a Transformer-based autoregressive model to extract the tuples corresponding to the target summary given a source document. In inference, a greedy approach is proposed to select source sentences to cover extracted tuples, eliminating the need for a fixed number. Our experiments on CNN/DM and NYT demonstrate the method’s superiority over strong baselines. Through the zero-shot setting for testing the generalization of models to diverse summary lengths across datasets, we further show our method outperforms baselines, including ChatGPT. + 2024.inlg-main.10 + ge-etal-2024-extractive-summarization + + + Evaluating <fixed-case>RDF</fixed-case>-to-text Generation Models for <fixed-case>E</fixed-case>nglish and <fixed-case>R</fixed-case>ussian on Out Of Domain Data + AnnaNikiforovskaya + ClaireGardent + 134–144 + While the WebNLG dataset has prompted much research on generation from knowledge graphs, little work has examined how well models trained on the WebNLG data generalise to unseen data and work has mostly been focused on English. In this paper, we introduce novel benchmarks for both English and Russian which contain various ratios of unseen entities and properties. These benchmarks also differ from WebNLG in that some of the graphs stem from Wikidata rather than DBpedia. Evaluating various models for English and Russian on these benchmarks shows a strong decrease in performance while a qualitative analysis highlights the various types of errors induced by non i.i.d data. + 2024.inlg-main.11 + 2024.inlg-main.11.Supplementary_Attachment.pdf + nikiforovskaya-gardent-2024-evaluating-rdf + + + Forecasting Implicit Emotions Elicited in Conversations + YurieKoga + ShunsukeKando + YusukeMiyao + 145–152 + This paper aims to forecast the implicit emotion elicited in the dialogue partner by a textual input utterance. Forecasting the interlocutor’s emotion is beneficial for natural language generation in dialogue systems to avoid generating utterances that make the users uncomfortable. Previous studies forecast the emotion conveyed in the interlocutor’s response, assuming it will explicitly reflect their elicited emotion. However, true emotions are not always expressed verbally. We propose a new task to directly forecast the implicit emotion elicited by an input utterance, which does not rely on this assumption. We compare this task with related ones to investigate the impact of dialogue history and one’s own utterance on predicting explicit and implicit emotions. Our result highlights the importance of dialogue history for predicting implicit emotions. It also reveals that, unlike explicit emotions, implicit emotions show limited improvement in predictive performance with one’s own utterance, and that they are more difficult to predict than explicit emotions. We find that even a large language model (LLM) struggles to forecast implicit emotions accurately. + 2024.inlg-main.12 + 2024.inlg-main.12.Supplementary_Attachment.pdf + koga-etal-2024-forecasting-implicit + + + <fixed-case>G</fixed-case>erman Voter Personas Can Radicalize <fixed-case>LLM</fixed-case> Chatbots via the Echo Chamber Effect + MaximilianBleick + NilsFeldhus + AljoschaBurchardt + SebastianMöller + 153–164 + We investigate the impact of LLMs on political discourse with a particular focus on the influence of generated personas on model responses. We find an echo chamber effect from LLM chatbots when provided with German-language biographical information of politicians and voters in German politics, leading to sycophantic responses and the reinforcement of existing political biases. Findings reveal that personas of certain political party, such as those of the ‘Alternative für Deutschland’ party, exert a stronger influence on LLMs, potentially amplifying extremist views. Unlike prior studies, we cannot corroborate a tendency for larger models to exert stronger sycophantic behaviour. We propose that further development should aim at reducing sycophantic behaviour in LLMs across all sizes and diversifying language capabilities in LLMs to enhance inclusivity. + 2024.inlg-main.13 + bleick-etal-2024-german-voter + + + Quantifying Memorization and Detecting Training Data of Pre-trained Language Models using <fixed-case>J</fixed-case>apanese Newspaper + ShotaroIshihara + HiromuTakahashi + 165–179 + Dominant pre-trained language models (PLMs) have demonstrated the potential risk of memorizing and outputting the training data. While this concern has been discussed mainly in English, it is also practically important to focus on domain-specific PLMs. In this study, we pre-trained domain-specific GPT-2 models using a limited corpus of Japanese newspaper articles and evaluated their behavior. Experiments replicated the empirical finding that memorization of PLMs is related to the duplication in the training data, model size, and prompt length, in Japanese the same as in previous English studies. Furthermore, we attempted membership inference attacks, demonstrating that the training data can be detected even in Japanese, which is the same trend as in English. The study warns that domain-specific PLMs, sometimes trained with valuable private data, can ”copy and paste” on a large scale. + 2024.inlg-main.14 + ishihara-takahashi-2024-quantifying-memorization + + + Should We Fine-Tune or <fixed-case>RAG</fixed-case>? Evaluating Different Techniques to Adapt <fixed-case>LLM</fixed-case>s for Dialogue + SimoneAlghisi + MassimoRizzoli + GabrielRoccabruna + Seyed MahedMousavi + GiuseppeRiccardi + 180–197 + We study the limitations of Large Language Models (LLMs) for the task of response generation in human-machine dialogue. Several techniques have been proposed in the literature for different dialogue types (e.g., Open-Domain). However, the evaluations of these techniques have been limited in terms of base LLMs, dialogue types and evaluation metrics. In this work, we extensively analyze different LLM adaptation techniques when applied to different dialogue types. We have selected two base LLMs, Llama-2 and Mistral, and four dialogue types Open-Domain, Knowledge-Grounded, Task-Oriented, and Question Answering. We evaluate the performance of in-context learning and fine-tuning techniques across datasets selected for each dialogue type. We assess the impact of incorporating external knowledge to ground the generation in both scenarios of Retrieval-Augmented Generation (RAG) and gold knowledge. We adopt consistent evaluation and explainability criteria for automatic metrics and human evaluation protocols. Our analysis shows that there is no universal best-technique for adapting large language models as the efficacy of each technique depends on both the base LLM and the specific type of dialogue. Last but not least, the assessment of the best adaptation technique should include human evaluation to avoid false expectations and outcomes derived from automatic metrics. + 2024.inlg-main.15 + 2024.inlg-main.15.Supplementary_Attachment.pdf + alghisi-etal-2024-fine-tune + + + Automating True-False Multiple-Choice Question Generation and Evaluation with Retrieval-based Accuracy Differential + Chen-JuiYu + Wen HungLee + Lin TseKe + Shih-WeiGuo + Yao-ChungFan + 198–212 + Creating high-quality True-False (TF) multiple-choice questions (MCQs), with accurate distractors, is a challenging and time-consuming task in education. This paper introduces True-False Distractor Generation (TFDG), a pipeline that leverages pre-trained language models and sentence retrieval techniques to automate the generation of TF-type MCQ distractors. Furthermore, the evaluation of generated TF questions presents a challenge. Traditional metrics like BLEU and ROUGE are unsuitable for this task. To address this, we propose a new evaluation metric called Retrieval-based Accuracy Differential (RAD). RAD assesses the discriminative power of TF questions by comparing model accuracy with and without access to reference texts. It quantitatively evaluates how well questions differentiate between students with varying knowledge levels. This research benefits educators and assessment developers, facilitating the efficient automatic generation of high-quality TF-type MCQs and their reliable evaluation. + 2024.inlg-main.16 + yu-etal-2024-automating-true + + + Transfer-Learning based on Extract, Paraphrase and Compress Models for Neural Abstractive Multi-Document Summarization + YlliasChali + ElozinoEgonmwan + 213–221 + Recently, transfer-learning by unsupervised pre-training and fine-tuning has shown great success on a number of tasks. The paucity of data for multi-document summarization (MDS) in the news domain, especially makes this approach practical. However, while existing literature mostly formulate unsupervised learning objectives tailored for/around the summarization problem we find that MDS can benefit directly from models pre-trained on other downstream supervised tasks such as sentence extraction, paraphrase generation and sentence compression. We carry out experiments to demonstrate the impact of zero-shot transfer-learning from these downstream tasks on MDS. Since it is challenging to train end-to-end encoder-decoder models on MDS due to i) the sheer length of the input documents, and ii) the paucity of training data. We hope this paper encourages more work on these downstream tasks as a means to mitigating the challenges in neural abstractive MDS. + 2024.inlg-main.17 + chali-egonmwan-2024-transfer-learning + + + Enhancing Presentation Slide Generation by <fixed-case>LLM</fixed-case>s with a Multi-Staged End-to-End Approach + SambaranBandyopadhyay + HimanshuMaheshwari + AnandhaveluNatarajan + ApoorvSaxena + 222–229 + Generating presentation slides from a long document with multimodal elements such as text and images is an important task. This is time consuming and needs domain expertise if done manually. Existing approaches for generating a rich presentation from a document are often semi-automatic or only put a flat summary into the slides ignoring the importance of a good narrative. In this paper, we address this research gap by proposing a multi-staged end-to-end model which uses a combination of LLM and VLM. We have experimentally shown that compared to applying LLMs directly with state-of-the-art prompting, our proposed multi-staged solution is better in terms of automated metrics and human evaluation. + 2024.inlg-main.18 + 2024.inlg-main.18.Supplementary_Attachment.pdf + bandyopadhyay-etal-2024-enhancing-presentation + + + Is Machine Psychology here? On Requirements for Using Human Psychological Tests on Large Language Models + LeaLöhn + NiklasKiehne + AlexanderLjapunov + Wolf-TiloBalke + 230–242 + In an effort to better understand the behavior of large language models (LLM), researchers recently turned to conducting psychological assessments on them. Several studies diagnose various psychological concepts in LLMs, such as psychopathological symptoms, personality traits, and intellectual functioning, aiming to unravel their black-box characteristics. But can we safely assess LLMs with tests that were originally designed for humans? The psychology domain looks back on decades of developing standards of appropriate testing procedures to ensure reliable and valid measures. We argue that analogous standardization processes are required for LLM assessments, given their differential functioning as compared to humans. In this paper, we propose seven requirements necessary for testing LLMs. Based on these, we critically reflect a sample of 25 recent machine psychology studies. Our analysis reveals (1) the lack of appropriate methods to assess test reliability and construct validity, (2) the unknown strength of construct-irrelevant influences, such as the contamination of pre-training corpora with test material, and (3) the pervasive issue of non-reproducibility of many studies. The results underscore the lack of a general methodology for the implementation of psychological assessments of LLMs and the need to redefine psychological constructs specifically for large language models rather than adopting them from human psychology. + 2024.inlg-main.19 + lohn-etal-2024-machine-psychology + + + Exploring the impact of data representation on neural data-to-text generation + David M.Howcroft + Lewis N.Watson + OlesiaNedopas + DimitraGkatzia + 243–253 + A relatively under-explored area in research on neural natural language generation is the impact of the data representation on text quality. Here we report experiments on two leading input representations for data-to-text generation: attribute-value pairs and Resource Description Framework (RDF) triples. Evaluating the performance of encoder-decoder seq2seq models as well as recent large language models (LLMs) with both automated metrics and human evaluation, we find that the input representation does not seem to have a large impact on the performance of either purpose-built seq2seq models or LLMs. Finally, we present an error analysis of the texts generated by the LLMs and provide some insights into where these models fail. + 2024.inlg-main.20 + 2024.inlg-main.20.Supplementary_Attachment.pdf + howcroft-etal-2024-exploring-impact + + + Automatically Generating <fixed-case>I</fixed-case>si<fixed-case>Z</fixed-case>ulu Words From <fixed-case>I</fixed-case>ndo-<fixed-case>A</fixed-case>rabic Numerals + ZolaMahlaza + TadiwaMagwenzi + C. MariaKeet + LangaKhumalo + 254–271 + Artificial conversational agents are deployed to assist humans in a variety of tasks. Some of these tasks require the capability to communicate numbers as part of their internal and abstract representations of meaning, such as for banking and scheduling appointments. They currently cannot do so for isiZulu because there are no algorithms to do so due to a lack of speech and text data and the transformation is complex and it may include dependence on the type of noun that is counted. We solved this by extracting and iteratively improving on the rules for speaking and writing numerals as words and creating two algorithms to automate the transformation. Evaluation of the algorithms by two isiZulu grammarians showed that six out of seven number categories were 90-100% correct. The same software was used with an additional set of rules to create a large monolingual text corpus, made up of 771 643 sentences, to enable future data-driven approaches. + 2024.inlg-main.21 + mahlaza-etal-2024-automatically-generating + + + (Mostly) Automatic Experiment Execution for Human Evaluations of <fixed-case>NLP</fixed-case> Systems + CraigThomson + AnyaBelz + 272–279 + Human evaluation is widely considered the most reliable form of evaluation in NLP, but recent research has shown it to be riddled with mistakes, often as a result of manual execution of tasks. This paper argues that such mistakes could be avoided if we were to automate, as much as is practical, the process of performing experiments for human evaluation of NLP systems. We provide a simple methodology that can improve both the transparency and reproducibility of experiments. We show how the sequence of component processes of a human evaluation can be defined in advance, facilitating full or partial automation, detailed preregistration of the process, and research transparency and repeatability. + 2024.inlg-main.22 + thomson-belz-2024-mostly-automatic + + + Generating Hotel Highlights from Unstructured Text using <fixed-case>LLM</fixed-case>s + Srinivas RameshKamath + FahimeSame + SaadMahamood + 280–288 + We describe our implementation and evaluation of the Hotel Highlights system which has been deployed live by trivago. This system leverages a large language model (LLM) to generate a set of highlights from accommodation descriptions and reviews, enabling travellers to quickly understand its unique aspects. In this paper, we discuss our motivation for building this system and the human evaluation we conducted, comparing the generated highlights against the source input to assess the degree of hallucinations and/or contradictions present. Finally, we outline the lessons learned and the improvements needed. + 2024.inlg-main.23 + kamath-etal-2024-generating-hotel + + + <fixed-case>T</fixed-case>ext2<fixed-case>T</fixed-case>raj2<fixed-case>T</fixed-case>ext: Learning-by-Synthesis Framework for Contextual Captioning of Human Movement Trajectories + HikaruAsano + RyoYonetani + TaikiSekii + HirokiOuchi + 289–302 + This paper presents Text2Traj2Text, a novel learning-by-synthesis framework for captioning possible contexts behind shopper’s trajectory data in retail stores. Our work will impact various retail applications that need better customer understanding, such as targeted advertising and inventory management. The key idea is leveraging large language models to synthesize a diverse and realistic collection of contextual captions as well as the corresponding movement trajectories on a store map. Despite learned from fully synthesized data, the captioning model can generalize well to trajectories/captions created by real human subjects. Our systematic evaluation confirmed the effectiveness of the proposed framework over competitive approaches in terms of ROUGE and BERT Score metrics. + 2024.inlg-main.24 + asano-etal-2024-text2traj2text-learning + + + n-gram <fixed-case>F</fixed-case>-score for Evaluating Grammatical Error Correction + ShotaKoyama + RyoNagata + HiroyaTakamura + NaoakiOkazaki + 303–313 + M2 and its variants are the most widely used automatic evaluation metrics for grammatical error correction (GEC), which calculate an F-score using a phrase-based alignment between sentences. However, it is not straightforward at all to align learner sentences containing errors to their correct sentences. In addition, alignment calculations are computationally expensive. We propose GREEN, an alignment-free F-score for GEC evaluation. GREEN treats a sentence as a multiset of n-grams and extracts edits between sentences by set operations instead of computing an alignment. Our experiments confirm that GREEN performs better than existing methods for the corpus-level metrics and comparably for the sentence-level metrics even without computing an alignment. GREEN is available at https://github.com/shotakoyama/green. + 2024.inlg-main.25 + koyama-etal-2024-n-gram + + + Personalized Cloze Test Generation with Large Language Models: Streamlining <fixed-case>MCQ</fixed-case> Development and Enhancing Adaptive Learning + Chih-HsuanShen + Yi-LiKuo + Yao-ChungFan + 314–319 + Cloze multiple-choice questions (MCQs) are essential for assessing comprehension in educational settings, but manually designing effective distractors is time-consuming. Addressing this, recent research has automated distractor generation, yet such methods often neglect to adjust the difficulty level to the learner’s abilities, resulting in non-personalized assessments. This study introduces the Personalized Cloze Test Generation (PCGL) Framework, utilizing Large Language Models (LLMs) to generate cloze tests tailored to individual proficiency levels. Our PCGL Framework simplifies test creation by generating both question stems and distractors from a single input word and adjusts the difficulty to match the learner’s proficiency. The framework significantly reduces the effort in creating tests and enhances personalized learning by dynamically adjusting to the needs of each learner. + 2024.inlg-main.26 + 2024.inlg-main.26.Supplementary_Attachment.pdf + shen-etal-2024-personalized-cloze + + + Pipeline Neural Data-to-text with Large Language Models + Chinonso CynthiaOsuji + BrianTimoney + ThiagoCastro Ferreira + BrianDavis + 320–329 + Previous studies have highlighted the advantages of pipeline neural architectures over end-to-end models, particularly in reducing text hallucination. In this study, we extend prior research by integrating pretrained language models (PLMs) into a pipeline framework, using both fine-tuning and prompting methods. Our findings show that fine-tuned PLMs consistently generate high quality text, especially within end-to-end architectures and at intermediate stages of the pipeline across various domains. These models also outperform prompt-based ones on automatic evaluation metrics but lag in human evaluations. Compared to the standard five-stage pipeline architecture, a streamlined three-stage pipeline, which only include ordering, structuring, and surface realization, achieves superior performance in fluency and semantic adequacy according to the human evaluation. + 2024.inlg-main.27 + osuji-etal-2024-pipeline-neural + + + Reduction-Synthesis: Plug-and-Play for Sentiment Style Transfer + ShengXu + FumiyoFukumoto + YoshimiSuzuki + 330–343 + Sentiment style transfer (SST), a variant of text style transfer (TST), has recently attracted extensive interest. Some disentangling-based approaches have improved performance, while most still struggle to properly transfer the input as the sentiment style is intertwined with the content of the text. To alleviate the issue, we propose a plug-and-play method that leverages an iterative self-refinement algorithm with a large language model (LLM). Our approach separates the straightforward Seq2Seq generation into two phases: (1) Reduction phase which generates a style-free sequence for a given text, and (2) Synthesis phase which generates the target text by leveraging the sequence output from the first phase. The experimental results on two datasets demonstrate that our transfer strategy is effective for challenging SST cases where the baseline methods perform poorly. Our code is available online. + 2024.inlg-main.28 + 2024.inlg-main.28.Supplementary_Attachment.zip + xu-etal-2024-reduction-synthesis + + + Resilience through Scene Context in Visual Referring Expression Generation + SimeonJunker + SinaZarrieß + 344–357 + Scene context is well known to facilitate humans’ perception of visible objects. In this paper, we investigate the role of context in Referring Expression Generation (REG) for objects in images, where existing research has often focused on distractor contexts that exert pressure on the generator. We take a new perspective on scene context in REG and hypothesize that contextual information can be conceived of as a resource that makes REG models more resilient and facilitates the generation of object descriptions, and object types in particular. We train and test Transformer-based REG models with target representations that have been artificially obscured with noise to varying degrees. We evaluate how properties of the models’ visual context affect their processing and performance. Our results show that even simple scene contexts make models surprisingly resilient to perturbations, to the extent that they can identify referent types even when visual information about the target is completely missing. + 2024.inlg-main.29 + junker-zarriess-2024-resilience-scene + + + The Unreasonable Ineffectiveness of Nucleus Sampling on Mitigating Text Memorization + LukaBorec + PhilippSadler + DavidSchlangen + 358–370 + This work analyses the text memorization behavior of large language models (LLMs) when subjected to nucleus sampling. Stochastic decoding methods like nucleus sampling are typically applied to overcome issues such as monotonous and repetitive text generation, which are often observed with maximization-based decoding techniques. We hypothesize that nucleus sampling might also reduce the occurrence of memorization patterns, because it could lead to the selection of tokens outside the memorized sequence. To test this hypothesis we create a diagnostic dataset with a known distribution of duplicates that gives us some control over the likelihood of memorisation of certain parts of the training data. Our analysis of two GPT-Neo models fine-tuned on this dataset interestingly shows that (i) an increase of the nucleus size reduces memorization only modestly, and (ii) even when models do not engage in “hard” memorization – a verbatim reproduction of training samples – they may still display “soft” memorization whereby they generate outputs that echo the training data but without a complete one-by-one resemblance. + 2024.inlg-main.30 + borec-etal-2024-unreasonable-ineffectiveness + + + <fixed-case>CADGE</fixed-case>: Context-Aware Dialogue Generation Enhanced with Graph-Structured Knowledge Aggregation + ChenTang + HongboZhang + TylerLoakman + BohaoYang + StefanGoetze + ChenghuaLin + 371–383 + Commonsense knowledge is crucial to many natural language processing tasks. Existing works usually incorporate graph knowledge with conventional graph neural networks (GNNs), resulting in a sequential pipeline that compartmentalizes the encoding processes for textual and graph-based knowledge. This compartmentalization does, however, not fully exploit the contextual interplay between these two types of input knowledge. In this paper, a novel context-aware graph-attention model (Context-aware GAT) is proposed, designed to effectively assimilate global features from relevant knowledge graphs through a context-enhanced knowledge aggregation mechanism. Specifically, the proposed framework employs an innovative approach to representation learning that harmonizes heterogeneous features by amalgamating flattened graph knowledge with text data. The hierarchical application of graph knowledge aggregation within connected subgraphs, complemented by contextual information, to bolster the generation of commonsense-driven dialogues is analyzed. Empirical results demonstrate that our framework outperforms conventional GNN-based language models in terms of performance. Both, automated and human evaluations affirm the significant performance enhancements achieved by our proposed model over the concept flow baseline. + 2024.inlg-main.31 + tang-etal-2024-cadge-context + + + Context-aware Visual Storytelling with Visual Prefix Tuning and Contrastive Learning + YingjinSong + DenisPaperno + AlbertGatt + 384–401 + Visual storytelling systems generate multi-sentence stories from image sequences. In this task, capturing contextual information and bridging visual variation bring additional challenges. We propose a simple yet effective framework that leverages the generalization capabilities of pretrained foundation models, only training a lightweight vision-language mapping network to connect modalities, while incorporating context to enhance coherence. We introduce a multimodal contrastive objective that also improves visual relevance and story informativeness. Extensive experimental results, across both automatic metrics and human evaluations, demonstrate that the stories generated by our framework are diverse, coherent, informative, and interesting. + 2024.inlg-main.32 + song-etal-2024-context-aware + + + Enhancing Editorial Tasks: A Case Study on Rewriting Customer Help Page Contents Using Large Language Models + AleksandraGabryszak + DanielRöder + ArneBinder + LucaSion + LeonhardHennig + 402–411 + In this paper, we investigate the use of large language models (LLMs) to enhance the editorial process of rewriting customer help pages. We introduce a German-language dataset comprising Frequently Asked Question-Answer pairs, presenting both raw drafts and their revisions by professional editors. On this dataset, we evaluate the performance of four large language models (LLM) through diverse prompts tailored for the rewriting task. We conduct automatic evaluations of content and text quality using ROUGE, BERTScore, and ChatGPT. Furthermore, we let professional editors assess the helpfulness of automatically generated FAQ revisions for editorial enhancement. Our findings indicate that LLMs can produce FAQ reformulations beneficial to the editorial process. We observe minimal performance discrepancies among LLMs for this task, and our survey on helpfulness underscores the subjective nature of editors’ perspectives on editorial refinement. + 2024.inlg-main.33 + gabryszak-etal-2024-enhancing-editorial + + + Customizing Large Language Model Generation Style using Parameter-Efficient Finetuning + XinyueLiu + HarshitaDiddee + DaphneIppolito + 412–426 + One-size-fits-all large language models (LLMs) are increasingly being used to help people with their writing. However, the style these models are trained to write in may not suit all users or use cases. LLMs would be more useful as writing assistants if their idiolect could be customized to match each user. In this paper, we explore whether parameter-efficient finetuning (PEFT) with Low-Rank Adaptation can effectively guide the style of LLM generations. We use this method to customize LLaMA-2 to ten different authors and show that the generated text has lexical, syntactic, and surface alignment with the target author but struggles with content memorization. Our findings highlight the potential of PEFT to support efficient, user-level customization of LLMs. + 2024.inlg-main.34 + liu-etal-2024-customizing-large + + + Towards Fine-Grained Citation Evaluation in Generated Text: A Comparative Analysis of Faithfulness Metrics + WeijiaZhang + MohammadAliannejadi + YifeiYuan + JiahuanPei + Jia-hongHuang + EvangelosKanoulas + 427–439 + Large language models (LLMs) often produce unsupported or unverifiable content, known as “hallucinations.” To mitigate this, retrieval-augmented LLMs incorporate citations, grounding the content in verifiable sources. Despite such developments, manually assessing how well a citation supports the associated statement remains a major challenge. Previous studies use faithfulness metrics to estimate citation support automatically but are limited to binary classification, overlooking fine-grained citation support in practical scenarios. To investigate the effectiveness of faithfulness metrics in fine-grained scenarios, we propose a comparative evaluation framework that assesses the metric effectiveness in distinguishing citations between three-category support levels: full, partial, and no support. Our framework employs correlation analysis, classification evaluation, and retrieval evaluation to measure the alignment between metric scores and human judgments comprehensively. Our results show no single metric consistently excels across all evaluations, revealing the complexity of assessing fine-grained support. Based on the findings, we provide practical recommendations for developing more effective metrics. + 2024.inlg-main.35 + zhang-etal-2024-towards-fine-grained + + + Audio-visual training for improved grounding in video-text <fixed-case>LLM</fixed-case>s + Shivprasad RajendraSagare + HemachandranS + KinshukSarabhai + PrashantUllegaddi + RajeshkumarSa + 440–445 + Recent advances in multimodal LLMs, have led to several video-text models being proposed for critical video-related tasks. However, most of the previous works support visual input only, essentially muting the audio signal in the video. Few models that support both audio and visual input, are not explicitly trained on audio data. Hence, the effect of audio towards video understanding is largely unexplored. To this end, we propose a model architecture that handles audio-visual inputs explicitly. We train our model with both audio and visual data from a video instruction-tuning dataset. Comparison with vision-only baselines, and other audio-visual models showcase that training on audio data indeed leads to better grounding of responses. For better evaluation of audio-visual models, we also release a human-annotated benchmark dataset, with audio-aware question-answer pairs. + 2024.inlg-main.36 + sagare-etal-2024-audio-visual + + + ai<fixed-case>X</fixed-case>plain <fixed-case>SDK</fixed-case>: A High-Level and Standardized Toolkit for <fixed-case>AI</fixed-case> Assets + ShreyasSharma + LucasPavanelli + ThiagoCastro Ferreira + MohamedAl-Badrashiny + HassanSawaf + 446–452 + The aiXplain SDK is an open-source Python toolkit which aims to simplify the wide and complex ecosystem of AI resources. The toolkit enables access to a wide selection of AI assets, including datasets, models, and metrics, from both academic and commercial sources, which can be selected, executed and evaluated in one place through different services in a standardized format with consistent documentation provided. The study showcases the potential of the proposed toolkit with different code examples and by using it on a user journey where state-of-the-art Large Language Models are fine-tuned on instruction prompt datasets, outperforming their base versions. + 2024.inlg-main.37 + sharma-etal-2024-aixplain-sdk + + + Referring Expression Generation in Visually Grounded Dialogue with Discourse-aware Comprehension Guiding + BramWillemsen + GabrielSkantze + 453–469 + We propose an approach to referring expression generation (REG) in visually grounded dialogue that is meant to produce referring expressions (REs) that are both discriminative and discourse-appropriate. Our method constitutes a two-stage process. First, we model REG as a text- and image-conditioned next-token prediction task. REs are autoregressively generated based on their preceding linguistic context and a visual representation of the referent. Second, we propose the use of discourse-aware comprehension guiding as part of a generate-and-rerank strategy through which candidate REs generated with our REG model are reranked based on their discourse-dependent discriminatory power. Results from our human evaluation indicate that our proposed two-stage approach is effective in producing discriminative REs, with higher performance in terms of text-image retrieval accuracy for reranked REs compared to those generated using greedy decoding. + 2024.inlg-main.38 + willemsen-skantze-2024-referring-expression + + + The <fixed-case>G</fixed-case>ricean Maxims in <fixed-case>NLP</fixed-case> - A Survey + LeaKrause + Piek T.J.M.Vossen + 470–485 + In this paper, we provide an in-depth review of how the Gricean maxims have been used to develop and evaluate Natural Language Processing (NLP) systems. Originating from the domain of pragmatics, the Gricean maxims are foundational principles aimed at optimising communicative effectiveness, encompassing the maxims of Quantity, Quality, Relation, and Manner. We explore how these principles are operationalised within NLP through the development of data sets, benchmarks, qualitative evaluation and the formulation of tasks such as Data-to-text, Referring Expressions, Conversational Agents, and Reasoning with a specific focus on Natural Language Generation (NLG). We further present current works on the integration of these maxims in the design and assessment of Large Language Models (LLMs), highlighting their potential influence on enhancing model performance and interaction capabilities. Additionally, this paper identifies and discusses relevant challenges and opportunities, with a special emphasis on the cultural adaptation and contextual applicability of the Gricean maxims. While they have been widely used in different NLP applications, we present the first comprehensive survey of the Gricean maxims’ impact. + 2024.inlg-main.39 + krause-vossen-2024-gricean-maxims + + + Leveraging Plug-and-Play Models for Rhetorical Structure Control in Text Generation + YukaYokogawa + TatsuyaIshigaki + HiroyaTakamura + YusukeMiyao + IchiroKobayashi + 486–493 + We propose a method that extends a BART-based language generator using a plug-and-play model to control the rhetorical structure of generated text. Our approach considers rhetorical relations between clauses and generates sentences that reflect this structure using plug-and-play language models. We evaluated our method using the Newsela corpus, which consists of texts at various levels of English proficiency. Our experiments demonstrated that our method outperforms the vanilla BART in terms of the correctness of output discourse and rhetorical structures. In existing methods, the rhetorical structure tends to deteriorate when compared to the baseline, the vanilla BART, as measured by n-gram overlap metrics such as BLEU. However, our proposed method does not exhibit this significant deterioration, demonstrating its advantage. + 2024.inlg-main.40 + 2024.inlg-main.40.Supplementary_Attachment.pdf + yokogawa-etal-2024-leveraging-plug + + + Multilingual Text Style Transfer: Datasets & Models for <fixed-case>I</fixed-case>ndian Languages + SourabrataMukherjee + Atul Kr.Ojha + AkankshaBansal + DeepakAlok + John P.McCrae + OndrejDusek + 494–522 + Text style transfer (TST) involves altering the linguistic style of a text while preserving its style-independent content. This paper focuses on sentiment transfer, a popular TST subtask, across a spectrum of Indian languages: Hindi, Magahi, Malayalam, Marathi, Punjabi, Odia, Telugu, and Urdu, expanding upon previous work on English-Bangla sentiment transfer. We introduce dedicated datasets of 1,000 positive and 1,000 negative style-parallel sentences for each of these eight languages. We then evaluate the performance of various benchmark models categorized into parallel, non-parallel, cross-lingual, and shared learning approaches, including the Llama2 and GPT-3.5 large language models (LLMs). Our experiments highlight the significance of parallel data in TST and demonstrate the effectiveness of the Masked Style Filling (MSF) approach in non-parallel techniques. Moreover, cross-lingual and joint multilingual learning methods show promise, offering insights into selecting optimal models tailored to the specific language and task requirements. To the best of our knowledge, this work represents the first comprehensive exploration of the TST task as sentiment transfer across a diverse set of languages. + 2024.inlg-main.41 + mukherjee-etal-2024-multilingual-text + + + Are Large Language Models Actually Good at Text Style Transfer? + SourabrataMukherjee + Atul Kr.Ojha + OndrejDusek + 523–539 + We analyze the performance of large language models (LLMs) on Text Style Transfer (TST), specifically focusing on sentiment transfer and text detoxification across three languages: English, Hindi, and Bengali. Text Style Transfer involves modifying the linguistic style of a text while preserving its core content. We evaluate the capabilities of pre-trained LLMs using zero-shot and few-shot prompting as well as parameter-efficient finetuning on publicly available datasets. Our evaluation using automatic metrics, GPT-4 and human evaluations reveals that while some prompted LLMs perform well in English, their performance in on other languages (Hindi, Bengali) remains average. However, finetuning significantly improves results compared to zero-shot and few-shot prompting, making them comparable to previous state-of-the-art. This underscores the necessity of dedicated datasets and specialized models for effective TST. + 2024.inlg-main.42 + mukherjee-etal-2024-large-language + + + Towards Effective Long Conversation Generation with Dynamic Topic Tracking and Recommendation + TrevorAshby + AdithyaKulkarni + JingyuanQi + MinqianLiu + EunahCho + VaibhavKumar + LifuHuang + 540–556 + During conversations, the human flow of thoughts may result in topic shifts and evolution. In open-domain dialogue systems, it is crucial to track the topics discussed and recommend relevant topics to be included in responses to have effective conversations. Furthermore, topic evolution is needed to prevent stagnation as conversation length increases. Existing open-domain dialogue systems do not pay sufficient attention to topic evolution and shifting, resulting in performance degradation due to ineffective responses as conversation length increases. To address the shortcomings of existing approaches, we propose EvolvConv. EvolvConv conducts real-time conversation topic and user preference tracking and utilizes the tracking information to evolve and shift topics depending on conversation status. We conduct extensive experiments to validate the topic evolving and shifting capabilities of EvolvConv as conversation length increases. Un-referenced evaluation metric UniEval compare EvolvConv with the baselines. Experimental results show that EvolvConv maintains a smooth conversation flow without abruptly shifting topics; the probability of topic shifting ranges between 5%-8% throughout the conversation. EvolvConv recommends 4.77% more novel topics than the baselines, and the topic evolution follows balanced topic groupings. Furthermore, we conduct user surveys to test the practical viability of EvolvConv. User survey results reveal that responses generated by EvolvConv are preferred 47.8% of the time compared to the baselines and comes second to real human responses. + 2024.inlg-main.43 + ashby-etal-2024-towards-effective + + + Automatic Metrics in Natural Language Generation: A survey of Current Evaluation Practices + PatriciaSchmidtova + SaadMahamood + SimoneBalloccu + OndrejDusek + AlbertGatt + DimitraGkatzia + David M.Howcroft + OndrejPlatek + AdarsaSivaprasad + 557–583 + Automatic metrics are extensively used to evaluate Natural Language Processing systems. However, there has been increasing focus on how the are used and reported by practitioners within the field. In this paper, we have conducted a survey on the use of automatic metrics, focusing particularly on natural language generation tasks. We inspect which metrics are used as well as why they are chosen and how their use is reported. Our findings from this survey reveal significant shortcomings, including inappropriate metric usage, lack of implementation details and missing correlations with human judgements. We conclude with recommendations that we believe authors should follow to enable more rigour within the field. + 2024.inlg-main.44 + schmidtova-etal-2024-automatic-metrics + + + A Comprehensive Analysis of Memorization in Large Language Models + HirokazuKiyomaru + IssaSugiura + DaisukeKawahara + SadaoKurohashi + 584–596 + This paper presents a comprehensive study that investigates memorization in large language models (LLMs) from multiple perspectives. Experiments are conducted with the Pythia and LLM-jp model suites, both of which offer LLMs with over 10B parameters and full access to their pre-training corpora. Our findings include: (1) memorization is more likely to occur with larger model sizes, longer prompt lengths, and frequent texts, which aligns with findings in previous studies; (2) memorization is less likely to occur for texts not trained during the latter stages of training, even if they frequently appear in the training corpus; (3) the standard methodology for judging memorization can yield false positives, and texts that are infrequent yet flagged as memorized typically result from causes other than true memorization. + 2024.inlg-main.45 + kiyomaru-etal-2024-comprehensive-analysis + + + Generating Attractive Ad Text by Facilitating the Reuse of Landing Page Expressions + HidetakaKamigaito + SoichiroMurakami + PeinanZhang + HiroyaTakamura + ManabuOkumura + 597–608 + Ad text generation is vital for automatic advertising in various fields through search engine advertising (SEA) to avoid the cost problem caused by laborious human efforts for creating ad texts. Even though ad creators create the landing page (LP) for advertising and we can expect its quality, conventional approaches with reinforcement learning (RL) mostly focus on advertising keywords rather than LP information. This work investigates and shows the effective usage of LP information as a reward in RL-based ad text generation through automatic and human evaluations. Our analysis of the actually generated ad text shows that LP information can be a crucial reward by appropriately scaling its value range to improve ad text generation performance. + 2024.inlg-main.46 + kamigaito-etal-2024-generating-attractive + + + Differences in Semantic Errors Made by Different Types of Data-to-text Systems + RudaliHuidrom + AnyaBelz + MichelaLorandi + 609–621 + In this paper, we investigate how different semantic, or content-related, errors made by different types of data-to-text systems differ in terms of number and type. In total, we examine 15 systems: three rule-based and 12 neural systems including two large language models without training or fine-tuning. All systems were tested on the English WebNLG dataset version 3.0. We use a semantic error taxonomy and the brat annotation tool to obtain word-span error annotations on a sample of system outputs. The annotations enable us to establish how many semantic errors different (types of) systems make and what specific types of errors they make, and thus to get an overall understanding of semantic strengths and weaknesses among various types of NLG systems. Among our main findings, we observe that symbolic (rule and template-based) systems make fewer semantic errors overall, non-LLM neural systems have better fluency and data coverage, but make more semantic errors, while LLM-based systems require improvement particularly in addressing superfluous. + 2024.inlg-main.47 + huidrom-etal-2024-differences-semantic + + + Leveraging Large Language Models for Building Interpretable Rule-Based Data-to-Text Systems + JędrzejWarczyński + MateuszLango + OndrejDusek + 622–630 + We introduce a simple approach that uses a large language model (LLM) to automatically implement a fully interpretable rule-based data-to-text system in pure Python. Experimental evaluation on the WebNLG dataset showed that such a constructed system produces text of better quality (according to the BLEU and BLEURT metrics) than the same LLM prompted to directly produce outputs, and produces fewer hallucinations than a BART language model fine-tuned on the same data. Furthermore, at runtime, the approach generates text in a fraction of the processing time required by neural approaches, using only a single CPU. + 2024.inlg-main.48 + 2024.inlg-main.48.Supplementary_Attachment.pdf + warczynski-etal-2024-leveraging-large + + + Explainability Meets Text Summarization: A Survey + MahdiDhaini + EgeErdogan + SmarthBakshi + GjergjiKasneci + 631–645 + Summarizing long pieces of text is a principal task in natural language processing with Machine Learning-based text generation models such as Large Language Models (LLM) being particularly suited to it. Yet these models are often used as black-boxes, making them hard to interpret and debug. This has led to calls by practitioners and regulatory bodies to improve the explainability of such models as they find ever more practical use. In this survey, we present a dual-perspective review of the intersection between explainability and summarization by reviewing the current state of explainable text summarization and also highlighting how summarization techniques are effectively employed to improve explanations. + 2024.inlg-main.49 + dhaini-etal-2024-explainability-meets + + + Generating Faithful and Salient Text from Multimodal Data + TahsinaHashem + WeiqingWang + Derry TantiWijaya + Mohammed EunusAli + Yuan-FangLi + 646–662 + While large multimodal models (LMMs) have obtained strong performance on many multimodal tasks, they may still hallucinate while generating text. Their performance on detecting salient features from visual data is also unclear. In this paper, we develop a framework to generate faithful and salient text from mixed-modal data, which includes images and structured data ( represented in knowledge graphs or tables). Specifically, we train a vision critic model to identify hallucinated and non-salient features from the image modality. The critic model also generates a list of salient image features. This information is used in the post editing step to improve the generation quality. Experiments on two datasets show that our framework improves LMMs’ generation quality on both faithfulness and saliency, outperforming recent techniques aimed at reducing hallucination. The dataset and code are available at https://github.com/TahsinaHashem/FaithD2T. + 2024.inlg-main.50 + 2024.inlg-main.50.Supplementary_Attachment.pdf + hashem-etal-2024-generating-faithful + + + Investigating Paraphrase Generation as a Data Augmentation Strategy for Low-Resource <fixed-case>AMR</fixed-case>-to-Text Generation + Marco AntonioSobrevilla Cabezudo + Marcio LimaInacio + Thiago Alexandre SalgueiroPardo + 663–675 + Abstract Meaning Representation (AMR) is a meaning representation (MR) designed to abstract away from syntax, allowing syntactically different sentences to share the same AMR graph. Unlike other MRs, existing AMR corpora typically link one AMR graph to a single reference. This paper investigates the value of paraphrase generation in low-resource AMR-to-Text generation by testing various paraphrase generation strategies and evaluating their impact. The findings show that paraphrase generation significantly outperforms the baseline and traditional data augmentation methods, even with fewer training instances. Human evaluations indicate that this strategy often produces syntactic-based paraphrases and can exceed the performance of previous approaches. Additionally, the paper releases a paraphrase-extended version of the AMR corpus. + 2024.inlg-main.51 + sobrevilla-cabezudo-etal-2024-investigating-paraphrase + + + Zooming in on Zero-Shot Intent-Guided and Grounded Document Generation using <fixed-case>LLM</fixed-case>s + PritikaRamu + PranshuGaur + RishitaEmandi + HimanshuMaheshwari + DanishJaved + AparnaGarimella + 676–694 + Repurposing existing content on-the-fly to suit author’s goals for creating initial drafts is crucial for document creation. We introduce the task of intent-guided and grounded document generation: given a user-specified intent (e.g., section title) and a few reference documents, the goal is to generate section-level multimodal documents spanning text and images, grounded on the given references, in a zero-shot setting. We present a data curation strategy to obtain general-domain samples from Wikipedia, and collect 1,000 Wikipedia sections consisting of textual and image content along with appropriate intent specifications and references. We propose a simple yet effective planning-based prompting strategy, Multimodal Plan-And-Write (MM-PAW), to prompt LLMs to generate an intermediate plan with text and image descriptions, to guide the subsequent generation. We compare the performances of MM-PAW and a text-only variant of it with those of zero-shot Chain-of-Thought (CoT) using recent close and open-domain LLMs. Both of them lead to significantly better performances in terms of content relevance, structure, and groundedness to the references, more so in the smaller models (upto 12.5 points increase in Rouge 1-F1) than in the larger ones (upto 4 points increase in R1-F1). They are particularly effective in improving relatively smaller models’ performances, to be on par or higher than those of their larger counterparts for this task. + 2024.inlg-main.52 + 2024.inlg-main.52.Supplementary_Attachment.pdf + ramu-etal-2024-zooming-zero + + + Zero-shot cross-lingual transfer in instruction tuning of large language models + NadezhdaChirkova + VassilinaNikoulina + 695–708 + Instruction tuning (IT) is widely used to teach pretrained large language models (LLMs) to follow arbitrary instructions, but is under-studied in multilingual settings. In this work, we conduct a systematic study of zero-shot cross-lingual transfer in IT, when an LLM is instruction-tuned on English-only data and then tested on user prompts in other languages. We advocate for the importance of evaluating various aspects of model responses in multilingual instruction following and investigate the influence of different model configuration choices. We find that cross-lingual transfer does happen successfully in IT even if all stages of model training are English-centric, but only if multiliguality is taken into account in hyperparameter tuning and with large enough IT data. English-trained LLMs are capable of generating correct-language, comprehensive and helpful responses in other languages, but suffer from low factuality and may occasionally have fluency errors. + 2024.inlg-main.53 + chirkova-nikoulina-2024-zero-shot + +
+ + + Proceedings of the 17th International Natural Language Generation Conference: System Demonstrations + SaadMahamood + Nguyen LeMinh + DaphneIppolito + Association for Computational Linguistics +
Tokyo, Japan
+ September + 2024 + 2024.inlg-demos + inlg + + + 2024.inlg-demos.0 + inlg-2024-demos + + + Be My Mate: Simulating Virtual Students for collaboration using Large Language Models + SergiSolera-Monforte + PabloArnau-González + MiguelArevalillo-Herráez + 1–3 + Advancements in machine learning, particularly Large Language Models (LLMs), offer new opportunities for enhancing education through personalized assistance. We introduce “Be My Mate,” an agent that leverages LLMs to simulate virtual peer students in online collaborative education. The system includes a subscription module for real-time updates and a conversational module for generating supportive interactions. Key challenges include creating temporally realistic interactions and credible error generation. The initial demonstration shows promise in enhancing student engagement and learning outcomes. + 2024.inlg-demos.1 + solera-monforte-etal-2024-mate-simulating + + + <fixed-case>MTS</fixed-case>witch: A Web-based System for Translation between Molecules and Texts + NijiaHan + ZimuWang + YuqiWang + HaiyangZhang + DaiyunHuang + WeiWang + 4–6 + We introduce MTSwitch, a web-based system for the bidirectional translation between molecules and texts, leveraging various large language models (LLMs). It supports two crucial tasks, including molecule captioning (explaining the properties of a molecule) and molecule generation (designing a molecule based on specific properties). To the best of our knowledge, MTSwitch is currently the first accessible system that allows users to translate between molecular representations and descriptive text contents. The system and a screencast can be found in https://github.com/hanninaa/MTSwitch. + 2024.inlg-demos.2 + han-etal-2024-mtswitch-web + + + <fixed-case>V</fixed-case>ideo<fixed-case>RAG</fixed-case>: Scaling the context size and relevance for video question-answering + Shivprasad RajendraSagare + PrashantUllegaddi + NachikethK S + NavanithR + KinshukSarabhai + Rajesh KumarS A + 7–8 + Recent advancements have led to the adaptation of several multimodal large language models (LLMs) for critical video-related use cases, particularly in Video Question-Answering (QA). However, most of the previous models sample only a limited number of frames from video due to the context size limit of backbone LLM. Another approach of applying temporal pooling to compress multiple frames, is also shown to saturate and does not scale well. These limitations cause videoQA on long videos to perform very poorly. To address this, we present VideoRAG, a system to utilize recently popularized Retrieval Augmented Generation (RAG) pipeline to select the top-k frames from video, relevant to the user query. We have observed a qualitative improvement in our experiments, indicating a promising direction to pursue. Additionally, our findings indicate that videoRAG demonstrates superior performance when addressing needle-in-the-haystack questions in long videos. Our extensible system allows for trying multiple strategies for indexing, ranking, and adding QA models. + 2024.inlg-demos.3 + sagare-etal-2024-videorag-scaling + + + <fixed-case>QCET</fixed-case>: An Interactive Taxonomy of Quality Criteria for Comparable and Repeatable Evaluation of <fixed-case>NLP</fixed-case> Systems + AnyaBelz + SimonMille + CraigThomson + RudaliHuidrom + 9–12 + Four years on from two papers (Belz et al., 2020; Howcroft et al., 2020) that first called out the lack of standardisation and comparability in the quality criteria assessed in NLP system evaluations, researchers still use widely differing quality criteria names and definitions, meaning that it continues to be unclear when the same aspect of quality is being assessed in two evaluations. While normalised quality criteria were proposed at the time, the list was unwieldy and using it came with a steep learning curve. In this demo paper, our aim is to address these issues with an interactive taxonomy tool that enables quick perusal and selection of the quality criteria, and provides decision support and examples of use at each node. + 2024.inlg-demos.4 + belz-etal-2024-qcet-interactive + + + factgenie: A Framework for Span-based Evaluation of Generated Texts + ZdeněkKasner + OndrejPlatek + PatriciaSchmidtova + SimoneBalloccu + OndrejDusek + 13–15 + We present ‘factgenie‘: a framework for annotating and visualizing word spans in textual model outputs. Annotations can capture various span-based phenomena such as semantic inaccuracies or irrelevant text. With ‘factgenie‘, the annotations can be collected both from human crowdworkers and large language models. Our framework consists of a web interface for data visualization and gathering text annotations, powered by an easily extensible codebase. + 2024.inlg-demos.5 + kasner-etal-2024-factgenie-framework + + + Filling Gaps in <fixed-case>W</fixed-case>ikipedia: Leveraging Data-to-Text Generation to Improve Encyclopedic Coverage of Underrepresented Groups + SimonMille + MassimilianoPronesti + CraigThomson + MichelaLorandi + SophieFitzpatrick + RudaliHuidrom + MohammedSabry + AmyO’Riordan + AnyaBelz + 16–19 + Wikipedia is known to have systematic gaps in its coverage that correspond to under-resourced languages as well as underrepresented groups. This paper presents a new tool to support efforts to fill in these gaps by automatically generating draft articles and facilitating post-editing and uploading to Wikipedia. A rule-based generator and an input-constrained LLM are used to generate two alternative articles, enabling the often more fluent, but error-prone, LLM-generated article to be content-checked against the more reliable, but less fluent, rule-generated article. + 2024.inlg-demos.6 + mille-etal-2024-filling-gaps + +
+ + + Proceedings of the 17th International Natural Language Generation Conference: Tutorial Abstract + AnyaBelz + JoãoSedo + CraigThomson + SimonMille + RudaliHuidrom + Association for Computational Linguistics +
Tokyo, Japan
+ September + 2024 + 2024.inlg-tutorials + inlg + + + 2024.inlg-tutorials.0 + inlg-2024-tutorials + + + The <fixed-case>INLG</fixed-case> 2024 Tutorial on Human Evaluation of <fixed-case>NLP</fixed-case> System Quality: Background, Overall Aims, and Summaries of Taught Units + AnyaBelz + JoãoSedoc + CraigThomson + SimonMille + RudaliHuidrom + 1–12 + Following numerous calls in the literature for improved practices and standardisation in human evaluation in Natural Language Processing over the past ten years, we held a tutorial on the topic at the 2024 INLG Conference. The tutorial addressed the structure, development, design, implementation, execution and analysis of human evaluations of NLP system quality. Hands-on practical sessions were run, designed to facilitate assimilation of the material presented. Slides, lecture recordings, code and data have been made available on GitHub (https://github.com/Human-Evaluation-Tutorial/INLG-2024-Tutorial). In this paper, we provide summaries of the content of the eight units of the tutorial, alongside its research context and aims. + 2024.inlg-tutorials.1 + belz-etal-2024-inlg + +
+ + + 2024.aiwolfdial-1 + 2024.practicald2t-1 + + +
diff --git a/data/xml/2024.practicald2t.xml b/data/xml/2024.practicald2t.xml new file mode 100644 index 0000000000..20fdd60b65 --- /dev/null +++ b/data/xml/2024.practicald2t.xml @@ -0,0 +1,52 @@ + + + + + Proceedings of the 2nd Workshop on Practical LLM-assisted Data-to-Text Generation + SimoneBalloccu + ZdeněkKasner + OndřejPlátek + PatríciaSchmidtová + KristýnaOnderková + MateuszLango + OndřejDušek + LucieFlek + EhudReiter + DimitraGkatzia + SimonMille + Association for Computational Linguistics +
Tokyo, Japan
+ September + 2024 + 2024.practicald2t-1 + practicald2t + ws + + + 2024.practicald2t-1.0 + practicald2t-2024-1 + + + Beyond the Hype: Identifying and Analyzing Math Word Problem-Solving Challenges for Large Language Models + Romina SoledadAlbornoz-De Luise + DavidArnau + PabloArnau-González + MiguelArevalillo-Herráez + 1–6 + Despite not being explicitly trained for this purpose, models like Mistral and LLaMA have demonstrated impressive results across numerous tasks, including generating solutions to Mathematical Word Problems (MWPs). A MWP involves translating a textual description into a mathematical model or equation that solving it. However, these models face challenges in accurately interpreting and utilizing the numerical information present in the MWP statements, which can lead to errors in the generated solutions. To better understand the limitations of LLMs, we analyzed the MWP where models failed to accurately solve problems from the SVAMP dataset. By categorizing these MWPs, we identify specific types of problems where the models are most prone to errors, providing insights into the underlying challenges faced by LLMs in problem-solving scenarios and open new modeling opportunities. By understanding the expected errors, researchers can design strategies to adequately model problems more effectively and choose the most suitable LLM for solving them taking into account each model’s strengths and weaknesses. + 2024.practicald2t-1.1 + albornoz-de-luise-etal-2024-beyond + + + Enhancing Situation Awareness through Model-Based Explanation Generation + KonstantinosGavriilidis + IoannisKonstas + HelenHastie + WeiPang + 7–16 + Robots are often deployed in remote locations for tasks such as exploration, where users cannot directly perceive the agent and its environment. For Human-In-The-Loop applications, operators must have a comprehensive understanding of the robot’s current state and its environment to take necessary actions and effectively assist the agent. In this work, we compare different explanation styles to determine the most effective way to convey real-time updates to users. Additionally, we formulate these explanation styles as separate fine-tuning tasks and assess the effectiveness of large language models in delivering in-mission updates to maintain situation awareness. The code and dataset for this work are available at:——— + 2024.practicald2t-1.2 + gavriilidis-etal-2024-enhancing + +
+
diff --git a/data/yaml/sigs/siggen.yaml b/data/yaml/sigs/siggen.yaml index 9ab9d7efcb..98c7fe2a85 100644 --- a/data/yaml/sigs/siggen.yaml +++ b/data/yaml/sigs/siggen.yaml @@ -2,6 +2,10 @@ Name: Special Interest Group on Natural Language Generation (SIGGEN) ShortName: SIGGEN URL: https://aclweb.org/aclwiki/SIGGEN Meetings: + - 2024: + - 2024.inlg-main + - 2024.inlg-demos + - 2024.inlg-tutorials # Proceedings of the 17th International Natural Language Generation Conference: Tutorial Abstract - 2023: - 2023.inlg-main - 2023.inlg-demos diff --git a/data/yaml/venues/aiwolfdial.yaml b/data/yaml/venues/aiwolfdial.yaml new file mode 100644 index 0000000000..0717230d25 --- /dev/null +++ b/data/yaml/venues/aiwolfdial.yaml @@ -0,0 +1,2 @@ +acronym: AIWolfDial +name: The 2nd International AIWolfDial Workshop diff --git a/data/yaml/venues/practicald2t.yaml b/data/yaml/venues/practicald2t.yaml new file mode 100644 index 0000000000..90406eec81 --- /dev/null +++ b/data/yaml/venues/practicald2t.yaml @@ -0,0 +1,2 @@ +acronym: PracticalD2T +name: The 2nd Workshop on Practical LLM-assisted Data-to-Text Generation