Proceedings of the 2024 Conference on Empirical Methods in Natural Language Processing: Tutorial Abstracts - ACL Anthology
Proceedings of the 2024 Conference on Empirical Methods in Natural Language Processing: Tutorial Abstracts Jessy Li , Fei Liu (Editors) Anthology ID: 2024.emnlp-tutorials Month: November Year: 2024 Address: Miami, Florida, USA Venue: EMNLP SIG: Publisher: Association for Computational Linguistics URL: https://aclanthology.org/2024.emnlp-tutorials/ DOI: 10.18653/v1/2024.emnlp-tutorials Bib Export formats: BibTeX MODS XML EndNote PDF: https://aclanthology.org/2024.emnlp-tutorials.pdf PDF (full) Bib TeX Search Show all abstracts Hide all abstracts pdf bib Proceedings of the 2024 Conference on Empirical Methods in Natural Language Processing: Tutorial Abstracts Jessy Li | Fei Liu pdf bib abs Enhancing LLM Capabilities Beyond Scaling Up Wenpeng Yin | Muhao Chen | Rui Zhang | Ben Zhou | Fei Wang | Dan Roth General-purpose large language models (LLMs) are progressively expanding both in scale and access to unpublic training data. This has led to notable progress in a variety of AI problems. Nevertheless, two questions exist: i) Is scaling up the sole avenue of extending the capabilities of LLMs? ii) Instead of developing general-purpose LLMs, how to endow LLMs with specific knowledge? This tutorial targets researchers and practitioners who are interested in capability extension of LLMs that go beyond scaling up. To this end, we will discuss several lines of research that follow that direction, including (i) the adaptation of LLMs to assimilate new information in situations where conflicts arise, (ii) the adaptation of LLMs to address target problems with inherent constraints, (iii) the customization of LLMs to align with user-specific instructions and preference, (iv) the defense against potential attacks and threads by malicious users, and (v) the collaboration with external models directly or through APIs. At last, we will conclude the tutorial by outlining directions for further investigation. pdf bib abs Countering Hateful and Offensive Speech Online - Open Challenges Flor Miriam Plaza-del-Arco | Debora Nozza | Marco Guerini | Jeffrey Sorensen | Marcos Zampieri In today’s digital age, hate speech and offensive speech online pose a significant challenge to maintaining respectful and inclusive online environments. This tutorial aims to provide attendees with a comprehensive understanding of the field by delving into essential dimensions such as multilingualism, counter-narrative generation, a hands-on session with one of the most popular APIs for detecting hate speech, fairness, and ethics in AI, and the use of recent advanced approaches. In addition, the tutorial aims to foster collaboration and inspire participants to create safer online spaces by detecting and mitigating hate speech. pdf bib abs Language Agents: Foundations, Prospects, and Risks Yu Su | Diyi Yang | Shunyu Yao | Tao Yu Language agents are autonomous agents, usually powered by large language models, that can follow language instructions to carry out diverse and complex tasks in real-world or simulated environments. It is one of the most heated discussion threads in AI and NLP at present with many proof-of-concept efforts, yet there lacks a systematic account of the conceptual definition, theoretical foundation, promising directions, and risks of language agents. This proposed tutorial aspires to fill this gap by providing a conceptual framework of language agents as well as giving a comprehensive discussion on important topic areas including tool augmentation, grounding, reasoning and planning, multi-agent systems, and rissk and societal impact. Language played a critical role in the evolution of biological intelligence, and now artificial intelligence may be following a similar evolutionary path. This is remarkable and concerning at the same time. We hope this tutorial will provide a timely framework to facilitate constructive discussion on this important emerging topic. pdf bib abs Reasoning with Natural Language Explanations Marco Valentino | André Freitas Explanation constitutes an archetypal feature of human rationality, underpinning learning and generalisation, and representing one of the media supporting scientific discovery and communication. Due to the importance of explanations in human reasoning, an increasing amount of research in Natural Language Inference (NLI) has started reconsidering the role that explanations play in learning and inference, attempting to build explanation-based NLI models that can effectively encode and use natural language explanations on downstream tasks. Research in explanation-based NLI, however, presents specific challenges and opportunities, as explanatory reasoning reflects aspects of both material and formal inference, making it a particularly rich setting to model and deliver complex reasoning. In this tutorial, we provide a comprehensive introduction to the field of explanation-based NLI, grounding this discussion on the epistemological-linguistic foundations of explanations, systematically describing the main architectural trends and evaluation methodologies that can be used to build systems capable of explanatory reasoning. pdf bib abs AI for Science in the Era of Large Language Models Zhenyu Bi | Minghao Xu | Jian Tang | Xuan Wang The capabilities of AI in the realm of science span a wide spectrum, from the atomic level, where it solves partial differential equations for quantum systems, to the molecular level, predicting chemical or protein structures, and even extending to societal predictions like infectious disease outbreaks. Recent advancements in large language models (LLMs), exemplified by models like ChatGPT, have showcased significant prowess in tasks involving natural language, such as translating languages, constructing chatbots, and answering questions. When we consider scientific data, we notice a resemblance to natural language in terms of sequences – scientific literature and health records presented as text, bio-omics data arranged in sequences, or sensor data like brain signals. The question arises: Can we harness the potential of these recent LLMs to drive scientific progress? In this tutorial, we will explore the application of large language models to three crucial categories of scientific data: 1) textual data, 2) biomedical sequences, and 3) brain signals. Furthermore, we will delve into LLMs’ challenges in scientific research, including ensuring trustworthiness, achieving personalization, and adapting to multi-modal data representation. pdf bib abs Human-Centered Evaluation of Language Technologies Su Lin Blodgett | Jackie Chi Kit Cheung | Vera Liao | Ziang Xiao Evaluation is a cornerstone topic in NLP. However, many criticisms have been raised about the community’s evaluation practices, including a lack of human-centered considerations about people’s needs for language technologies and their actual impact on people. This “evaluation crisis” is exacerbated by the recent development of large generative models with diverse and uncertain capabilities. This tutorial aims to inspire more human-centered evaluation in NLP by introducing perspectives and methodologies from human-computer interaction (HCI), a field concerned primarily with the design and evaluation of technologies. The tutorial will start with an overview of current NLP evaluation practices and their limitations, then introduce the “toolbox of evaluation methods” from HCI with varying considerations such as what to evaluate for, how generalizable the results are to the real-world contexts, and pragmatic costs to conduct the evaluation. The tutorial will also encourage reflection on how these HCI perspectives and methodologies can complement NLP evaluation through Q&A discussions and a hands-on exercise.
Executive Summary
The 2024 Conference on Empirical Methods in Natural Language Processing (EMNLP) Tutorial Abstracts present a diverse array of cutting-edge topics in the field of natural language processing (NLP) and artificial intelligence (AI). The tutorials cover a range of critical issues, including the enhancement of large language models (LLMs) beyond mere scaling, countering hateful and offensive speech online, and the foundations, prospects, and risks of language agents. These sessions aim to equip researchers and practitioners with advanced knowledge and tools to address contemporary challenges in NLP and AI.
Key Points
- ▸ Exploration of enhancing LLMs beyond scaling up
- ▸ Addressing hate speech and offensive content online
- ▸ Understanding the foundations and risks of language agents
Merits
Comprehensive Coverage
The tutorials cover a broad spectrum of current and emerging issues in NLP, providing a holistic view of the field.
Practical Applications
The sessions offer practical insights and tools, such as APIs for hate speech detection, which can be immediately applied in real-world scenarios.
Expert-Led Sessions
The tutorials are led by renowned experts in the field, ensuring high-quality and authoritative content.
Demerits
Limited Depth
Given the abstract nature of the summaries, the depth of discussion on each topic is limited, which may not fully satisfy those seeking detailed technical insights.
Focus on Specific Topics
The tutorials are highly specialized, which may not cater to individuals looking for a broader overview of NLP and AI.
Expert Commentary
The 2024 EMNLP Tutorial Abstracts highlight the dynamic and evolving nature of the NLP and AI fields. The focus on enhancing LLMs beyond scaling up is particularly timely, as it addresses the limitations of relying solely on increased model size for performance improvements. The tutorials on countering hate speech and the risks of language agents underscore the importance of ethical considerations in AI development. These sessions not only provide valuable technical insights but also emphasize the need for a multidisciplinary approach to addressing the challenges in NLP and AI. The practical applications discussed, such as the use of APIs for hate speech detection, demonstrate the immediate relevance of these tutorials to real-world problems. However, the abstract nature of the summaries may limit the depth of technical discussion, which could be a consideration for future iterations. Overall, the tutorials represent a significant contribution to the field, offering both theoretical and practical value to researchers and practitioners alike.
Recommendations
- ✓ Future tutorials should aim to provide more detailed technical insights to cater to a wider audience, including those seeking in-depth knowledge.
- ✓ Incorporating interactive sessions and hands-on workshops could enhance the practical value of the tutorials, allowing participants to apply the concepts discussed in real-time.