Workshop on Large Language Models' Interpretability and Trustworthiness (LLMIT)

Saha, T., Ganguly, D. , Saha, S. and Mitra, P. (2023) Workshop on Large Language Models' Interpretability and Trustworthiness (LLMIT). In: 32nd ACM International Conference on Information and Knowledge Management (CIKM 2023), Birmingham, UK, 21-25 Oct 2023, pp. 5290-5293. ISBN 9798400701245 (doi: 10.1145/3583780.3615311)

[img] Text
304935.pdf - Accepted Version

522kB

Abstract

Large language models (LLMs), when scaled from millions to billions of parameters, have been demonstrated to exhibit the so-called 'emergence' effect, in that they are not only able to produce semantically correct and coherent text, but are also able to adapt themselves surprisingly well with small changes in contexts supplied as inputs (commonly called prompts). Despite producing semantically coherent and potentially relevant text for a given context, LLMs are vulnerable to yield incorrect information. This misinformation generation, or the so-called hallucination problem of an LLM, gets worse when an adversary manipulates the prompts to their own advantage, e.g., generating false propaganda to disrupt communal harmony, generating false information to trap consumers with target consumables etc. Not only does the consumption of an LLM-generated hallucinated content by humans pose societal threats, such misinformation, when used as prompts, may lead to detrimental effects for in-context learning (also known as few-shot prompt learning). With reference to the above-mentioned problems of LLM usage, we argue that it is necessary to foster research on topics related to not only identifying misinformation from LLM-generated content, but also to mitigate the propagation effects of this generated misinformation on downstream predictive tasks thus leading to more robust and effective leveraging in-context learning.

Item Type:Conference Proceedings
Status:Published
Refereed:Yes
Glasgow Author(s) Enlighten ID:Ganguly, Dr Debasis
Authors: Saha, T., Ganguly, D., Saha, S., and Mitra, P.
College/School:College of Science and Engineering > School of Computing Science
ISBN:9798400701245
Copyright Holders:Copyright © 2023 The Authors
Publisher Policy:Reproduced in accordance with the copyright policy of the publisher
Related URLs:

University Staff: Request a correction | Enlighten Editors: Update this record