Recognition: unknown
LLM Harms: A Taxonomy and Discussion
read the original abstract
This study addresses categories of harm surrounding Large Language Models (LLMs) in the field of artificial intelligence. It addresses five categories of harms addressed before, during, and after development of AI applications: pre-development, direct output, Misuse and Malicious Application, and downstream application. By underscoring the need to define risks of the current landscape to ensure accountability, transparency and navigating bias when adapting LLMs for practical applications. It proposes mitigation strategies and future directions for specific domains and a dynamic auditing system guiding responsible development and integration of LLMs in a standardized proposal.
This paper has not been read by Pith yet.
Forward citations
Cited by 1 Pith paper
-
From Notepad AI to Social Media: How Can Text Style Transformation Mitigate Social Harm?
A framework transforms aggressive social media text into neutral styles while preserving semantics, measured by a new Emotion Drift Index to reduce online harm.
discussion (0)
Sign in with ORCID, Apple, or X to comment. Anyone can read and Pith papers without signing in.