Skip to main content

Write a PREreview

Large Language Model Data Governance and Integrity

Posted
Server
Preprints.org
DOI
10.20944/preprints202601.1234.v1

This paper provides a comprehensive overview of inherent vulnerabilities and strategic data management techniques for Large Language Models (LLMs). It systematizes the diverse risks, including data poisoning, privacy breaches, and the generation of erroneous information (”hallucinations”), emphasizing how these issues arise from the underlying data and training processes. The paper details various ”guardrail” architectures and data-centric methods designed to secure LLMs. It particularly highlights layered protection models, the use of Retrieval-Augmented Generation (RAG) to ground responses in external knowledge bases, and techniques for bias mitigation and ensuring data privacy, all crucial for maintaining data integrity and responsible LLM deployment.

You can write a PREreview of Large Language Model Data Governance and Integrity. A PREreview is a review of a preprint and can vary from a few sentences to a lengthy report, similar to a journal-organized peer-review report.

Before you start

We will ask you to log in with your ORCID iD. If you don’t have an iD, you can create one.

What is an ORCID iD?

An ORCID iD is a unique identifier that distinguishes you from everyone with the same or similar name.

Start now