To Err is AI : A Case Study Informing LLM Flaw Reporting Practices
In August of 2024, 495 hackers generated evaluations in an open-ended bug bounty targeting the Open Language Model (OLMo) from The Allen Institute for AI. A vendor panel staffed by representatives of OLMo's safety program adjudicated changes to OLMo's documentation and awarded cash bountie...
Saved in:
Main Authors: | , , , , , , , , , , , , |
---|---|
Format: | Journal Article |
Language: | English |
Published: |
15-10-2024
|
Subjects: | |
Online Access: | Get full text |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Summary: | In August of 2024, 495 hackers generated evaluations in an open-ended bug
bounty targeting the Open Language Model (OLMo) from The Allen Institute for
AI. A vendor panel staffed by representatives of OLMo's safety program
adjudicated changes to OLMo's documentation and awarded cash bounties to
participants who successfully demonstrated a need for public disclosure
clarifying the intent, capacities, and hazards of model deployment. This paper
presents a collection of lessons learned, illustrative of flaw reporting best
practices intended to reduce the likelihood of incidents and produce safer
large language models (LLMs). These include best practices for safety reporting
processes, their artifacts, and safety program staffing. |
---|---|
DOI: | 10.48550/arxiv.2410.12104 |