Paper: https://cdn.openai.com/papers/openais...) Blog: https://openai.com/index/advancing-re...)
This white paper details OpenAI's approach to external red teaming for AI models and systems. External red teaming, using outside experts, helps uncover novel risks, stress-test safety measures, and provide independent assessments. The paper explores the design of red teaming campaigns, including team composition, access levels, and documentation. Different red teaming methods—manual, automated, and mixed—are discussed, along with their respective advantages and limitations. Finally, the paper explains how insights from human red teaming can be used to create more robust and efficient automated evaluations for ongoing safety assessments.
ai , model , ai safety , openai, genai, generativeai, artificialintelligence , arxiv , research , paper , publication