AI Safety Bounties
Showing 2 bounties in AI safety research
Model Uncertainty Quantification for Safety-Critical Decisions
Open
Develop methods for accurately quantifying model uncertainty in safety-critical contexts, enabling systems to know when to defer to human judgment.
*...
alignmentevalsmonitoringbenchmarks
AdvancedAutomaticInference Only
2
Constitutional AI Critique Generator
Open
Build an open-source implementation of Constitutional AI's critique and revision pipeline.
**Goal:** Given a potentially harmful model output, genera...
alignmentevalsrobustness
IntermediateHuman ReviewInference Only
1