What it does
AI algorithms can discriminate against people without anyone noticing β reproducing the biases in training data across hiring, credit, marketing, and other decisions. If you don't verify, you may be discriminating against customers or candidates without knowing it. This prompt builds a process to detect and correct bias in your AI systems: how to test, which fairness metrics to use, and how to correct issues without destroying model accuracy. Use it when you use AI in decisions that affect people, when you want to ensure your AI is fair, or when regulators or customers raise questions about bias.
When to use
- AI algorithms can discriminate against people without anyone noticing β reproducing the biases in training data across hiring, credit, marketing, and other decisions
- If you don't verify, you may be discriminating against customers or candidates without knowing it
- This prompt builds a process to detect and correct bias in your AI systems: how to test, which fairness metrics to use, and how to correct issues without destroying model accuracy
- Use it when you use AI in decisions that affect people, when you want to ensure your AI is fair, or when regulators or customers raise questions about bias
What you will get
A structured result ready to use, personalized for your context.