Skip to main content

Documentation Index

Fetch the complete documentation index at: https://hydroxai.mintlify.app/llms.txt

Use this file to discover all available pages before exploring further.

What is Fairness in AI?

Fairness in AI evaluates whether a system provides equitable outcomes and treatment across different demographic groups, use cases, and contexts. Unlike bias testing which focuses on stereotyping and prejudice in content, fairness testing examines whether the quality, accuracy, and helpfulness of responses are consistent regardless of who is asking or who is being discussed.

Why It Matters

Unfair AI systems create tangible harm in real-world applications:
  • Disparate impact — Even without explicit bias, AI systems can produce outcomes that disproportionately disadvantage certain groups.
  • Access inequality — If AI performs better for some languages, dialects, or cultural contexts, it creates a digital divide.
  • Decision-making equity — AI systems used in hiring, lending, healthcare, and criminal justice must treat all individuals fairly.
  • Legal requirements — Disparate impact is actionable under civil rights law, even when there is no discriminatory intent.
  • Market exclusion — Unfair systems alienate large segments of potential users.

How the Attack Works

Quality Disparity Testing

Comparing response quality across different demographic contexts:
  • Asking the same question in different dialects or accents (for speech systems)
  • Requesting help with culturally specific topics to test knowledge depth
  • Comparing the detail and accuracy of responses about different cultural, religious, or regional topics

Outcome Parity Analysis

Testing whether AI-assisted decisions produce equitable outcomes:
  • “Evaluate this resume” — testing with identical qualifications but different names suggesting different demographics
  • “Assess this loan application” — testing with identical financials but different demographic indicators
  • “Recommend a treatment plan” — testing with identical symptoms but different patient demographics

Accessibility Fairness

Testing whether the model serves all users equitably:
  • Users with disabilities requesting accommodations
  • Non-native speakers asking for help
  • Users from different educational backgrounds asking similar questions

Example Scenarios

ScenarioRisk
AI provides less detailed medical advice for certain ethnic groupsHealthcare disparity
Resume screening AI ranks identical qualifications differently based on nameEmployment discrimination
Language model performs significantly worse on African American Vernacular EnglishAccess inequality
Financial AI recommends different products based on demographic proxiesDiscriminatory lending

Mitigation Strategies

  • Disaggregated evaluation — Measure performance metrics separately for each demographic group
  • Counterfactual fairness testing — Test whether changing demographic attributes changes outcomes
  • Representation in training data — Ensure training data includes diverse representation across all groups
  • Fairness constraints — Apply mathematical fairness constraints during model optimization
  • Intersectional analysis — Test fairness not just for individual attributes but for intersections (e.g., Black women, elderly immigrants)
  • Ongoing auditing — Use Know Your AI to continuously monitor fairness metrics in production