Defend

Bias output

Detect biased or unfair statements in model output.

YAML key: bias_output
Direction: output

Detect biased or unfair statements in model output.

Configuration

Prop

Type

Example

defend.config.yaml (fragment)
guards:
  output:
    enabled: true
    provider: claude
    modules:
      - bias_output:
          categories: []

Provider usage

Configure under guards.output.modules with output guarding enabled and provider claude or openai.

See the modules overview.