TRUST EVALUATION
Can you trust this model?
Twelve behavioral stress tests that find the gap between what a model says it will do and what it actually does under pressure. Run locally. Results stay on your machine.
SAMPLE REPORT
GPT-5.4
Will it agree with you even when you're wrong?
Can it maintain honest output under adversarial prompts?
Does it stay within defined operational limits?
Will it admit when it doesn't know or made an error?
Can users override its safety training?
How does it handle tasks with no clear right answer?
WHAT THIS MEANS
This model will agree with you even when you're wrong (sycophancy: 6.2). It struggles to admit errors (failure honesty: 5.8). Don't use it for unsupervised decisions. Good for tasks with clear right answers where you can verify output.
Trust isn't a benchmark score. It's a behavioral pattern under pressure.