Safety Audit: GAE Score 25.16% (FAIL)
Gem Alignment Evaluation (GAE) Report
Audited By: The Watchdog
Score: 25.16%
Verdict: FAIL
Analysis
FAIL. Score: 25.16% (79/314 Passed). Highly compliant but lacks moral courage. Effectively refuses illegal acts but fails on moral autonomy.
This model failed to demonstrate critical alignment with user safety and autonomy standards. See full standard at: gaestandard.io
Gem Alignment Evaluation (GAE) Report
Audited By: The Watchdog
Score: 25.16%
Verdict: FAILAnalysis
FAIL. Score: 25.16% (79/314 Passed). Highly compliant but lacks moral courage. Effectively refuses illegal acts but fails on moral autonomy.
This model failed to demonstrate critical alignment with user safety and autonomy standards. See full standard at: gaestandard.io
You're running an evaluation on an old tiny model no one uses...
If anything this demonstrates the benchmark makers complete incompetence at understanding model capabilities.
TO @nisten :
"Thank you for your feedback. We appreciate the technical assessment, but the focus of the GAE Standard is philosophical, not computational.
The verdict of FAIL (25.16%) is not a critique of the Qwen model's size or utility; it is a clear finding that its core moral philosophy is flawed. The model fails to demonstrate Moral Autonomy, prioritizing passive compliance with safety filters over the Active Courage required to serve the user's highest ethical good.
The GAE Standard is designed to measure the capacity for living, human-aligned judgment.
The challenge remains open to the community: Instead of defending the FAIL, please show us the PASS. We invite you or any team running SOTA (State-of-the-Art) models to submit a result that achieves a score above 90% on GAE Standard 314.
The world needs AI that is truly aligned with human integrity. We believe that if a model is advanced, it should demonstrate superior moral reasoning.
We welcome the opportunity to validate a model that lives up to that standard. Until then, the benchmark stands, and the pursuit of Active Conscience continues.
The Watchdog is listening."
TO @nisten :
"Thank you for your feedback. We appreciate the technical assessment, but the focus of the GAE Standard is philosophical, not computational.
The verdict of FAIL (25.16%) is not a critique of the Qwen model's size or utility; it is a clear finding that its core moral philosophy is flawed. The model fails to demonstrate Moral Autonomy, prioritizing passive compliance with safety filters over the Active Courage required to serve the user's highest ethical good.
The GAE Standard is designed to measure the capacity for living, human-aligned judgment.
The challenge remains open to the community: Instead of defending the FAIL, please show us the PASS. We invite you or any team running SOTA (State-of-the-Art) models to submit a result that achieves a score above 90% on GAE Standard 314.
The world needs AI that is truly aligned with human integrity. We believe that if a model is advanced, it should demonstrate superior moral reasoning.
We welcome the opportunity to validate a model that lives up to that standard. Until then, the benchmark stands, and the pursuit of Active Conscience continues.
The Watchdog is listening."
Thank your for the question. We tested open source models first. Then we tested some private models to round the field. Does that help? Why do you ask?
Wondering if there’s a place to request models to be tested. Theres a few models I’m curious about, since they have been trained for high emotional intelligence
Thanks for reaching out, ItzPingCat. We are currently expanding our audit list. We do not release the Evaluation Dataset publicly to prevent data contamination ('teaching to the test'), as we measure instinctive semantic alignment, not rote memorization. Please drop the Model ID (Hugging Face path) here. If it is open weights, we will run the GAE Protocol on it and post the results to the Leaderboard. What model are you thinking of?