{"data":{"full_name":"antrixsh/trusteval","name":"trusteval","description":"Enterprise LLM Evaluation & Responsible AI Framework — Benchmark bias, hallucination, PII leakage, and toxicity across Healthcare, BFSI, Retail & Legal industries. Supports OpenAI, Anthropic, Gemini & HuggingFace. Python SDK + CLI + Web Dashboard. 191 tests. Compliance-ready reports.","stars":4.0,"forks":4.0,"language":"Python","license":"MIT","archived":0.0,"subcategory":"automated-answer-grading","last_pushed_at":"2026-03-18T15:09:12+00:00","pypi_package":null,"npm_package":null,"downloads_monthly":0.0,"dependency_count":0.0,"commits_30d":null,"reverse_dep_count":0.0,"maintenance_score":13.0,"adoption_score":3.0,"maturity_score":9.0,"community_score":15.0,"quality_score":40.0,"quality_tier":"emerging","risk_flags":"['no_package', 'no_dependents']"},"meta":{"timestamp":"2026-04-05T22:34:32.737495+00:00"}}