{"data":{"full_name":"back2matching/turboquant","name":"turboquant","description":"First open-source TurboQuant KV cache compression for LLM inference. Drop-in for HuggingFace. pip install turboquant.","stars":5.0,"forks":1.0,"language":"Python","license":"NOASSERTION","archived":0.0,"subcategory":"llm-quantization-methods","last_pushed_at":"2026-03-27T01:57:05+00:00","pypi_package":"turboquant","npm_package":null,"downloads_monthly":0.0,"dependency_count":4.0,"commits_30d":null,"reverse_dep_count":0.0,"maintenance_score":13.0,"adoption_score":4.0,"maturity_score":18.0,"community_score":12.0,"quality_score":47.0,"quality_tier":"emerging","risk_flags":"[]"},"meta":{"timestamp":"2026-04-07T10:06:19.760095+00:00"}}