{"data":{"full_name":"RahulSChand/gpu_poor","name":"gpu_poor","description":"Calculate token/s & GPU memory requirement for any LLM.  Supports llama.cpp/ggml/bnb/QLoRA quantization","stars":1396.0,"forks":87.0,"language":"JavaScript","license":null,"archived":0.0,"subcategory":"llm-inference-engines","last_pushed_at":"2024-12-03T20:52:37+00:00","pypi_package":null,"npm_package":null,"downloads_monthly":0.0,"dependency_count":0.0,"commits_30d":null,"reverse_dep_count":0.0,"maintenance_score":0.0,"adoption_score":10.0,"maturity_score":8.0,"community_score":16.0,"quality_score":34.0,"quality_tier":"emerging","risk_flags":"['no_license', 'stale_6m', 'no_package', 'no_dependents']"},"meta":{"timestamp":"2026-04-08T10:31:16.682939+00:00"}}