{"data":{"full_name":"emredeveloper/Mem-LLM","name":"Mem-LLM","description":"Mem-LLM is a Python library for building memory-enabled AI assistants that run entirely on local LLMs, combining persistent multi-user conversation history with configurable knowledge bases, storage backends, and Ollama model support—perfect for privacy-first, production-ready workflows.","stars":7.0,"forks":0.0,"language":"Python","license":"Apache-2.0","archived":0.0,"subcategory":"llm-evaluation-benchmarking","last_pushed_at":"2026-03-05T20:11:40+00:00","pypi_package":"mem-llm","npm_package":null,"downloads_monthly":470.0,"dependency_count":6.0,"commits_30d":null,"reverse_dep_count":0.0,"maintenance_score":10.0,"adoption_score":10.0,"maturity_score":18.0,"community_score":0.0,"quality_score":38.0,"quality_tier":"emerging","risk_flags":"[]"},"meta":{"timestamp":"2026-04-06T20:33:36.473506+00:00"}}