📄 Our paper is out: Optimizing Knowledge Graph-LLM Interface
AIE at SF and Budapest Data+ML Forum - June 2024. Contact: info@topoteretes.com
📄 Read Our Research Paper
We've published our findings on optimizing the interface between Knowledge Graphs and LLMs for complex reasoning. In this paper, we present systematic hyperparameter optimization results using cognee's modular framework across multiple QA benchmarks.
AI Memory Benchmark Results
Understanding how well different AI memory systems retain and utilize context across interactions is crucial for enhancing LLM performance.
We have conducted a comprehensive evaluation of cognee AI memory system against other leading tools, including Dreamify (our proprietary tool), cognee (in vanilla setting), Zep/Graphiti, and Mem0. This analysis provides a detailed comparison of performance metrics, helping developers select the best AI memory solution for their applications.
The evaluation results are based on the following metrics:
Key Performance Metrics
Results for Cognee (Dreamify)
0.89
Human-LLM Correctness
0.75
DeepEval Correctness
0.71
DeepEval F1
0.54
DeepEval EM
Benchmark Comparison
Dreamify: Our Hyperparam framework increases accuracy even more
Cognee with Dreamify shows significant performance improvements across all metrics:
  • Human-LLM Correctness: ~+6% (0.84 → 0.89)
  • DeepEval Correctness: ~+32% (0.57 → 0.75)
  • DeepEval F1: ~+255% (0.20 → 0.71)
  • DeepEval EM: ~ +1250% (0.04 → 0.54)
Comprehensive Metrics Comparison
Dive Deeper
What is Next?
Continuous improvement is key. We are actively enhancing our benchmarks, integrating new metrics, and evaluating additional AI memory solutions. Stay tuned for updates and more detailed analysis.
Have questions or want help optimizing your AI system?