The full system isn't open source yet - still deciding on licensing. But the benchmark repo has:
- Complete results (500/500 on LongMemEval)
- Raw logs showing each question/answer
- Comparison with baselines
Happy to answer questions about the approach. The core insight: intelligent context organization beats raw context volume. No LLM calls for memory extraction - pure embedding-based retrieval using RudraDB (https://rudradb.com).
If you want to verify independently, I can provide API access.
Full raw benchmark logs published: https://github.com/AceIQ360/AceIQ360-Benchmark
Unlike others who only share top-line numbers, I published every question, answer, and judgment. You can verify the 100% claim yourself.
Great claims. Where is the code to check them at?
Hi Thanks
Benchmark results and methodology here: https://github.com/AceIQ360/AceIQ360-Benchmark
The full system isn't open source yet - still deciding on licensing. But the benchmark repo has: - Complete results (500/500 on LongMemEval) - Raw logs showing each question/answer - Comparison with baselines
Happy to answer questions about the approach. The core insight: intelligent context organization beats raw context volume. No LLM calls for memory extraction - pure embedding-based retrieval using RudraDB (https://rudradb.com).
If you want to verify independently, I can provide API access.