banner
vasilijee.bsky.social
@vasilijee.bsky.social
Founder @cognee.bsky.social | cognee.ai

OSS: github.com/topoteretes/...
Community: discord.gg/m63hxKsp4p
⚡ Learn the BaseRetriever pattern
⚡ See real code snippets
⚡ Take the “Which Retriever Are You?” quiz

Read to get smarter answers? Let me know which retriever you are 🙂

dub.sh/cognee-retri...
Cognee - Semantic Search & Knowledge Graph Retrieval Tactics | Cognee
Drive results with semantic search and knowledge graph retrieval; explore AI retrievers, vector databases, and GraphRAG to turn data into answers—read now!
dub.sh
June 18, 2025 at 3:43 PM
Bottom line: if you’re building agents, assitants, or automated workflows, it’s time to evolve from “data lake” to AI memory “lake”.

- Read the deep dive ➡️ dub.sh/file-based-m...

- GitHub ➡️ github.com/topoteretes/...

- Join us on Discord ➡️ discord.com/invite/tV7pr...
dub.sh
June 12, 2025 at 2:48 PM
We also introduce dreamify - our optimization engine that tunes chunk sizes, retriever configs & prompts in real time for max accuracy and latency ✨
June 12, 2025 at 2:48 PM
Why file-based?

• Cheap, cloud-native (S3, GCS)

• Scales linearly with data growth

• Easy diff + version control

• Plays nicely with existing ETL & BI stacks
June 12, 2025 at 2:48 PM
It’s a living system:

1️⃣ User adds data

2️⃣ Data is cognified

3️⃣ Search & reasoning improve

4️⃣ Feedback flows in

5️⃣ System self-optimizes

…and the loop keeps compounding value. ♻️
June 12, 2025 at 2:48 PM
🔑 Key insight: Data → Memory → Intelligence

Our pipeline “cognifies” every file into graphs, giving agents memory - just like a human mind. So let’s see how 👇🏼
June 12, 2025 at 2:48 PM
First, why care about AI memory?

LLMs are brilliant—until they meet your fragmented data. They forget, hallucinate, or drown in silos. File-based AI memory bridges that gap, turning raw files into contextual intelligence. 📂🧠
June 12, 2025 at 2:48 PM
Taken together, the results support the use of hyperparameter optimization as a routine part of deploying retrieval-augmented QA systems. Gains are possible and sometimes substantial, but they are also dependent on task design, metric selection, and evaluation procedure.
June 3, 2025 at 2:01 PM
We evaluate on three established multi-hop QA benchmarks: HotPotQA, TwoWikiMultiHop, and Musique. Each configuration is scored using one of three metrics: exact match (EM), token-level F1, or correctness.
June 3, 2025 at 2:01 PM
We present a structured study of hyperparameter
optimization in graph-based RAG systems, with a focus on tasks that combine unstructured inputs, knowledge graph construction, retrieval, and generation.
June 3, 2025 at 2:01 PM
Building AI memory and data pipelines to populate them is tricky. The performance of these pipelines depends
heavily on a wide range of configuration choices, including chunk size, retriever type, top-k thresholds, and prompt templates.
June 3, 2025 at 2:01 PM
Why does AI memory matter?

LLMs can’t give us details about our data, they "forget" or simply don’t know the details.
June 3, 2025 at 2:01 PM
Full write-up → www.cognee.ai/blog/fundame...

If you’re exploring how to blend vectors and graphs for richer retrieval, we build exactly that at @cognee.bsky.social - DMs open for a chat!
Cognee - Vector Databases Explained: A Smarter Way to Search by Meaning
Learn vector databases, how vector stores like Pinecone power semantic search and AI applications by indexing embeddings. Maximize their benefits with cognee now!
www.cognee.ai
May 21, 2025 at 4:34 PM
@PGvector

If your need something cheap and a way to get started, pgvector is the key. If you need something to run in production with large volumes, well, maybe you will run into trouble there. Still, it will do a lot of heavy lifting for you
May 21, 2025 at 4:34 PM