I wanted to see exactly how the X "For You" algorithm decides whether to recommend your tweet or not, so I built this 'playable' CRT-styled simulator that runs a simpler version of X's recommendation pipeline right in your browser.
It's a "playable documentary" that allows you to type raw text and watch it physically travel through tokenization, embedding, safety filtering, and final ranking, all based on the X-AI reference docs.
Under the hood it's running real local ML inference (a tiny embedding model using Transformers.js) to generate embeddings and score content on the fly - all locally on your machine.
Would love some feedback!
I had Claude create a task system, hook it up to a `/task` command, then can use `bun ralph` to churn through any open tasks.
Try it out and feel free to contribute!
Hey HN,
I’ve spent the last few months building Remember-Me, an open-source "Sovereign Brain" stack designed to run entirely offline on consumer hardware.
The core thesis is simple: Don't rent your cognition.
Most RAG (Retrieval Augmented Generation) implementations are just "grep for embeddings." They are messy, imprecise, and prone to hallucination. I wanted to solve the "Context integrity" problem at the architectural layer.
The Tech Stack (How it works):
QDMA (Quantum Dream Memory Architecture): instead of a flat vector DB, it uses a hierarchical projection engine. It separates "Hot" (Recall) from "Cold" (Storage) memory, allowing for effectively infinite context window management via compression.
CSNP (Context Switching Neural Protocol) - The Hallucination Killer: This is the most important part. Every memory fragment is hashed into a Merkle Chain. When the LLM retrieves context, the system cryptographically verifies the retrieval against the immutable ledger.
If the hash doesn't match the chain: The retrieval is rejected.
Result: The AI visually cannot "make things up" about your past because it is mathematically constrained to the ledger. Local Inference: Built on top of llama.cpp server. It runs Llama-3 (or any GGUF) locally. No API keys. No data leaving your machine.
Features:
Zero-Dependency: Runs on Windows/Linux with just Python and a GPU (or CPU).
Visual Interface: Includes a Streamlit-based "Cognitive Interface" to visualize memory states. Open Source: MIT License. This is an attempt to give "Agency" back to the user. I believe that if we want AGI, it needs to be owned by us, not rented via an API.
Repository: https://github.com/merchantmoh-debug/Remember-Me-AI
I’d love to hear your feedback on the Merkle-verification approach. Does constraining the context window effectively solve the "trust" issue for you?
It's fully working - Fully tested. If you tried to Git Clone before without luck - As this is not my first Show HN on this - Feel free to try again.
To everyone who HATES AI slop; Greedy corporations and having their private data stuck on cloud servers.
You're welcome.
Cheers, Mohamad