So I tested it. I spun up 3 teammates backend, frontend, and tester on a full-stack project. Before the test, I pre-loaded the project knowledge: /nemp:init # auto-detected Next.js, TypeScript, Prisma /nemp:save auth "JWT with 15min access tokens, httpOnly refresh" /nemp:save api-style "RESTful, snake_case, versioned at /api/v1" /nemp:save testing "Pytest with fixtures, 80% coverage minimum" Then each teammate ran /nemp:context with their relevant keywords as their first action. The backend agent ran /nemp:context auth and immediately knew the JWT strategy. The frontend agent ran /nemp:context stack and knew it was Next.js + TypeScript. The tester ran /nemp:context testing and found the test conventions. It worked because the memory is just a JSON file on disk (.nemp/memories.json). No server, no port, no database. Every agent reads the same file. One saves a finding, the others discover it with a keyword search. End result: 34 files committed, 100 tests passing, zero conflicts between the 3 agents. Nobody had to re-discover what another agent already knew. The whole thing is a Claude Code plugin, two commands to install, no API keys, no cloud, nothing leaves your machine. Demo of the full test: https://vimeo.com/1162546825?share=copy&fl=sv&fe=ci I'm a solo developer building this in the open. If this is useful to you, a star on the repo would mean a lot, it helps other developers find it. https://github.com/SukinShetty/Nemp-memory Happy to answer any questions about the architecture or how it performed with agent teams.