Nothing leaves your machine. Embeddings run locally via Ollama (or optionally Gemini/NIM). Sub-200ms semantic search over 1,000+ docs on M-series Mac.
Front-ends: Raycast extension, `recall` CLI, and an MCP server for AI assistants.
Would love feedback on the architecture and any ideas for new connectors.