1 pointby mnemoxai7 hours ago1 comment
  • mnemoxai7 hours ago
    Author here. Some context on why I built this:

    I use Claude Code and Cursor daily. The pattern I kept hitting: describe an idea → AI starts coding → hours later I search GitHub and realize there are already 10+ mature projects doing the same thing.

    The fix was simple: search BEFORE coding. But I wanted it automated — not a manual GitHub search, but something the AI agent does by default as part of its workflow.

    Technical details: - Keyword extraction uses a 3-stage dictionary pipeline (90+ intent anchors, 80+ synonym expansions, 150+ Chinese term mappings). No LLM needed for the MCP path — fast, free, deterministic. - Scoring is intentionally simple and explainable: weighted average across sources, not ML. You can read the formula in scoring/engine.py. - All sources are searched in parallel with asyncio.gather. Deep mode hits 5 APIs concurrently. - Zero storage — nothing is logged or saved. Protocol, not SaaS.

    The most useful integration pattern: add "always run idea_check before building anything new" to your CLAUDE.md or Cursor rules. The agent will automatically check before writing code.

    Honest limitation: dictionary-based keyword extraction has a coverage ceiling. Niche domains outside tech (e.g., "pet acupuncture booking app") may get less precise results. The web demo at mnemox.ai/check uses LLM extraction for better coverage.

    Would love feedback on the scoring formula — it's the part I'm least confident about. Is a single 0-100 number useful, or would you prefer something more granular?