$ -weight: 500;">git clone https://github.com/vinaybudideti/intent-atoms.-weight: 500;">git cd intent-atoms python -m venv venv && source venv/bin/activate -weight: 500;">pip -weight: 500;">install -r requirements.txt python tests/benchmark_100.py # run the 100-query benchmark yourself COMMAND_BLOCK: -weight: 500;">git clone https://github.com/vinaybudideti/intent-atoms.-weight: 500;">git cd intent-atoms python -m venv venv && source venv/bin/activate -weight: 500;">pip -weight: 500;">install -r requirements.txt python tests/benchmark_100.py # run the 100-query benchmark yourself COMMAND_BLOCK: -weight: 500;">git clone https://github.com/vinaybudideti/intent-atoms.-weight: 500;">git cd intent-atoms python -m venv venv && source venv/bin/activate -weight: 500;">pip -weight: 500;">install -r requirements.txt python tests/benchmark_100.py # run the 100-query benchmark yourself - Tier 1 — Direct hit (similarity > 0.85): Return cached response. Zero cost. ~97ms. This caught 54 of 100 queries. - Tier 2 — Adapt (similarity 0.70–0.85): Take the closest cached response and use Haiku to tweak it for the new query. ~$0.002. This caught 35 queries. - Tier 3 — Full miss (similarity < 0.70): Fall through to atom-level decomposition. Only 9 queries reached this tier. - Embeddings: sentence-transformers/all-mpnet-base-v2 (768-dim, runs locally — no API cost for embedding) - Vector search: FAISS IndexFlatIP (cosine similarity via inner product on normalized vectors) - LLM providers: Anthropic Claude — Haiku for cheap operations (decompose, adapt, compose), Sonnet for generation - API: FastAPI with async support - Dashboard: React + Recharts, deployable to Vercel - Persistence: JSON metadata + binary FAISS index files