SHARE
SHARE
SHARE
Gaurav Bhattacharya
Gaurav Bhattacharya

CEO, Jeeva AI

July 14, 2025

How Jeeva AI Hits <2s Response Latency Using Hybrid Vector Search

How Jeeva AI Hits <2s Response Latency Using Hybrid Vector Search

How Jeeva AI Hits <2s Response Latency Using Hybrid Vector Search

How Jeeva AI Hits <2s Response Latency Using Hybrid Vector Search

Gaurav Bhattacharya
Gaurav Bhattacharya
Gaurav Bhattacharya

CEO, Jeeva AI

July 14, 2025

Real-Time AI Search, Powered by Jeeva
Real-Time AI Search, Powered by Jeeva
Real-Time AI Search, Powered by Jeeva
Real-Time AI Search, Powered by Jeeva

Hook: You Can’t Wait Three Seconds for That Perfect Prospect

Imagine you’re a sales rep, about to follow up on a hot lead. You need that perfect list of prospects now. But if it takes more than two seconds to load, your flow breaks, your focus shifts, and your conversion odds drop. In 2025, user patience is razor-thin: 53% of mobile visitors bail if a page takes over 3 seconds, and every extra second slashes form fills, conversions, and SEO dwell time. For modern AI-driven sales teams, sub-2s answers aren’t a luxury; they're the new standard.

This is the story of how Jeeva AI built a lightning-fast, billion-scale retrieval pipeline using hybrid vector search and a series of engineering tricks that turn “good” latency into “great.”

Why Sub-2s Is the New Non-Negotiable

Google’s Core Web Vitals update now factors API-driven content delays into page quality and SEO scores. If your sales agent platform delivers “top 5 prospects” in over 2 seconds, you risk both higher abandonment and lower search rankings. For real-time workflows, speed isn’t just about UX it's about more deals, more pipeline, and happier users.

What is Hybrid Vector Search?

Hybrid vector search combines two retrieval techniques lexical (BM25) and semantic (vector) fusing their results for higher recall and relevance than either method alone. This fusion keeps results fast and on-topic, even at a billion-document scale.

Inside Jeeva AI’s <2s Retrieval Pipeline

Here’s how Jeeva AI achieves consistently fast, accurate results even with huge datasets:

1. Trigger:
Agent receives a real-time context (ICP filters, intent, keywords).

2. Embedding Cache:
Query embeddings are pre-computed for 40% of recurring prompts; misses fall back to a GPU encoder (<35 ms).

3. Metadata Filtering (≈ 20 ms):
Lexical WHERE clauses (e.g., industry, ARR band) trim the pool from 120M to 3M.

4. Hybrid Search Stage-1 (≈ 50 ms):

  • Sparse BM25 on Elastic cluster returns top-800 docs.

  • Dense ANN on Pinecone returns top-800 vectors.

  • Convex-combination fusion creates a 400-item shortlist.

5. Re-rank with Cross-Encoder (≈ 70 ms):
A mini-LM model re-scores the top 50 candidates.

6. Edge Cache & Serialization (≈ 15 ms):
Results are streamed as JSON to the agent UI.

Result:
Total p95 wall time is 1.78s, leaving headroom for network variance.

Algorithms & Infrastructure Choices

Layer

Tech

Why We Chose It

Hot tier

Pinecone HNSW pods

p50 <10ms at 1B vectors, minimal cold-start latency

Warm tier

Weaviate (BM25 + vector fusion)

Built-in convex combination, serverless autoscale

Cold tier

DiskANN on Azure Cosmos vector

~5ms query time, 80% cheaper RAM footprint

Re-rank

MiniLM-L6-v3 cross-encoder (T4 GPU)

15× cheaper than A100, <1ms/token, enables real-time re-ranking

Pipeline orchestration runs on Kubernetes with KEDA, auto-scaling worker pods based on Pub/Sub queue depth to prevent back-pressure during traffic spikes.

Engineering Tricks That Shave the Last Millisecond

Trick

Latency Win

How it Works

Locality-aware sharding

–120 ms RTT

Pinecone routes queries to RAM-resident shards, avoiding cross-AZ hops

Batch-and-dash embeds

–40 ms

Batch encode queued queries on GPU for better throughput

Early exit on high-score gap

–25 ms

Skip re-ranking if top-1/average >1.5

Edge caching ICP filters

–30 ms

Store frequent queries (e.g., “US SaaS CEOs”) in Cloudflare KV

gRPC instead of REST

–15 ms

Binary Protobuf cuts wire payload by ~60%

These micro-optimizations collectively trim ≈200ms off the p95 path transforming 1.9s into a market-leading 1.7s experience.

Benchmarks (May 2025 Jeeva Internal Test)

Corpus Size

Search Mode

Median Latency

Recall@10

10M

Pure BM25

430 ms

0.62

10M

Pure HNSW

55 ms

0.71

10M

Hybrid (fusion)

72 ms

0.85

1B

DiskANN (cold)

5 ms

0.76

1B

Jeeva pipeline

1.78 s

0.87

Hybrid adds only ~17 ms over ANN alone, but improves topical recall by 14 percentage points worth every millisecond.

Business Impact

  • Lead-suggestion modal loads 35% faster after hybrid rollout, increasing rep click-through by 18%.

  • Support CPU savings: Elastic 8 nodes replace OpenSearch at 1/3 the size, with 12× query-time gains.

  • Lower infra cost: DiskANN’s SSD-based graph cut RAM costs by ~60% while holding 95% recall.

Performance improvements

Latency Benchmarks in US-East vs. EU-Central Data-Centres

Jeeva AI’s hybrid pipeline maintains sub-2s latency in both US-East and EU-Central regions, with gRPC and edge caching mitigating transatlantic hops. For APAC clients, early exit logic and regionally pinned shards sustain a 1.95s p95 SLA even at peak volumes.

How Jeeva Queried 1 Billion Vectors in <2 Seconds (How-To)

  1. Pre-compute and cache frequent query embeddings.

  2. Slice candidate pool with fast metadata filters.

  3. Use hybrid BM25 + ANN search to create a ranked shortlist.

  4. Re-rank top results with a lightweight mini-LM cross-encoder.

  5. Edge-cache high-frequency queries and serialize output in lightweight JSON.

  6. Monitor wall time and tune for network variance using KEDA autoscaling.

FAQ: Hybrid Vector Search & Latency

Q1. What is hybrid vector search?
Hybrid search combines lexical (BM25) and semantic (vector) retrieval, fusing results for higher recall and relevance. (Elastic)

Q2. How fast can modern vector databases respond?
Managed engines like Pinecone show p50 latencies under 10 ms and p99 under 50 ms—even at billion-scale. (Aloa)

Q3. Isn’t DiskANN just for research?
No. Platforms like Azure Cosmos and Milvus deliver ~5 ms queries with 95% recall on billion-vector corpora. (Milvus)

Q4. Why aim for <2s end-to-end?
UX studies show 53% of users abandon if waits exceed 3 seconds; for AI agents, fast answers = higher conversions. 

Q5. How does Jeeva keep costs sane?
DiskANN on SSD, aggressive caching, and early-exit logic dramatically cut both cloud costs and carbon footprint.

Fuel Your Growth with AI

Fuel Your Growth with AI

Ready to elevate your sales strategy? Discover how Jeeva’s AI-powered tools streamline your sales process, boost productivity, and drive meaningful results for your business.

Ready to elevate your sales strategy? Discover how Jeeva’s AI-powered tools streamline your sales process, boost productivity, and drive meaningful results for your business.

Stay Ahead with Jeeva

Stay Ahead with Jeeva

Get the latest AI sales insights and updates delivered to your inbox.

Get the latest AI sales insights and updates delivered to your inbox.