approximate nearest neighbor vector search with warm/cold tiering
Executes sub-10ms vector similarity search on pre-computed embeddings using approximate nearest neighbor (ANN) algorithms with a two-tier memory architecture: hot data cached in NVMe SSD/memory for p50 latency of 8ms, cold data retrieved from S3 object storage on first access. Supports topk result limiting and operates at scale across 500M+ documents per namespace with observed throughput of 25k+ queries/second.
Unique: Separates compute and storage layers with S3-backed tiered caching (NVMe SSD + memory for hot data, object storage for cold), enabling 10x cost reduction vs alternatives while maintaining sub-10ms p50 latency on warm queries through intelligent cache management rather than keeping all vectors in-memory
vs alternatives: Cheaper than Pinecone/Weaviate at scale because it uses S3 for persistent storage instead of expensive managed vector storage, while maintaining competitive latency through SSD caching for frequently accessed namespaces
bm25 full-text search with metadata filtering
Performs keyword-based document retrieval using BM25 ranking algorithm combined with optional metadata filtering to narrow result sets by document attributes. Operates independently from vector search or in hybrid mode, with measured p50 latency of 343ms on warm namespaces. Metadata filter syntax and exact filtering capabilities are undocumented but support structured attribute-based result narrowing.
Unique: Integrates BM25 full-text search as a first-class capability alongside vector search within the same API, enabling hybrid search queries that combine both ranking signals without requiring separate search infrastructure or post-processing to merge results
vs alternatives: Simpler than maintaining separate Elasticsearch/Meilisearch instances for keyword search because full-text and vector search are unified in a single API with shared namespace isolation and S3 storage
api authentication and access control
Secures API access using API key-based authentication with undocumented header format and encoding. Supports role-based access control (RBPR) at Scale tier with SSO (single sign-on), and fine-grained permissions at Enterprise tier. Specific authentication mechanisms, token formats, and permission models are completely undocumented.
Unique: Tiered authentication where Launch uses basic API keys, Scale adds RBAC and SSO, and Enterprise adds fine-grained permissions, but all authentication mechanisms are undocumented making integration difficult
vs alternatives: unknown — cannot compare authentication security or usability to alternatives without API specification
multi-region deployment and data residency
Supports deployment across multiple AWS regions with data residency controls, but specific regions, latency characteristics, and failover behavior are completely undocumented. Region selection appears to be tied to S3 bucket location.
Unique: unknown — insufficient data on region availability, replication strategy, and failover behavior
vs alternatives: unknown — cannot assess multi-region capabilities without documentation
customer support and sla guarantees
Provides tiered support with Launch offering community support, Scale offering 8-5 business hours support with private Slack channel, and Enterprise offering 24/7 support with 99.95% uptime SLA. Specific response times, escalation procedures, and SLA terms are undocumented.
Unique: Tiered support model where Launch includes community support, Scale adds business hours support with private Slack, and Enterprise adds 24/7 support with 99.95% SLA, but SLA terms and support response times are undocumented
vs alternatives: More accessible than Pinecone for startups because Launch tier includes community support, though 24/7 support requires Enterprise tier like most SaaS products
hybrid vector + full-text search with combined ranking
Executes simultaneous vector and full-text search queries and combines their ranking signals to produce a unified result set that balances semantic similarity with keyword relevance. Implementation details of ranking combination (weighted sum, learning-to-rank, etc.) are undocumented, but enables use cases requiring both semantic and keyword precision without separate round-trips.
Unique: Provides native hybrid search combining vector and full-text signals in a single query without requiring application-level result merging or separate API calls, with unified ranking across both modalities within the same namespace isolation model
vs alternatives: More efficient than querying vector and full-text search separately and merging results in application code because ranking is unified server-side, reducing latency and eliminating deduplication logic
namespace-based multi-tenancy and data isolation
Isolates documents and queries into logical namespaces, enabling secure multi-tenant deployments where each tenant's data is completely segregated at the API level. Supports up to 100M+ namespaces with independent vector/full-text indexes, metadata schemas, and cache policies. Namespaces can be pinned (up to 256) to keep data in warm cache, or unpinned to use cold S3 storage for cost optimization.
Unique: Implements namespace-based isolation with optional pinning to control which tenants' data stays in warm cache vs cold S3, enabling fine-grained cost optimization where high-value tenants get guaranteed low latency while others use cheaper cold storage
vs alternatives: More cost-efficient than per-tenant Pinecone instances because multiple tenants share infrastructure with namespace isolation, and pinning allows selective warm caching instead of keeping all data hot
s3-backed persistent storage with tiered caching
Stores all vector and document data durably in AWS S3 object storage while maintaining a two-tier cache layer (NVMe SSD + memory) for hot data. On first query to a namespace, data is loaded from S3 into cache; subsequent queries hit the faster cache layer. Namespaces can be explicitly pinned to keep data in warm cache, or unpinned to allow cache eviction and S3 fallback for cost savings.
Unique: Decouples compute and storage by using S3 as the durable backend with intelligent tiered caching (NVMe SSD + memory) for hot data, enabling 10x cost reduction vs in-memory vector databases while maintaining sub-10ms latency for frequently accessed data through automatic cache management
vs alternatives: Cheaper than Weaviate/Milvus at scale because persistent storage is S3 (pay-per-GB) instead of expensive managed storage, while SSD caching prevents S3 latency from impacting warm queries
+5 more capabilities