Why did we open-source our inference engine? Read the post

nomic-ai/nomic-embed-text-v2-moe

Architecture
Parameters
137M
Tasks
Encode
Outputs
Dense
Dimensions
Dense: 768
Max Sequence Length
2,048 tokens
License

Benchmarks

CQADupstackPhysicsRetrieval

scientific retrieval en

Performance L4 b1 c16
Corpus TPS 13.0K
Corpus p50 149.6ms
Query TPS 1.2K
Query p50 143.2ms

CosQA

technology retrieval en

Performance L4 b1 c16
Corpus TPS 807
Corpus p50 595.7ms
Query TPS 139
Query p50 634.4ms

NanoFiQA2018Retrieval

finance retrieval en

Quality
ndcg at 10 0.5207
map at 10 0.4283
mrr at 10 0.5634
Performance L4 b1 c16
Corpus TPS 20.1K
Corpus p50 135.4ms
Query TPS 1.7K
Query p50 119.2ms

SCIDOCS

scientific retrieval en

Performance L4 b1 c16
Corpus TPS 2.4K
Corpus p50 1.3s
Query TPS 74
Query p50 1.7s

StackOverflowQA

technology retrieval en

Performance L4 b1 c16
Corpus TPS 24.1K
Corpus p50 145.6ms
Query TPS 33.4K
Query p50 142.9ms

Self-hosted inference for search & document processing

Cut API costs by 50x, boost quality with 85+ SOTA models, and keep your data in your own cloud.

Github
1.5K

Contact us

Tell us about your use case and we'll get back to you shortly.