News & Archive

Deep dives, tutorials, and technical explorations from the ZeroEntropy team.

Zemail: Semantic Gmail Search on Claude Code & Cowork
Apr 15, 2026AI

Zemail: Semantic Gmail Search on Claude Code & Cowork

Zemail is a free Claude Code/Cowork plugin that builds a local semantic index of your Gmail inbox. Keyword search can't find the email you're thinking of. A reranker can.

Read
The Best Embedding Model for Conversational AI in 2026
Apr 10, 2026Search

The Best Embedding Model for Conversational AI in 2026

zembed-1 achieves the highest conversational retrieval benchmark score with a +35% advantage over OpenAI, bridging the gap between informal user queries and formal knowledge bases.

Read
The Best Embedding Model for Finance in 2026: Why zembed-1 Wins
Apr 10, 2026Search

The Best Embedding Model for Finance in 2026: Why zembed-1 Wins

zembed-1 outperforms all benchmarked competitors on finance-domain retrieval, with a 32k context window, flexible compression, and Elo-calibrated relevance for regulatory compliance, earnings analysis, and investment research.

Read
The Best Embedding Model for Code in 2026: zembed-1 Tops the Leaderboard
Apr 10, 2026Search

The Best Embedding Model for Code in 2026: zembed-1 Tops the Leaderboard

zembed-1 achieves the highest benchmark score of any embedding model in the code domain at 0.6452 NDCG@10, while simultaneously leading every other domain -- the only model that is best-in-class everywhere.

Read
The Best Embedding Model for Healthcare in 2026: zembed-1 Leads the Field
Apr 10, 2026Search

The Best Embedding Model for Healthcare in 2026: zembed-1 Leads the Field

zembed-1 achieves 0.6260 NDCG@10 on healthcare retrieval benchmarks, leading competitors by up to +31.8%, with multilingual support, 32k context, and self-hosting for HIPAA compliance.

Read
The Best Embedding Model for Legal in 2026: zembed-1 Sets the Standard
Apr 10, 2026Search

The Best Embedding Model for Legal in 2026: zembed-1 Sets the Standard

zembed-1 achieves 0.6723 NDCG@10 on legal retrieval benchmarks, outperforming all competitors by up to +31.8%, with Elo-calibrated relevance, 32k context, and quantization for massive legal corpora.

Read
The Best Embedding Model for Manufacturing in 2026
Apr 10, 2026Search

The Best Embedding Model for Manufacturing in 2026

zembed-1 leads manufacturing domain benchmarks with a +12.9% advantage over the nearest competitor, powering industrial AI retrieval for maintenance, quality, and engineering applications.

Read
The Best Embedding Model for RAG and Enterprise Search in 2026
Apr 10, 2026Search

The Best Embedding Model for RAG and Enterprise Search in 2026

zembed-1 achieves the highest NDCG@10 on MSMARCO and leads every domain-specific benchmark, making it the ideal foundation for enterprise RAG systems.

Read
The Best Embedding Model for STEM and Mathematics in 2026
Apr 10, 2026Search

The Best Embedding Model for STEM and Mathematics in 2026

zembed-1 leads STEM and mathematics retrieval benchmarks, outperforming OpenAI by 35% and all competitors in scientific document search.

Read
The Best Embedding Model of 2026: Why zembed-1 Is in a Class of Its Own
Apr 10, 2026Search

The Best Embedding Model of 2026: Why zembed-1 Is in a Class of Its Own

zembed-1 is the first embedding model to lead every domain benchmark simultaneously — finance, healthcare, legal, manufacturing, code, STEM, and conversational — while topping MSMARCO at 0.946 NDCG@10.

Read
The Best Multilingual Embedding Model in 2026: zembed-1 Was Built for the World
Apr 10, 2026Search

The Best Multilingual Embedding Model in 2026: zembed-1 Was Built for the World

zembed-1 was built with over 50% non-English training data, delivering true multilingual parity and cross-lingual retrieval across all major world languages with 0.946 NDCG@10 on MSMARCO.

Read
harrier-27b: Can 27B Parameters Beat zembed-1?
Apr 8, 2026Search

harrier-27b: Can 27B Parameters Beat zembed-1?

A head-to-head evaluation of ZeroEntropy's zembed-1 against all three Harrier embedding models across 24 diverse datasets, using graded LLM relevance judgments.

Read
Smarter Context Compression for LLM Pipelines: zerank-2 as a Calibrated Classifier
Apr 2, 2026AI

Smarter Context Compression for LLM Pipelines: zerank-2 as a Calibrated Classifier

How to use zerank-2's calibrated relevance scores as a binary classifier for context compression, document routing, and multi-label classification — at 50-100x less cost than LLM classification.

Read
Beyond Binary: A New Version of the MTEB
Mar 28, 2026Search

Beyond Binary: A New Version of the MTEB

We re-annotated 24 MTEB datasets with LLM pointwise scoring on a 0-10 scale using three independent judges. With graded relevance, the retrieval leaderboard looks meaningfully different.

Read
Bi-Encoders vs Cross-Encoders
Mar 18, 2026Search

Bi-Encoders vs Cross-Encoders

Why production search uses two models, not one — and what happens when you pair them.

Read
zembed-1 vs voyage-4
Mar 5, 2026Search

zembed-1 vs voyage-4

A thorough evaluation comparing the performance of ZeroEntropy's zembed-1 and Voyage's latest voyage-4 embedding models.

Read
"Let's eat, grandma" vs "let's eat grandma": how embedding models encode the world
Mar 2, 2026AI

"Let's eat, grandma" vs "let's eat grandma": how embedding models encode the world

A deep dive into how embedding models encode meaning, why famous training examples create the illusion of capability, and what consistent behavior across 10k+ nouns tells us about genuine understanding.

Read
Introducing zembed-1: The World's Best Text-Embedding Model
Mar 2, 2026Search

Introducing zembed-1: The World's Best Text-Embedding Model

Introducing zembed-1: The World's Best Text-Embedding Model

Read
2026's Top 10 Embedding Companies Powering Search Technology
Feb 23, 2026AI

2026's Top 10 Embedding Companies Powering Search Technology

Read
Best Embedding Model for Legal Document Search in 2026
Feb 23, 2026AI

Best Embedding Model for Legal Document Search in 2026

Read
The Geometric Limitations of Vector Embeddings in Retrieval Systems
Feb 23, 2026AI

The Geometric Limitations of Vector Embeddings in Retrieval Systems

Read
How Assembled Powers High-Quality AI Customer Support with ZeroEntropy
Feb 12, 2026AI

How Assembled Powers High-Quality AI Customer Support with ZeroEntropy

After integrating ZeroEntropy's reranking into their retrieval pipeline and validating it on live production traffic, Assembled migrated to 100% of their reranking volume through ZeroEntropy.

Read
Prompting Best Practices For Instruction-Following Rerankers
Jan 22, 2026AI

Prompting Best Practices For Instruction-Following Rerankers

In this guide, we will go over best practices around prompting a reranker for state-of-the-art relevance.

Read
Open-source alternatives to Cohere Rerank in 2026
Jan 16, 2026AI

Open-source alternatives to Cohere Rerank in 2026

Learn about the open-source and open-weight rerankers to use as an alternative to closed-source models.

Read
Latency Performance Assessment of zerank-2
Dec 9, 2025AI

Latency Performance Assessment of zerank-2

In this blog, we go over the latency performance of ZeroEntropy's latest reranker model zerank-2.

Read
Introducing zerank-2: The Most Accurate Multilingual Instruction-Following Reranker
Nov 18, 2025AI

Introducing zerank-2: The Most Accurate Multilingual Instruction-Following Reranker

zerank-2 is ZeroEntropy's latest multilingual cross-encoder model, tuned for instruction-following, and score calibration.

Read
The Latency Myth: Why Reranking Is Still the Smartest Optimization You Can Make
Oct 24, 2025AI

The Latency Myth: Why Reranking Is Still the Smartest Optimization You Can Make

Yes, technically, rerankers add a layer to your search pipeline. Yet, reranking improves both efficiency and quality once you consider the full retrieval-generation loop.

Read
Context Engineering Webinar: Everything You Missed
Oct 22, 2025AI

Context Engineering Webinar: Everything You Missed

Thanks to everyone who joined our first Context Engineering Webinar!

Read
How Vera Health Achieved State-of-the-Art Clinical Accuracy Using ZeroEntropy
Oct 14, 2025AI

How Vera Health Achieved State-of-the-Art Clinical Accuracy Using ZeroEntropy

Vera Health sets a new accuracy record on medical question answering using ZeroEntropy Search and Rerank APIs.

Read
Equall Improves Legal Document Structuring and Retrieval Accuracy with ZeroEntropy
Oct 9, 2025AI

Equall Improves Legal Document Structuring and Retrieval Accuracy with ZeroEntropy

ZeroEntropy's reranker consistently outperformed every alternative we tested — recall@10, recall@5, across datasets. It's fast, stable, and a no-brainer for legal document pipelines.

Read
Implementing ZeroEntropy Reranking with turbopuffer Retrieval
Oct 5, 2025AI

Implementing ZeroEntropy Reranking with turbopuffer Retrieval

Learn how to implement a two-step search pipeline for fast and accurate retrieval, using turbopuffer and ZeroEntropy.

Read
Paper TLDR: How we trained zerank-1 with the zELO method
Sep 19, 2025AI

Paper TLDR: How we trained zerank-1 with the zELO method

This is a short TLDR summarizing the main concepts in the zELO paper. We describe the training approach of ZeroEntropy's reranker zerank-1.

Read
Mem0 Improves Memory Retrieval Accuracy with ZeroEntropy
Sep 18, 2025AI

Mem0 Improves Memory Retrieval Accuracy with ZeroEntropy

Mem0 migrated their production rerank traffic to ZeroEntropy's zerank-1, a critical component of their retrieval stack.

Read
On The Geometric Limit of Dense Single Vector Embeddings
Sep 6, 2025AI

On The Geometric Limit of Dense Single Vector Embeddings

Learn about why embeddings are usually not enough in a search pipeline, and how you can fix some of the most common pitfalls embeddings have.

Read
Should You Use LLMs for Reranking? A Deep Dive into Pointwise, Listwise, and Cross-Encoders
Sep 5, 2025AI

Should You Use LLMs for Reranking? A Deep Dive into Pointwise, Listwise, and Cross-Encoders

Are LLMs cheap, calibrated, and fast enough for reranking tasks? How should you select the best results from a candidate set of k=100, or even k=200? We go over the pros and cons of using LLMs as rerankers and what the best approach is.

Read
My AskAI Improves Support Agent Latency and Accuracy with ZeroEntropy
Sep 4, 2025AI

My AskAI Improves Support Agent Latency and Accuracy with ZeroEntropy

My AskAI replaced its existing reranker with ZeroEntropy's zerank‑1 across production traffic. Results: faster responses at scale, a measurable lift in answer quality, and lower cost.

Read
Ultimate Guide to Choosing the Best Reranking Model in 2026
Aug 21, 2025AI

Ultimate Guide to Choosing the Best Reranking Model in 2026

How to pick a reranker model for your RAG stack in 2025?

Read
How to Do RAG with Mastra and ZeroEntropy
Aug 17, 2025AI

How to Do RAG with Mastra and ZeroEntropy

Building Production RAG with Mastra and ZeroEntropy Reranking

Read
Why Evaluation Metrics for Reranking Matter in Search Quality
Aug 12, 2025AI

Why Evaluation Metrics for Reranking Matter in Search Quality

Learn why evaluation metrics for reranking are crucial in improving search engine performance. Discover how they ensure more accurate and relevant search results.

Read
ZeroEntropy's zerank-1 vs. Jina AI's jina-reranker-m0
Aug 7, 2025AI

ZeroEntropy's zerank-1 vs. Jina AI's jina-reranker-m0

How does Jina AI's jina-reranker-m0 compare to ZeroEntropy's zerank-1? In this blog post, we compare both on the latency, accuracy, and pricing standpoints.

Read
Best Reranker for HR Knowledge Bases
Jul 30, 2025AI

Best Reranker for HR Knowledge Bases

Automate employee Q&A, HR policy lookups, job search and match, with smarter reranking.

Read
Latency Benchmark: Cohere rerank 3.5 vs. ZeroEntropy zerank-1
Jul 22, 2025AI

Latency Benchmark: Cohere rerank 3.5 vs. ZeroEntropy zerank-1

zerank-1 delivers the fastest, most accurate reranking for AI search—up to 14× faster than Jina and 12% faster than Cohere. Try it via API or Hugging Face.

Read
2-Norm Vector
Jul 20, 2025AI

2-Norm Vector

Understanding 2-Norm Vectors in Vector Similarity Calculations

Read
Should You Use an LLM as a Reranker? Pros, Cons, and Benchmarks
Jul 20, 2025AI

Should You Use an LLM as a Reranker? Pros, Cons, and Benchmarks

Explore whether using an LLM as a reranker makes sense — with tradeoffs, latency impacts, and real-world data.

Read
Best Reranker for Healthcare AI
Jul 15, 2025AI

Best Reranker for Healthcare AI

Boost clinical search accuracy and patient data retrieval with domain-specific reranking.

Read
Best Reranker for Legal Document Search
Jul 14, 2025AI

Best Reranker for Legal Document Search

Transform legal research workflows with intelligent, explainable rerankers.

Read
Announcing ZeroEntropy's First Rerankers: zerank-1 and zerank-1-small
Jul 10, 2025AI

Announcing ZeroEntropy's First Rerankers: zerank-1 and zerank-1-small

Meet zerank-1: ZeroEntropy's powerful new reranker, outperforming Cohere and Gemini with up to 28% higher precision—now live via API and Hugging Face.

Read
ZeroEntropy Raises $4.2M Seed Round to Make AI Retrieval Truly Intelligent
Jul 9, 2025General

ZeroEntropy Raises $4.2M Seed Round to Make AI Retrieval Truly Intelligent

ZeroEntropy raises $4.2M to reinvent AI retrieval—enabling lightning-fast, accurate search across unstructured data. Backed by YC, Initialized & more.

Read
Improving Retrieval with ELO Scores
Jul 8, 2025AI

Improving Retrieval with ELO Scores

ZeroEntropy uses chess-inspired ELO scores to train the next-gen reranker—boosting accuracy with nuanced, scalable relevance scoring. Try zerank-1 today.

Read
What is a reranker and do I need one?
Jun 24, 2025AI

What is a reranker and do I need one?

Learn what a reranker is and how it boosts precision in RAG pipelines—surface the right results faster, reduce hallucinations, and improve LLM output quality.

Read
Deep Dive:  The Architecture of ZeroEntropy v1
May 14, 2025AI

Deep Dive: The Architecture of ZeroEntropy v1

Discover how ZeroEntropy v1 blends BM25, dense embeddings, and LLMs for lightning-fast, accurate hybrid search over unstructured documents.

Read
AGI requires better retrieval, not just better LLMs
Dec 2, 2024AI

AGI requires better retrieval, not just better LLMs

AGI needs more than LLMs—it needs smarter retrieval. Learn how to identify failure modes in RAG and evaluate search accuracy with ZeroEntropy's benchmarks.

Read
LlamaChunk: A General and Cost Efficient Approach to Semantic Chunking
Dec 1, 2024AI

LlamaChunk: A General and Cost Efficient Approach to Semantic Chunking

Learn how LlamaChunk delivers fast, accurate semantic chunking for RAG—outperforming regex and embedding methods with LLM-guided document splitting.

Read
LegalBench-RAG, the First Open-Source Retrieval Benchmark for the Legal Domain
Nov 29, 2024AI

LegalBench-RAG, the First Open-Source Retrieval Benchmark for the Legal Domain

LegalBench-RAG is the first open-source benchmark for legal RAG retrieval—6,800+ queries, 79M+ characters, human-annotated spans. Evaluate legal AI today.

Read
ZeroEntropy
The best AI teams retrieve with ZeroEntropy
Follow us on
GitHubTwitterSlackLinkedInDiscord