TensorPool Logo
Technical

How ZeroEntropy Trained the zerank Model Family on TensorPool

How ZeroEntropy's reranker models achieved SOTA performance through innovative training methods and elastic GPU access.

Tycho Svoboda
Tycho SvobodaCofounder & CEO
November 22, 2025
4 min read

Summarize with AI

ZeroEntropy's zerank series has emerged as dominant open-weight rerankers across various verticals like customer support, legal, finance, medical, and code. Their breakthrough came from a novel training approach using synthetic pairwise preference annotations turned into Elo scores for ranking. Both zerank-1 and the recently launched zerank-2 were trained on TensorPool.

Models at a Glance

zerank is a family of open source state-of-the-art rerankers. They are cross-encoders that are LoRA fine-tunes of the Qwen3 model family.

  • zerank-1 and zerank-2 are 4B parameters and finetuned on a Qwen3-4B base. They're released under the Creative Commons Attribution Non Commercial 4.0 license.
  • The compact variant, zerank-1-small, is built on a Qwen3-1.7B base that is open-source under the Apache 2.0 license; it has 1.7B parameters.

NDCG@10 Rerankers Comparison

Latency by Payload Size - ZeroEntropy vs Commercial Rerankers

zerank's state-of-the-art performance has led it to dominate in RAG deployments for critical domains such as customer support, legal, finance, medical, and code domains.

What Makes zerank Special

There are a couple characteristics we think are under appreciated:

Future Ready Training Pipeline

ZeroEntropy's novel training approach: zELO. Their pipeline combines synthetic pairwise supervision with an Elo-style ranking scheme. This means that training data can be regenerated as LLMs improve. This grants them a free lunch (aside from the token cost) without the logistical drag or expense of human annotation.

ZeroEntropy Punches Above Their Weight

The launch of zerank-1 in September 2024 showed how well ZeroEntropy is able to execute despite few resources. Cohere was navigating a complex reorganization following their $500M Series D (~$1.5B raised to date). ZeroEntropy came fresh out of their $4.2M seed and leapfrogged the large well-resourced incumbents.

How TensorPool Helped

TensorPool's CLI and elastic access to GPUs enabled ZeroEntropy to speed up their experimentation cycle. This enabled them to iterate faster, test more ideas, and focus on research rather than infrastructure logistics.

We chose TensorPool because they not only provide easy to use infrastructure, but also go out of their way to ensure we get the best bang for our buck."

— Nicholas Pipitone, CTO of ZeroEntropy

Conclusion

ZeroEntropy demonstrates that smart architecture and training methodology outpaces pure capital deployment. zELO, ZeroEntropy's novel training pipeline, is the backbone of their SOTA reranker today and is positioned to be even better tomorrow. TensorPool enabled ZeroEntropy to compress their research timeline, turning the R&D cost center into a competitive advantage.

The result is the zerank reranker family that's reshaping retrieval in domains where precision matters most—proving speed and research focus matter more than headcount or funding rounds.

To learn more about ZeroEntropy and the zELO training pipeline, check out their blog!

Ready to Get Started?

Sign up for TensorPool and start building on powerful GPU infrastructure.