Instructions to use zeroentropy/zerank-2-reranker with libraries, inference providers, notebooks, and local apps. Follow these links to get started.
- Libraries
- sentence-transformers
How to use zeroentropy/zerank-2-reranker with sentence-transformers:
from sentence_transformers import CrossEncoder model = CrossEncoder("zeroentropy/zerank-2-reranker") query = "Which planet is known as the Red Planet?" passages = [ "Venus is often called Earth's twin because of its similar size and proximity.", "Mars, known for its reddish appearance, is often referred to as the Red Planet.", "Jupiter, the largest planet in our solar system, has a prominent red spot.", "Saturn, famous for its rings, is sometimes mistaken for the Red Planet." ] scores = model.predict([(query, passage) for passage in passages]) print(scores) - Notebooks
- Google Colab
- Kaggle
vLLM support?
Hi,
Is there any vLLM support planned in the near future? I'd like to try this model locally but it doesn't seem possible for the moment
Thanks
Hi,
I can get it to work but the scores make no sense, so probably broken in a way. You can run it by adding command line args --hf-overrides '{"architectures": ["Qwen3ForSequenceClassification"], "classifier_from_token": ["no", "yes"], "is_original_qwen3_reranker": true}' . The reranking scores seem random though. Here's my test case:
curl -s -X POST "http://rtx5070:8000/v1/score" -H "Content-Type: application/json" \
-d '
{
"model": "zeroentropy/zerank-2",
"text_1": "What are the symptoms of diabetes?",
"text_2": [
"Common diabetes symptoms include increased thirst, frequent urination, fatigue, and blurred vision.",
"The Great Wall of China is over 13000 miles long.",
"Basketball was invented by James Naismith in 1891.",
"Type 2 diabetes can often be managed with diet and exercise.",
"Symptoms may also include slow-healing wounds and tingling in hands or feet."],
"use_activation": false
}'
This gives these scores:
Doc Raw Logit Softmax Score Content
0 4.25 0.986 Diabetes symptoms (thirst, urination...)
4 6.51 0.998 More symptoms (wounds, tingling)
3 2.87 0.947 Diabetes management
1 2.74 0.940 Great Wall (irrelevant)
2 2.67 0.936 Basketball (irrelevant)
The raw logits show better differentiation (4.25 vs 2.74 for relevant vs irrelevant), but after softmax they compress to 0.986 vs 0.940. In any case the scores are not great, with irrelevant documents getting half the score that a fully relevant document is getting. I assume this is caused by the way of invocation.