jina-reranker-v3: Last but Not Late Interaction for Document Reranking
Abstract
A multilingual document reranker using causal self-attention achieves state-of-the-art performance with a compact architecture.
jina-reranker-v3 is a 0.6B parameter multilingual document reranker that introduces a novel last but not late interaction. Unlike late interaction models such as ColBERT that perform separate encoding followed by multi-vector matching, our approach conducts causal self-attention between query and documents within the same context window, enabling rich cross-document interactions before extracting contextual embeddings from the last token of each document. This compact architecture achieves state-of-the-art BEIR performance with 61.94 nDCG@10 while being ten times smaller than generative listwise rerankers.
Community
Hi, really impressive work there!
I needed a clarification, on page 5, the equation for dual matching loss ℓdual is mentioned to be Eq. but there is no Eq 4, also the link is pointing to page 4
This is an automated message from the Librarian Bot. I found the following papers similar to this paper.
The following papers were recommended by the Semantic Scholar API
- MetaEmbed: Scaling Multimodal Retrieval at Test-Time with Flexible Late Interaction (2025)
- Conan-Embedding-v2: Training an LLM from Scratch for Text Embeddings (2025)
- From Ranking to Selection: A Simple but Efficient Dynamic Passage Selector for Retrieval Augmented Generation (2025)
- ViRanker: A BGE-M3&Blockwise Parallel Transformer Cross-Encoder for Vietnamese Reranking (2025)
- Granite Embedding R2 Models (2025)
- Multi-view-guided Passage Reranking with Large Language Models (2025)
- fact check AI at SemEval-2025 Task 7: Multilingual and Crosslingual Fact-checked Claim Retrieval (2025)
Please give a thumbs up to this comment if you found it helpful!
If you want recommendations for any Paper on Hugging Face checkout this Space
You can directly ask Librarian Bot for paper recommendations by tagging it in a comment:
@librarian-bot
recommend
Models citing this paper 3
Datasets citing this paper 0
No dataset linking this paper
Spaces citing this paper 0
No Space linking this paper