|
|
--- |
|
|
configs: |
|
|
- config_name: default |
|
|
data_files: |
|
|
- split: test |
|
|
path: qrels/test.jsonl |
|
|
- config_name: corpus |
|
|
data_files: |
|
|
- split: corpus |
|
|
path: corpus.jsonl |
|
|
- config_name: queries |
|
|
data_files: |
|
|
- split: queries |
|
|
path: queries.jsonl |
|
|
--- |
|
|
## Dataset Summary |
|
|
|
|
|
**NQ-Fa** is a Persian (Farsi) dataset created for the **Retrieval** task, specifically targeting **open-domain question answering**. It is a **translated version** of the original English **Natural Questions (NQ)** dataset and a central component of the [FaMTEB (Farsi Massive Text Embedding Benchmark)](https://huggingface.co/spaces/mteb/leaderboard), as part of the **BEIR-Fa** collection. |
|
|
|
|
|
- **Language(s):** Persian (Farsi) |
|
|
- **Task(s):** Retrieval (Question Answering) |
|
|
- **Source:** Translated from English NQ using Google Translate |
|
|
- **Part of FaMTEB:** Yes — under BEIR-Fa |
|
|
|
|
|
## Supported Tasks and Leaderboards |
|
|
|
|
|
This dataset evaluates how well **text embedding models** can retrieve relevant answer passages from Persian Wikipedia in response to **natural language questions**, originally issued to Google Search. Results are benchmarked on the **Persian MTEB Leaderboard** on Hugging Face Spaces (language filter: Persian). |
|
|
|
|
|
## Construction |
|
|
|
|
|
The construction process included: |
|
|
|
|
|
- Starting with the **Natural Questions (NQ)** English dataset, containing real user search queries |
|
|
- Using the **Google Translate API** to translate both questions and annotated Wikipedia passages into Persian |
|
|
- Retaining original query-passage mapping structure for retrieval evaluation |
|
|
|
|
|
As described in the *FaMTEB* paper, all BEIR-Fa datasets (including NQ-Fa) underwent: |
|
|
|
|
|
- **BM25 retrieval comparison** between English and Persian |
|
|
- **LLM-based translation quality check** using the GEMBA-DA framework |
|
|
|
|
|
These evaluations confirmed a **high level of translation quality**. |
|
|
|
|
|
## Data Splits |
|
|
|
|
|
Defined in the FaMTEB paper (Table 5): |
|
|
|
|
|
- **Train:** 0 samples |
|
|
- **Dev:** 0 samples |
|
|
- **Test:** 2,685,669 samples |
|
|
|
|
|
**Total:** ~2.69 million examples (according to metadata) |
|
|
|