NativQA

Framework and benchmark for culturally aligned multilingual natural question answering

Framework + benchmark for multilingual LLM evaluation

Build culturally aligned natural QA datasets grounded in native speakers and local context.

NativQA is a scalable, language-independent framework for constructing question answering datasets in native languages. It supports both evaluation and fine-tuning of large language models, with MultiNativQA as a public benchmark built from regionally grounded, native-speaker queries.

Quick install pip install nativqa-framework

Explore resources

Framework

NativQA Framework

Use the framework to create culturally and regionally aligned QA datasets for multilingual LLM evaluation and tuning.

Dataset

MultiNativQA Dataset

See dataset links, download metrics, language coverage, and topic distribution in a dedicated resources page.

Why NativQA

Native-speaker grounded

Queries are sourced from native speakers, making evaluation data closer to real local information needs.

Culturally aligned

The benchmark emphasizes region-specific and culturally situated questions that generic QA sets often miss.

Evaluation + tuning ready

The same framework supports both benchmarking open- and closed-source LLMs and creating fine-tuning data.

news

latest posts

Jul 16, 2024 Just a moment...

publications

  1. nativqa_framework.png
    NativQA Framework: Enabling llms with native, local, and everyday knowledge
    Firoj Alam, Md Arid Hasan, Sahinur Rahman Laskar , and 3 more authors
    arXiv preprint arXiv:2504.05995, 2025
  2. language_donut_chart.png
    NativQA: Multilingual Culturally-Aligned Natural Query for LLMs
    Md. Arid Hasan, Maram Hasanain, Fatema Ahmad , and 6 more authors
    In Findings of the Association for Computational Linguistics ACL 2025 , Jul 2025