Skip to content

FarExStance: The first and largest claim-based explainable stance detection dataset on Farsi

Notifications You must be signed in to change notification settings

Zarharan/FarExStance

Repository files navigation

FarExStance: Explainable Stance Detection for Farsi

This repository contains the data and code for the paper FarExStance: Explainable Stance Detection for Farsi, which will be presented at the 31st International Conference on Computational Linguistics (COLING 2025).

Introduction

FarExStance introduces the first and largest claim-based explainable stance detection dataset for Farsi. This dataset enables new research in stance detection by providing high-quality annotations and supporting evidence.

We conducted extensive experiments to establish baseline performance using various multilingual open-source and proprietary models, including small and large language models, retrieval-augmented generation (RAG), and parameter-efficient fine-tuning methods. Both automatic and human evaluations were performed to analyze the strengths and limitations of these approaches.

Our dataset, curated with manually labeled instances and supporting evidence, is publicly available to facilitate further research in areas such as explainable NLP and the social media domain.

Data

The dataset is organized in the data/ directory:

  • A list of 130+ Farsi news agency websites, used to collect political, economic, and sports news over six months, is available in news_agency_websites_list.json.
  • Training, development, and test sets for the article2claim and social media domains can be found in the data/b2c/ directory.
  • Training, development, and test sets for the head2claim task are in the data/h2c/ directory.

Reproducing the Experiments

To reproduce the experiments from our paper:

  1. Use the provided *.sh scripts with the appropriate arguments. For example:

    • Running huggingface_exp.sh reproduces the zero-shot and few-shot results of the explainable stance detection task on the test set using models like Command-R-32B and Llama-3.1-70B.
  2. The generated results (predicted stance and explanation) for all instances in the test set for each LLM are available in the corresponding directories under data/.

  3. A detailed explanation of each argument can be found in the rag_inference.py file.

Reference

If you use the tool or any information from this repository or the paper, please cite the paper using the format provided below.

@inproceedings{zarharan-etal-2025-farexstance,
    title = "{F}ar{E}x{S}tance: Explainable Stance Detection for {F}arsi",
    author = "Zarharan, Majid  and
      Hashemi, Maryam  and
      Behroozrazegh, Malika  and
      Eetemadi, Sauleh  and
      Pilehvar, Mohammad Taher  and
      Foster, Jennifer",
    editor = "Rambow, Owen  and
      Wanner, Leo  and
      Apidianaki, Marianna  and
      Al-Khalifa, Hend  and
      Eugenio, Barbara Di  and
      Schockaert, Steven",
    booktitle = "Proceedings of the 31st International Conference on Computational Linguistics",
    month = jan,
    year = "2025",
    address = "Abu Dhabi, UAE",
    publisher = "Association for Computational Linguistics",
    url = "https://aclanthology.org/2025.coling-main.676/",
    pages = "10125--10147",
    abstract = "We introduce FarExStance, a new dataset for explainable stance detection in Farsi. Each instance in this dataset contains a claim, the stance of an article or social media post towards that claim, and an extractive explanation which provides evidence for the stance label. We compare the performance of a fine-tuned multilingual RoBERTa model to several large language models in zero-shot, few-shot, and parameter-efficient fine-tuned settings on our new dataset. On stance detection, the most accurate models are the fine-tuned RoBERTa model, the LLM Aya-23-8B which has been fine-tuned using parameter-efficient fine-tuning, and few-shot Claude-3.5-Sonnet. Regarding the quality of the explanations, our automatic evaluation metrics indicate that few-shot GPT-4o generates the most coherent explanations, while our human evaluation reveals that the best Overall Explanation Score (OES) belongs to few-shot Claude-3.5-Sonnet. The fine-tuned Aya-32-8B model produced explanations most closely aligned with the reference explanations."
}

Releases

No releases published

Packages

No packages published