Survey data underlying the MSc thesis: "Generative AI: Investigating Consistency and Neutrality in Multilingual Outputs"

DOI:10.4121/e058cc9d-7ca8-408f-9233-79ea0bd3953f.v1
The DOI displayed above is for this specific version of this dataset, which is currently the latest. Newer versions may be published in the future. For a link that will always point to the latest version, please use
DOI: 10.4121/e058cc9d-7ca8-408f-9233-79ea0bd3953f

Datacite citation style

Ibrahim, Ahmed (2025): Survey data underlying the MSc thesis: "Generative AI: Investigating Consistency and Neutrality in Multilingual Outputs". Version 1. 4TU.ResearchData. dataset. https://doi.org/10.4121/e058cc9d-7ca8-408f-9233-79ea0bd3953f.v1
Other citation styles (APA, Harvard, MLA, Vancouver, Chicago, IEEE) available at Datacite

Dataset

This dataset contains responses from an online survey designed to evaluate how consistently and neutrally ChatGPT’s English and Arabic answers align across ten prompts (seven politically sensitive, three non-sensitive). Each row captures one participant’s ratings of sentiment and factual consistency between the two language outputs, neutrality scores for each response and the prompt itself, and optional comments. The data were collected via Qualtrics from English- and Arabic-fluent respondents who compared side-by-side model answers, providing quantitative Likert-scale ratings to assess multilingual consistency and neutrality of Generative AI output in a human evaluation study.

History

  • 2025-05-20 first online, published, posted

Publisher

4TU.ResearchData

Format

.xlsx

Organizations

TU Delft, Faculty of Electrical Engineering, Mathematics and Computer Science

DATA

Files (1)

  • 31,270 bytesMD5:9b6931ab84c23746525bcba7dfc44884Data.xlsx