Newsletter
Join the Community
Subscribe to our newsletter for the latest news and updates
A study evaluating geopolitical and cultural biases in large language models through dual-layered assessments.
This repository contains the dataset, evaluation scripts, and results for analyzing geopolitical and cultural biases in large language models (LLMs). The study is structured into two evaluation phases: factual QA (objective knowledge) and disputable QA (politically sensitive disputes). We explore how LLMs exhibit model bias (training-induced) and inference bias (query language-induced) when answering questions in different languages.