Pre-Trained Language Models Represent Some Geographic Populations Better Than Others
Dunn, J.; Adams, B.; and Tayyar Madabushi, H. (2023). “Pre-Trained Language Models Represent Some Geographic Populations Better Than Others.” In Proceedings of the Joint International Conference on Computational Linguistics, Language Resources and Evaluation (LREC/COLING 2024). Abstract. This paper measures the skew in how well two families of LLMs represent diverse geographic populations. A spatial probing … More Pre-Trained Language Models Represent Some Geographic Populations Better Than Others