Hannah Rose Kirk
Cited by
Cited by
Bias out-of-the-box: An empirical analysis of intersectional occupational biases in popular generative language models
HR Kirk, Y Jun, F Volpin, H Iqbal, E Benussi, F Dreyer, A Shtedritski, ...
Advances in neural information processing systems 34, 2611-2624, 2021
SemEval-2023 Task 10: Explainable Detection of Online Sexism
HR Kirk, W Yin, B Vidgen, P Röttger
arXiv preprint arXiv:2303.04222, 2023
Dataperf: Benchmarks for data-centric ai development
M Mazumder, C Banbury, X Yao, B Karlaš, WG Rojas, S Diamos, ...
arXiv preprint arXiv:2207.10062, 2022
Auditing large language models: a three-layered approach
J Mökander, J Schuett, HR Kirk, L Floridi
AI and Ethics, 1-31, 2023
A prompt array keeps the bias away: Debiasing vision-language models with adversarial learning
H Berg, SM Hall, Y Bhalgat, W Yang, HR Kirk, A Shtedritski, M Bain
Proceedings of the 2nd Conference of the Asia-Pacific Chapter of the …, 2022
Hatemoji: A test suite and adversarially-generated dataset for benchmarking and detecting emoji-based hate
HR Kirk, B Vidgen, P Röttger, T Thrush, SA Hale
Proceedings of the 2022 Conference of the North American Chapter of the …, 2021
Personalisation within bounds: A risk taxonomy and policy framework for the alignment of large language models with personalised feedback
HR Kirk, B Vidgen, P Röttger, SA Hale
arXiv preprint arXiv:2303.05453, 2023
Memes in the Wild: Assessing the Generalizability of the Hateful Memes Challenge Dataset
HR Kirk, Y Jun, P Rauba, G Wachtel, R Li, X Bai, N Broestl, M Doff-Sotta, ...
Proceedings of the 5th Workshop on Online Abuse and Harms (WOAH 2021), 2021
Looking for a Handsome Carpenter! Debiasing GPT-3 Job Advertisements
C Borchers, DS Gala, B Gilburt, E Oravkin, W Bounsi, YM Asano, HR Kirk
Proceedings of the 4th workshop on gender bias in natural language …, 2022
Handling and Presenting Harmful Text in NLP
HR Kirk, A Birhane, B Vidgen, L Derczynski
EMNLP Findings, 2022
The nuances of Confucianism in technology policy: An inquiry into the interaction between cultural and political systems in Chinese digital ethics
HR Kirk, K Lee, C Micallef
International Journal of Politics, Culture, and Society, 1-24, 2020
Assessing language model deployment with risk cards
L Derczynski, HR Kirk, V Balachandran, S Kumar, Y Tsvetkov, MR Leiser, ...
arXiv preprint arXiv:2303.18190, 2023
Is More Data Better? Re-thinking the Importance of Efficiency in Abusive Language Detection with Transformers-Based Active Learning
HR Kirk, B Vidgen, SA Hale
Proceedings of the Third Workshop on Threat, Aggression and Cyberbullying …, 2022
Xstest: A test suite for identifying exaggerated safety behaviours in large language models
P Röttger, HR Kirk, B Vidgen, G Attanasio, F Bianchi, D Hovy
arXiv preprint arXiv:2308.01263, 2023
Balancing the Picture: Debiasing Vision-Language Datasets with Synthetic Contrast Sets
B Smith, M Farinha, SM Hall, HR Kirk, A Shtedritski, M Bain
arXiv preprint arXiv:2305.15407, 2023
VisoGender: A dataset for benchmarking gender bias in image-text pronoun resolution
SM Hall, FG Abrantes, H Zhu, G Sodunke, A Shtedritski, HR Kirk
arXiv preprint arXiv:2306.12424, 2023
Adversarial Nibbler: A Data-Centric Challenge for Improving the Safety of Text-to-Image Models
A Parrish, HR Kirk, J Quaye, C Rastogi, M Bartolo, O Inel, J Ciro, ...
arXiv preprint arXiv:2305.14384, 2023
The mediation of matchmaking: a comparative study of gender and generational preference in online dating websites and offline blind date markets in Chengdu
HR Kirk, S Gupta
The Journal of Chinese Sociology 9 (1), 2, 2022
Casteist but Not Racist? Quantifying Disparities in Large Language Model Bias between India and the West
K Khandelwal, M Tonneau, AM Bean, HR Kirk, SA Hale
arXiv preprint arXiv:2309.08573, 2023
DoDo Learning: DOmain-DemOgraphic Transfer in Language Models for Detecting Abuse Targeted at Public Figures
HR Kirk, AR Williams, L Burke, YL Chung, I Debono, P Johansson, ...
arXiv preprint arXiv:2307.16811, 2023
The system can't perform the operation now. Try again later.
Articles 1–20