Passer à la navigation principale Passer à la recherche Passer au contenu principal

In-Context Impersonation Reveals Large Language Models' Strengths and Biases

  • Leonard Salewski
  • , Stephan Alaniz
  • , Isabel Rio-Torto
  • , Eric Schulz
  • , Zeynep Akata

Résultats de recherche: Le chapitre dans un livre, un rapport, une anthologie ou une collectionContribution à une conférenceRevue par des pairs

Résumé

In everyday conversations, humans can take on different roles and adapt their vocabulary to their chosen roles. We explore whether LLMs can take on, that is impersonate, different roles when they generate text in-context. We ask LLMs to assume different personas before solving vision and language tasks. We do this by prefixing the prompt with a persona that is associated either with a social identity or domain expertise. In a multi-armed bandit task, we find that LLMs pretending to be children of different ages recover human-like developmental stages of exploration. In a language-based reasoning task, we find that LLMs impersonating domain experts perform better than LLMs impersonating non-domain experts. Finally, we test whether LLMs' impersonations are complementary to visual information when describing different categories. We find that impersonation can improve performance: an LLM prompted to be a bird expert describes birds better than one prompted to be a car expert. However, impersonation can also uncover LLMs' biases: an LLM prompted to be a man describes cars better than one prompted to be a woman. These findings demonstrate that LLMs are capable of taking on diverse roles and that this in-context impersonation can be used to uncover their strengths and hidden biases. Our code is available at https://github.com/ExplainableML/in-context-impersonation.

langue originaleAnglais
titreAdvances in Neural Information Processing Systems 36 - 37th Conference on Neural Information Processing Systems, NeurIPS 2023
rédacteurs en chefA. Oh, T. Neumann, A. Globerson, K. Saenko, M. Hardt, S. Levine
EditeurNeural information processing systems foundation
ISBN (Electronique)9781713899921
étatPublié - 1 janv. 2023
Modification externeOui
Evénement37th Conference on Neural Information Processing Systems, NeurIPS 2023 - New Orleans, États-Unis
Durée: 10 déc. 202316 déc. 2023

Série de publications

NomAdvances in Neural Information Processing Systems
Volume36
ISSN (imprimé)1049-5258

Une conférence

Une conférence37th Conference on Neural Information Processing Systems, NeurIPS 2023
Pays/TerritoireÉtats-Unis
La villeNew Orleans
période10/12/2316/12/23

Empreinte digitale

Examiner les sujets de recherche de « In-Context Impersonation Reveals Large Language Models' Strengths and Biases ». Ensemble, ils forment une empreinte digitale unique.

Contient cette citation