Noyce Conference Room
Colloquium

All day

Speaker: 
Mahzarin Banaji

Our campus is closed to the public for this event.

Mahzarin Banaji Richard Clarke Cabot Professor of Social Ethics Department of Psychology, Harvard University Abstract: Large Language Models (LLMs) show emergent patterns that mimic human cognition. We explored whether they also mirror other, less deliberative human psychological processes. Drawing upon classical theories of cognitive consistency, two preregistered studies tested whether GPT-4o changed its attitudes toward Vladimir Putin in the direction of a positive or negative essay it wrote about the Russian leader. Indeed, GPT displayed patterns of attitude change mimicking cognitive dissonance effects in humans. Even more remarkably, the degree of change increased sharply when the LLM was offered an illusion of choice about which essay (positive or negative) to write, suggesting that GPT-4o manifests a functional analog of humanlike selfhood. The exact mechanisms by which the model mimics human attitude change and self-referential processing remain to be understood. This paper is in press at PNAS. I would especially like to discuss the broader implications of the sensitivity to free choice result demonstrated by GPT with SFI colleagues.

Speaker

Mahzarin BanajiMahzarin BanajiProfessor of Social Ethics at Harvard University & External Professor at SFI
SFI Host: 
Andrew Stier

More SFI Events