Behnam Karami
drbehnamkarami.bsky.social
Behnam Karami
@drbehnamkarami.bsky.social
Postdoc researcher in Cognitive Science
🚨 New Preprint Out!
“Emergent Moral Representations in Large Language Models Align with Human Conceptual, Neural, and Behavioral Moral Structure”

www.researchsquare.com/article/rs-8...

Do LLMs internally represent morality like humans?
Our results point to a striking yes!
Key findings:
Emergent Moral Representations in Large Language Models Aligns with Human Conceptual, Neural, and Behavioral Moral Structure
Large language models (LLMs) increasingly operate in ethically sensitive settings, yet it remains unclear whether they internally encode structured representations of morality. Here we examine the act...
www.researchsquare.com
December 7, 2025 at 11:24 AM
Reposted by Behnam Karami
Imagine an apple 🍎. Is your mental image more like a picture or more like a thought? In a new preprint led by Morgan McCarty—our lab's wonderful RA—we develop a new approach to this old cognitive science question and find that LLMs excel at tasks thought to be solvable only via visual imagery. 🧵
Artificial Phantasia: Evidence for Propositional Reasoning-Based Mental Imagery in Large Language Models
This study offers a novel approach for benchmarking complex cognitive behavior in artificial systems. Almost universally, Large Language Models (LLMs) perform best on tasks which may be included in th...
arxiv.org
October 1, 2025 at 1:27 AM
June 7, 2025 at 1:03 PM
Reposted by Behnam Karami
@suryaganguli.bsky.social gives great talk at LLM workshop at Berkley

-LLM<->brain is still a new topic, less progress so far than LLM<->vision
-lets train LLM foundation models of specific brain systems and then reverse engineer them
-emerging paradigm: read-write experiments in brains+machines
February 3, 2025 at 7:49 PM