Author
Listed:
- Pat Pataranutaporn
(a Media Lab, Massachusetts Institute of Technology , Cambridge , MA 02139-4307)
- Nattavudh Powdthavee
(b Division of Economics, Nanyang Technological University , Singapore 639818 , Singapore)
- Chayapatr Archiwaranguprok
(a Media Lab, Massachusetts Institute of Technology , Cambridge , MA 02139-4307)
- Pattie Maes
(a Media Lab, Massachusetts Institute of Technology , Cambridge , MA 02139-4307)
Abstract
Subjective well-being is central to economic, medical, and policy decision-making. We evaluate whether large language models (LLMs) can provide valid predictions of well-being across global populations. Using natural-language profiles from 64,000 individuals in 64 countries, we benchmark four leading LLMs against self-reports and statistical models. Unlike regressions, which estimate relationships from survey data, LLMs draw only on individual characteristics (e.g., sociodemographic, attitudinal, and psychological factors) together with associations encoded during pretraining, rather than from the survey’s subjective well-being responses. They produced plausible patterns consistent with known correlates such as income and health, but systematically underperformed relative to regressions and showed the largest errors in underrepresented countries, reflecting biases rooted in global digital and economic inequality. A preregistered experiment revealed that LLMs rely on surface-level linguistic associations rather than conceptual understanding, leading to predictable distortions in unfamiliar contexts. Injecting contextual information partly reduced—but did not remove—these biases. These findings demonstrate that while LLMs can simulate broad correlates of life satisfaction, they fail to capture its experiential and cultural depth. Accordingly, they should not be used as substitutes for human self-reports of well-being; doing so would risk reinforcing inequality and undermining human agency.
Suggested Citation
Pat Pataranutaporn & Nattavudh Powdthavee & Chayapatr Archiwaranguprok & Pattie Maes, 2025.
"Simulating human well-being with large language models: Systematic validation and misestimation across 64,000 individuals from 64 countries,"
Proceedings of the National Academy of Sciences, Proceedings of the National Academy of Sciences, vol. 122(48), pages 2519394122-, December.
Handle:
RePEc:nas:journl:v:122:y:2025:p:e2519394122
DOI: 10.1073/pnas.2519394122
Download full text from publisher
Corrections
All material on this site has been provided by the respective publishers and authors. You can help correct errors and omissions. When requesting a correction, please mention this item's handle: RePEc:nas:journl:v:122:y:2025:p:e2519394122. See general information about how to correct material in RePEc.
If you have authored this item and are not yet registered with RePEc, we encourage you to do it here. This allows to link your profile to this item. It also allows you to accept potential citations to this item that we are uncertain about.
We have no bibliographic references for this item. You can help adding them by using this form .
If you know of missing items citing this one, you can help us creating those links by adding the relevant references in the same way as above, for each refering item. If you are a registered author of this item, you may also want to check the "citations" tab in your RePEc Author Service profile, as there may be some citations waiting for confirmation.
For technical questions regarding this item, or to correct its authors, title, abstract, bibliographic or download information, contact: PNAS Product Team (email available below). General contact details of provider: http://www.pnas.org/ .
Please note that corrections may take a couple of weeks to filter through
the various RePEc services.