Unknown

Dataset Information

0

Scaling up psychology via Scientific Regret Minimization.


ABSTRACT: Do large datasets provide value to psychologists? Without a systematic methodology for working with such datasets, there is a valid concern that analyses will produce noise artifacts rather than true effects. In this paper, we offer a way to enable researchers to systematically build models and identify novel phenomena in large datasets. One traditional approach is to analyze the residuals of models-the biggest errors they make in predicting the data-to discover what might be missing from those models. However, once a dataset is sufficiently large, machine learning algorithms approximate the true underlying function better than the data, suggesting, instead, that the predictions of these data-driven models should be used to guide model building. We call this approach "Scientific Regret Minimization" (SRM), as it focuses on minimizing errors for cases that we know should have been predictable. We apply this exploratory method on a subset of the Moral Machine dataset, a public collection of roughly 40 million moral decisions. Using SRM, we find that incorporating a set of deontological principles that capture dimensions along which groups of agents can vary (e.g., sex and age) improves a computational model of human moral judgment. Furthermore, we are able to identify and independently validate three interesting moral phenomena: criminal dehumanization, age of responsibility, and asymmetric notions of responsibility.

SUBMITTER: Agrawal M 

PROVIDER: S-EPMC7183163 | biostudies-literature | 2020 Apr

REPOSITORIES: biostudies-literature

altmetric image

Publications

Scaling up psychology via Scientific Regret Minimization.

Agrawal Mayank M   Peterson Joshua C JC   Griffiths Thomas L TL  

Proceedings of the National Academy of Sciences of the United States of America 20200402 16


Do large datasets provide value to psychologists? Without a systematic methodology for working with such datasets, there is a valid concern that analyses will produce noise artifacts rather than true effects. In this paper, we offer a way to enable researchers to systematically build models and identify novel phenomena in large datasets. One traditional approach is to analyze the residuals of models-the biggest errors they make in predicting the data-to discover what might be missing from those  ...[more]

Similar Datasets

| S-EPMC4835314 | biostudies-literature
| S-EPMC7494847 | biostudies-literature
| S-EPMC2518954 | biostudies-literature
| S-EPMC4212974 | biostudies-literature
| S-EPMC4026152 | biostudies-literature
| S-EPMC7117690 | biostudies-literature
| S-EPMC5965697 | biostudies-literature
| S-EPMC8252642 | biostudies-literature
2024-03-26 | GSE262455 | GEO