Society often relies on social algorithms that adapt to human behavior. Yet scientists struggle to generalize the combined behavior of mutually-adapting humans and algorithms. This scientific challenge is a governance problem when algorithms amplify human responses to falsehoods. Could attempts to influence humans have second-order effects on algorithms? Using a large-scale field experiment, I test if influencing readers to fact-check unreliable sources causes news aggregation algorithms to promote or lessen the visibility of those sources. Interventions encouraged readers to fact-check articles or fact-check and provide votes to the algorithm. Across 1104 discussions, these encouragements increased human fact-checking and reduced vote scores on average. The fact-checking condition also caused the algorithm to reduce the promotion of articles over time by as much as -25 rank positions on average, enough to remove an article from the front page. Overall, this study offers a path for the science of human-algorithm behavior by experimentally demonstrating how influencing collective human behavior can also influence algorithm behavior.
Download full-text PDF |
Source |
---|---|
http://www.ncbi.nlm.nih.gov/pmc/articles/PMC10359256 | PMC |
http://dx.doi.org/10.1038/s41598-023-38277-5 | DOI Listing |
PLoS One
October 2024
Molecular Ecology Group, Water Research Institute, National Research Council of Italy, Verbania Pallanza, Italy.
Fear of spiders is a widespread condition often disproportionate to the actual danger spiders pose to humans. Likely rooted in evolutionary history, fear of spiders might also have a cultural component. Recent studies have shown that a significant fraction of spider-related media reports are misleading and sensationalistic.
View Article and Find Full Text PDFNat Hum Behav
December 2024
Department of Psychology, University of California, Berkeley, Berkeley, CA, USA.
J Am Med Dir Assoc
October 2024
Department of Biomedical Informatics, Vanderbilt University Medical Center, Nashville, TN, USA; Department of Computer Science, Vanderbilt University, Nashville, TN, USA.
Introduction: There are many myths regarding Alzheimer's disease (AD) that have been circulated on the internet, each exhibiting varying degrees of accuracy, inaccuracy, and misinformation. Large language models, such as ChatGPT, may be a valuable tool to help assess these myths for veracity and inaccuracy; however, they can induce misinformation as well.
Objective: This study assesses ChatGPT's ability to identify and address AD myths with reliable information.
PNAS Nexus
July 2024
JLU Giessen, 35394 Giessen, Germany.
Community-based fact-checking is a promising approach to fact-check social media content at scale. However, an understanding of whether users trust community fact-checks is missing. Here, we presented Americans with 36 misleading and nonmisleading social media posts and assessed their trust in different types of fact-checking interventions.
View Article and Find Full Text PDFSurg Today
October 2024
Division of Surgical Oncology, Department of Surgery, Nagoya University Graduate School of Medicine, 65 Tsurumai-Cho, Showa-Ku, Nagoya, 466-8550, Japan.
Purposes: We performed a conversation analysis of the speech conducted among the surgical team during three-dimensional (3D)-printed liver model navigation for thrice or more repeated hepatectomy (TMRH).
Methods: Seventeen patients underwent 3D-printed liver navigation surgery for TMRH. After transcription of the utterances recorded during surgery, the transcribed utterances were coded by the utterer, utterance object, utterance content, sensor, and surgical process during conversation.
Enter search terms and have AI summaries delivered each week - change queries or unsubscribe any time!