Journal of Online Trust and Safety https://tsjournal.org/index.php/jots <p>The Journal of Online Trust and Safety is a cross-disciplinary, open access, fast peer-review journal that publishes research on how consumer internet services are abused to cause harm and how to prevent those harms. </p> ojs/pkp en-US Journal of Online Trust and Safety 2770-3142 Securing Federated Platforms https://tsjournal.org/index.php/jots/article/view/171 <p class="p2">As the social media landscape undergoes broad transformation for the first time in over a decade, with alternative platforms like Mastodon, Bluesky, and Threads emerging where X has receded, many users and observers have celebrated the promise of these new services and their visions of alternative governance structures that empower consumers. Drawing on a large-scale textual analysis of platform moderation policies, capabilities, and transparency mechanisms, as well as semistructured group interviews with developers, administrators, and moderators of federated platforms, we found that federated platforms face considerable obstacles to robust and scalable governance, particularly with regard to persistent threats such as coordinated behavior and spam. Key barriers identified include underdeveloped moderation technologies and a lack of sustainable financial models for trust and safety work. We offer four solutions to the collective safety and security risks identified: (1) institutionalize shared responses to critical harms, (2) build transparent governance into the system, (3) invest in open-source tooling, and (4) enable data sharing across instances.</p> Yoel Roth Samantha Lai Copyright (c) 2024 Journal of Online Trust and Safety https://creativecommons.org/licenses/by-nc-sa/4.0 2024-02-28 2024-02-28 2 2 10.54501/jots.v2i2.171 Content Modeling in Multi-Platform Multilingual Social Media Data https://tsjournal.org/index.php/jots/article/view/136 <p class="p2">An increase in the use of social media as the primary news source for the general population has created an ecosystem in which organic conversation commingles with inorganically seeded and amplified narratives, which can include public relations and marketing activity but also covert and malign influence operations. An efficient and easily understandable analysis of such data is important, as it allows relevant stakeholders to protect online communities and free discussion while better identifying activity and content that may violate social media platform terms of service. To accomplish this, we propose a method of large-scale social media data analysis, which allows for multilingual conversations to be analyzed in depth across any number of social media platforms simultaneously. Our method uses a text embedding model, i.e., a natural language processing model that holds semantic and contextual understandings of language. The model uses an “understanding” of language to represent posts as coordinates in a high-dimensional space, such that posts with similar meanings are assigned coordinates close together. We then cluster and analyze the posts to identify online topics of conversation existing across multiple social media platforms. We explicitly show how our method can be applied to four different datasets, three consisting of Chinese social media posts related to the Belt and Road Initiative and one relating to the Russia-Ukraine war, and we find politically-influenced conversations that contain misleading information relating to the Chinese government and the Russia-Ukraine war.</p> Arman Setser Libby Lange Kyle Weiss Vlad Barash Copyright (c) 2024 Journal of Online Trust and Safety https://creativecommons.org/licenses/by-nc-sa/4.0 2024-02-28 2024-02-28 2 2 10.54501/jots.v2i2.136 Fingerprints of Conspiracy Theories: Identifying Signature Information Sources of a Misleading Narrative and Their Roles in Shaping Message Content and Dissemination https://tsjournal.org/index.php/jots/article/view/152 <p class="p2">This study investigates the role of information sources in the propagation and reception of misleading narratives on social media, focusing on the case of the Chemtrail conspiracy theory—a false claim that the trails in the sky behind airplanes are chemicals deliberately spread for sinister reasons. We collected data from Facebook Pages and Groups discussing the conspiracy theory. We specifically focused on identifying and analyzing “signature” information sources, which are repeatedly used by online communities engaged in the discussion of a misleading narrative but are not widely used by other communities. The findings indicate that messages referencing signature sources contain more death-, illness-, risk-, and health-related words, convey more negativity, and elicit more negative reactions from users, compared with those without signature sources. The insights from this study could contribute to the development of effective strategies to monitor and counter the spread of misleading narratives in digital spaces.</p> Soojong Kim Kwanho Kim Haoning Xue Copyright (c) 2024 Journal of Online Trust and Safety https://creativecommons.org/licenses/by-nc-sa/4.0 2024-02-28 2024-02-28 2 2 10.54501/jots.v2i2.152 Effects of Browsing Conditions and Visual Alert Design on Human Susceptibility to Deepfakes https://tsjournal.org/index.php/jots/article/view/144 <p class="p2">The increasing reach of deepfakes raises practical questions about people’s ability to detect false videos online. How vulnerable are people to deepfake videos? What technologies can help improve detection? Previous experiments that measure human deepfake detection historically omit a number of conditions that can exist in typical browsing conditions. Here, we operationalized four such conditions (low prevalence, brief presentation, low video quality, and divided attention), and found in a series of online experiments that all conditions lowered detection relative to baseline, suggesting that the current literature underestimates people’s susceptibility to deepfakes. Next, we examined how AI assistance could be integrated into the human decision process. We found that a model that exposes deepfakes by amplifying artifacts increases detection rates, and also leads to higher rates of incorporating AI feedback and higher final confidence than text-based prompts. Overall, this suggests that visual indicators that cause distortions on fake videos may be effective at mitigating the impact of falsified video.</p> Emilie Josephs Camilo Fosco Aude Oliva Copyright (c) 2024 Journal of Online Trust and Safety https://creativecommons.org/licenses/by-nc-sa/4.0 2024-02-28 2024-02-28 2 2 10.54501/jots.v2i2.144 Burden of Proof: Lessons Learned for Regulators from the Oversight Board’s Implementation Work https://tsjournal.org/index.php/jots/article/view/168 Naomi Shiffman Carly Miller Manuel Parra Yagnam Claudia Flores-Saviaga Copyright (c) 2024 Journal of Online Trust and Safety https://creativecommons.org/licenses/by-nc-sa/4.0 2024-02-28 2024-02-28 2 2 10.54501/jots.v2i2.168 Assuming Good Faith Online https://tsjournal.org/index.php/jots/article/view/169 Eric Goldman Copyright (c) 2024 Journal of Online Trust and Safety https://creativecommons.org/licenses/by-nc-sa/4.0 2024-02-28 2024-02-28 2 2 10.54501/jots.v2i2.169 Should Politicians Be Exempt from Fact-Checking? https://tsjournal.org/index.php/jots/article/view/170 Sarah Fisher Beatriz Kira Kiran Arabaghatta Basavaraj Jeffrey Howard Copyright (c) 2024 Journal of Online Trust and Safety https://creativecommons.org/licenses/by-nc-sa/4.0 2024-02-28 2024-02-28 2 2 10.54501/jots.v2i2.170 Bridging Theory & Practice: Examining the State of Scholarship Using the History of Trust and Safety Archive https://tsjournal.org/index.php/jots/article/view/173 Megan Knittel Amanda Menking Copyright (c) 2024 Journal of Online Trust and Safety https://creativecommons.org/licenses/by-nc-sa/4.0 2024-02-28 2024-02-28 2 2 10.54501/jots.v2i2.173