OSoMe Awesome Speakers

OSoMe Awesome Speakers is a virtual event that aims to provide a platform for leading scholars and researchers to share their work and insights about social media manipulation and misinformation. The series will bring together experts from various fields to discuss a range of topics related to information integrity. The series is an excellent opportunity for academics, journalists, and industry professionals to learn about the latest developments in social media research and engage in insightful discussions with experts. (Note: "OSoMe Awesome" is pronounced "awesome awesome.")

Speaker
Event Date
Talk Title
Joshua Tucker

Joshua Tucker

New York University
Professor of Politics, affiliated Professor of Russian and Slavic Studies, and affiliated Professor of Data Science
Wednesday, September 20, 2023
12:00 PM - 1:00 PM EDT

TALK TITLE: Exposure to the Russian Internet Research Agency foreign influence campaign on Twitter in the 2016 US election and its relationship to attitudes and voting behavior

ABSTRACT: There is widespread concern that foreign actors are using social media to interfere in elections worldwide. Yet data have been unavailable to investigate links between exposure to foreign influence campaigns and political behavior. Using longitudinal survey data from US respondents linked to their Twitter feeds, we quantify the relationship between exposure to the Russian foreign influence campaign and attitudes and voting behavior in the 2016 US election. We demonstrate, first, that exposure to Russian disinformation accounts was heavily concentrated: only 1% of users accounted for 70% of exposures. Second, exposure was concentrated among users who strongly identified as Republicans. Third, exposure to the Russian influence campaign was eclipsed by content from domestic news media and politicians. Finally, we find no evidence of a meaningful relationship between exposure to the Russian foreign influence campaign and changes in attitudes, polarization, or voting behavior. The results have implications for understanding the limits of election interference campaigns on social media.

BIO: Joshua A. Tucker is Professor of Politics, affiliated Professor of Russian and Slavic Studies, and affiliated Professor of Data Science at New York University. He is the Director of NYU’s Jordan Center for Advanced Study of Russia and co-Director of the NYU Center for Social Media and Politics; he also served as a co-author/editor of the award-winning politics and policy blog The Monkey Cage at The Washington Post for over a decade. His research focuses on the intersection of social media and politics, including partisan echo chambers, online hate speech, the effects of exposure to social media on political knowledge, online networks and protest, disinformation and fake news, how authoritarian regimes respond to online opposition, and Russian bots and trolls. He is the co-Chair of the external academic team for the U.S. 2020 Facebook & Instagram Election Study, serves on the advisory board of the American National Election Study, the Comparative Study of Electoral Systems, and numerous academic journals, and was the co-founder and co-editor of the Journal of Experimental Political Science. An internationally recognized scholar, he has been a keynote speaker for conferences in Sweden, Denmark, Italy, Brazil, the Netherlands, Russia, and the United States, and has given over 200 invited research presentations at top domestic and international universities and research centers. His most recent books are the co-authored Communism’s Shadow: Historical Legacies and Contemporary Political Attitudes (Princeton University Press, 2017), and the co-edited Social Media and Democracy: The State of the Field (Cambridge University Press, 2020).

Gianluca Stringhini

Gianluca Stringhini

Boston University
Associate Professor, Electrical and Computer Engineering Department
Wednesday, September 27, 2023
12:00 PM - 1:00 PM EDT

TALK TITLE: Computational Methods to Measure and Mitigate Online Disinformation

ABSTRACT: The Web has allowed disinformation to reach an unprecedented scale, allowing it to become ubiquitous and harm society in multiple ways. To be able to fully understand this phenomenon, we need computational tools able to trace false information, monitoring a plethora of online platforms and analyzing not only textual content but also images and videos. In this talk, I will present my group's efforts in developing tools to automatically monitor and model online disinformation. These tools allow us to recommend social media posts that should receive soft moderation, to identify false and misleading images posted online, and to detect inauthentic social network accounts that are likely involved in state-sponsored influence campaigns. I will then discuss our research on understanding the potentially unwanted consequences of suspending misbehaving users on social media.

BIO: Gianluca Stringhini is an Associate Professor in the Electrical and Computer Engineering Department at Boston University, holding affiliate appointments in the Computer Science Department, in the Faculty of Computing and Data Sciences, in the BU Center for Antiracist Research, and in the Center for Emerging Infectious Diseases Policy & Research. In his research Gianluca applies a data-driven approach to better understand malicious activity on the Internet. Through the collection and analysis of large-scale datasets, he develops novel and robust mitigation techniques to make the Internet a safer place. His research involves a mix of quantitative analysis, (some) qualitative analysis, machine learning, crime science, and systems design. Over the years, Gianluca has worked on understanding and mitigating malicious activities like malware, online fraud, influence operations, and coordinated online harassment. He received multiple prizes including an NSF CAREER Award in 2020, and his research won multiple Best Paper Awards. Gianluca has published over 100 peer reviewed papers including several in top computer security conferences like IEEE Security and Privacy, CCS, NDSS, and USENIX Security, as well as top measurement, HCI, and Web conferences such as IMC, ICWSM, CHI, CSCW, and WWW.

Luca Luceri

Luca Luceri

University of Southern California
Research Scientist
Wednesday, October 04, 2023
12:00 PM - 1:00 PM EDT

TALK TITLE: AI-Driven Approaches for Countering Influence Campaigns in Socio-Technical Systems

ABSTRACT: The proliferation of online platforms and social media has sparked a surge in information operations designed to manipulate public opinion on a massive scale, posing significant harm at both the individual and societal levels. In this talk, I will outline a research agenda focused on identifying, investigating, and mitigating orchestrated influence campaigns and deceptive activities within socio-technical systems. I will start by detailing my research efforts in designing AI-based approaches for detecting state-backed troll accounts on social media. Modeling human decision-making as a Markov Decision Process and using an Inverse Reinforcement Learning framework, I will illustrate how we can extract the incentives that social media users respond to and differentiate genuine users from state-sponsored operators. Next, I will delve into a set of innovative approaches I developed to uncover signals of inauthentic, coordinated behaviors. By combining embedding techniques to unveil unexpected similarities in the activity patterns of social media users, along with graph decomposition methods, I will show how we can reveal network structures that pinpoint coordinated groups orchestrating information operations. Through these approaches, I will provide actionable insights to inform regulators in shaping strategies to tame harm, discussing challenges and opportunities to improve the resilience of the information ecosystem, including the potential for interdisciplinary collaborations to address these complex issues.

This is a hybrid event, co-hosted by the Center for Complex Networks and Systems Research. The talk will be held in room 2005 of the Luddy AI building. Please indicate if you will be attending in-person as food will be provided.

BIO: Luca Luceri is a Research Scientist at the Information Sciences Institute (ISI) at the University of Southern California (USC). His research incorporates machine learning, data and network science, with a primary focus on detecting and mitigating online harms in socio-technical systems. He investigates deceptive and malicious behaviors on social media, with a particular emphasis on problems such as social media manipulation, (mis-)information campaigns, and Internet-mediated radicalization processes. His research advances AI/ML/NLP for social good, computational social science, and human-machine interaction. In his role as a Research Scientist at ISI, Luca Luceri serves as a co-PI of the DARPA-funded program INCAS, aiming to develop techniques to detect, characterize, and track geopolitical influence campaigns. Additionally, he is the co-PI of a Swiss NSF-sponsored project called CARISMA, which develops network models to simulate the effects of moderation policies to combat online harms.

Franziska Roesner

Franziska Roesner

University of Washington
Associate Professor, Paul G. Allen School of Computer Science & Engineering
Thursday, October 05, 2023
12:00 PM - 1:00 PM EDT

TALK TITLE: Supporting Effective Peer Response to Misinformation on Social Media

ABSTRACT: Online information is difficult to navigate. While large organizations and social media platforms have taken steps to amplify accurate information and combat rumors and misinformation online, much work to offer correct information and quality resources happens at a more local scale, with individuals and community moderators engaging with (mis)information posted by their peers. In this talk, I will discuss the role of peer-based misinformation response, drawing on results from qualitative and quantitative studies of people’s engagement with potential misinformation online, and sharing our vision and preliminary learnings developing a software tool (the “ARTT Guide”) that supports constructive online conversations and peer-based misinformation response.

BIO: Franziska (Franzi) Roesner is an Associate Professor in the Paul G. Allen School of Computer Science & Engineering at the University of Washington, where she co-directs the Security and Privacy Research Lab. Her research focuses broadly on computer security and privacy for end users of existing and emerging technologies. Her work has studied topics including online tracking and advertising, security and privacy for marginalized and vulnerable user groups, security and privacy in emerging augmented reality (AR) and IoT platforms, and online mis/disinformation. She is the recipient of a Google Security and Privacy Research Award and a Google Research Scholar Award, a Consumer Reports Digital Lab Fellowship, an MIT Technology Review ”Innovators Under 35” Award, an Emerging Leader Alumni Award from the University of Texas at Austin, and an NSF CAREER Award. She serves on the USENIX Security and USENIX Enigma Steering Committees (having previously co-chaired both conferences). More information at https://www.franziroesner.com.

Brendan Nyhan

Brendan Nyhan

Dartmouth College
James O. Freedman Presidential Professor, Department of Government
Wednesday, October 11, 2023
12:00 PM - 1:00 PM EDT

TALK TITLE: Subscriptions and external links help drive resentful users to alternative and extremist YouTube videos

ABSTRACT: Do online platforms facilitate the consumption of potentially harmful content? Using paired behavioral and survey data provided by participants recruited from a representative sample in 2020 (n=1,181), we show that exposure to alternative and extremist channel videos on YouTube is heavily concentrated among a small group of people with high prior levels of gender and racial resentment. These viewers often subscribe to these channels (prompting recommendations to their videos) and follow external links to them. In contrast, non-subscribers rarely see or follow recommendations to videos from these channels. Our findings suggest YouTube's algorithms were not sending people down "rabbit holes" during our observation window in 2020, possibly due to changes that the company made to its recommender system in 2019. However, the platform continues to play a key role in facilitating exposure to content from alternative and extremist channels among dedicated audiences.

BIO: Brendan Nyhan is the James O. Freedman Presidential Professor of Government at Dartmouth College. His research, which focuses on misperceptions about politics and health care, has been published in journals including the American Journal of Political Science, Journal of Politics, Nature Human Behaviour, Proceedings of the National Academy of Sciences, Pediatrics, and Vaccine. Nyhan has been named a Guggenheim Fellow by the Guggenheim Foundation, an Andrew Carnegie Fellow by the Carnegie Corporation of New York, and a Belfer Fellow by the Anti-Defamation League. He is also a co-founder of Bright Line Watch, a non-partisan group monitoring the state of American democracy; a contributor to The Upshot at The New York Times; and co-author of All the President's Spin, a New York Times bestseller that Amazon named one of the best political books of 2004. Nyhan received his Ph.D. from the Department of Political Science at Duke University and previously served as a Robert Wood Johnson Scholar in Health Policy Research and Professor of Public Policy at the Ford School of Public Policy at the University of Michigan.

Ceren Budak

Ceren Budak

University of Michigan
Associate Professor, School of Information
Wednesday, November 08, 2023
12:00 PM - 1:00 PM EDT

TALK TITLE: The Dynamics of Unfollowing Misinformation Spreaders

ABSTRACT: Many studies explore how people "come into" misinformation exposure. But much less is known about how people "come out of" misinformation exposure. Do people organically sever ties to misinformation spreaders? And what predicts doing so? We will discuss this topic in this talk, describing the following work. Over six months, we tracked the frequency and predictors of 1M followers unfollowing health misinformation spreaders on Twitter. We found that misinformation ties are persistent. The fact that users rarely unfollow misinformation spreaders suggests a need for external nudges and the importance of preventing exposure from arising in the first place. Although generally infrequent, the factors most associated with unfollowing misinformation spreaders are redundancy and ideology.

BIO: Ceren Budak is an Associate Professor at the School of Information at the University of Michigan. She utilizes network science, machine learning, and crowdsourcing methods and draws from scientific knowledge across multiple social science communities to contribute computational methods to the field of political communication. Her recent scholarship is focused on topics related to news production and consumption, election campaigns, and online social movements.

Renée DiResta

Renée DiResta

Stanford Internet Observatory
Technical Research Manager
Thursday, November 16, 2023
12:00 PM - 1:00 PM EDT

TALK TITLE: Generative AI and the Challenge of Unreality

ABSTRACT: Generative AI technology has reshaped the landscape of disinformation and influence operations, though the effect of unreality extends far beyond these domains. Just as social media altered the capacity for distribution, generative AI has transformed the capacity for creation. This talk explores how the ability to create virtually anything has significant implications for trust and safety on digital platforms, particularly in the realm of child safety (CSAM). It emphasizes that these changes are not mere hypotheticals but are already shaping our online environment, and discusses the ways that platform trust and safety teams, as well as policymakers, must grapple with these new challenges, as we all adapt to a world in which unreality becomes the norm.

BIO: Renée DiResta is the Technical Research Manager at the Stanford Internet Observatory, a cross-disciplinary program of research, teaching and policy engagement for the study of abuse in current information technologies. Renee’s work examines the spread of narratives across social and media networks, how distinct actor types leverage the information ecosystem to exert influence, and how policy, education, and design responses can be used to mitigate manipulative dynamics. Renée has advised Congress, the State Department, and other academic, civic, and business organizations. At the behest of SSCI, she led outside teams investigating both the Russia-linked Internet Research Agency’s multi-year effort to manipulate American society and elections, and the GRU influence campaign deployed alongside its hack-and-leak operations in the 2016 election. Renée is an Ideas contributor at Wired and The Atlantic, an Emerson Fellow, a 2019 Truman National Security Project fellow, a 2019 Mozilla Fellow in Media, Misinformation, and Trust, a 2017 Presidential Leadership Scholar, and a Council on Foreign Relations term member.

David Lazer

David Lazer

Northeastern University
University Distinguished Professor
Thursday, February 08, 2024
12:00 PM - 1:00 PM EDT

TALK TITLE: The Emergent Structure of the Online Information Ecosystem

ABSTRACT: The first part of this presentation examines the emergent and sometimes paradoxical logic of the internet news ecosystem, in particular: (1) collectively, news diets have become far more concentrated in a small number of outlets; (2) however, individuals have relatively diverse news diets—almost certainly far more diverse than was plausible pre-Internet (as measured by number of unique content producers); (3) the social-algorithmic curation system of the Internet tends to point people to content with their preferences, sometimes in unlikely places. The greater diversity of consumption of news measured by number of unique outlets may not actually result in diversity of content. The second part of the presentation will discuss the development of the National Internet Observatory a large, NSF-supported effort to create a privacy-preserving data collection/data analytic system for the broader research community.

BIO: David Lazer is University Distinguished Professor of Political Science and Computer Sciences, Northeastern University. Prior to coming to Northeastern University, he was on the faculty at the Harvard Kennedy School (1998-2009). In 2019, he was elected a fellow to the National Academy of Public Administration. He has published prominent work on misinformation, democratic deliberation, collective intelligence, computational social science, and algorithmic auditing, across a wide range of prominent journals such as Science, Nature, Proceedings of the National Academy of Science, the American Political Science Review, Organization Science, and the Administrative Science Quarterly. His research has received extensive coverage in the media, including the New York Times, NPR, the Washington Post, the Wall Street Journal, and CBS Evening News. He is a co-leader of the COVID States Project, one of the leading efforts to understand the social and political dimensions of the pandemic in the United States. He has been a PI on more than $30m of grants. Dr. Lazer has served in multiple leadership and editorial positions, including as a board member for the International Society for Computational Social Science (ISCSS), the International Network of Social Network Analysts (INSNA), reviewing editor for Science, associate editor of Social Networks and Network Science, numerous other editorial boards and program committees.

David Broniatowski

David Broniatowski

George Washington University
Associate Professor
Thursday, February 29, 2024
12:00 PM - 1:00 PM EDT

TALK TITLE: The Efficacy of Facebook's Vaccine Misinformation Policies and Architecture During the COVID-19 Pandemic

ABSTRACT: Online misinformation promotes distrust in science, undermines public health, and may drive civil unrest. During the coronavirus disease 2019 pandemic, Facebook—the world’s largest social media company—began to remove vaccine misinformation as a matter of policy. We evaluated the efficacy of these policies using a comparative interrupted time-series design. We found that Facebook removed some antivaccine content, but we did not observe decreases in overall engagement with antivaccine content. Provaccine content was also removed, and antivaccine content became more misinformative, more politically polarized, and more likely to be seen in users’ newsfeeds. We explain these findings as a consequence of Facebook’s system architecture, which provides substantial flexibility to motivated users who wish to disseminate misinformation through multiple channels. Facebook’s architecture may therefore afford antivaccine content producers several means to circumvent the intent of misinformation removal policies.

BIO: David A. Broniatowski, Director of the Decision Making and Systems Architecture Laboratory, conducts research in decision making under risk, group decision making, the design and analysis of complex systems, and behavioral epidemiology. This research program draws upon a wide range of techniques including formal mathematical modeling, experimental design, automated text analysis and natural language processing, social and technical network analysis, and big data. His work on systematic distortions of public opinion about vaccines on social media by state-sponsored trolls has been widely reported in the academic and popular press.

David Rand

David Rand

Massachusetts Institute of Technology
Professor of Management Science, Sloan School
Wednesday, March 06, 2024
12:00 PM - 1:00 PM EDT

TALK TITLE: Reducing misinformation sharing on social media using digital ads

ABSTRACT: Interventions to reduce misinformation sharing have been a major focus in recent years. Developing “content-neutral” interventions that do not require specific fact-checks or warnings related to individual false claims is particularly important in developing scalable solutions. Here, we provide the first evaluations of a content-neutral intervention to reduce misinformation sharing conducted at scale in the field. Specifically, across two on-platform randomized controlled trials, one on Meta’s Facebook (N=33,043,471) and the other on Twitter (N=75,763), we find that simple messages reminding people to think about accuracy —delivered to large numbers of users using digital advertisements— reduce misinformation sharing, with effect sizes on par with what is typically observed in digital advertising experiments. On Facebook, in the hour after receiving an accuracy prompt ad, we found a 2.6% reduction in the probability of being a misinformation sharer among users who had shared misinformation the week prior to the experiment. On Twitter, over more than a week of receiving 3 accuracy prompt ads per day, we similarly found a 3.7% to 6.3% decrease in the probability of sharing low-quality content among active users who shared misinformation pre-treatment. These findings suggest that content-neutral interventions that prompt users to consider accuracy have the potential to complement existing content-specific interventions in reducing the spread of misinformation online.

BIO: David Rand is the Erwin H. Schell Professor and a Professor of Management Science and Brain and Cognitive Sciences at MIT, an affiliate of the MIT Institute for Data, Systems, and Society, and the director of the Human Cooperation Laboratory and the Applied Cooperation Team. Bridging the fields of cognitive science, behavioral economics, and social psychology, David’s research combines behavioral experiments run online and in the field with mathematical and computational models to understand people’s attitudes, beliefs, and choices. His work uses a cognitive science perspective grounded in the tension between more intuitive versus deliberative modes of decision-making. He focuses on illuminating why people believe and share misinformation and “fake news,” understanding political psychology and polarization, and promoting human cooperation. David was named to Wired magazine’s Smart List 2012 of “50 people who will change the world,” chosen as a 2012 Pop!Tech Science Fellow, received the 2015 Arthur Greer Memorial Prize for Outstanding Scholarly Research, was selected as fact-checking researcher of the year in 2017 by the Poyner Institute’s International Fact-Checking Network, and received the 2020 FABBS Early Career Impact Award from the Society for Judgment and Decision Making. Papers he has coauthored have been awarded Best Paper of the Year in Experimental Economics, Social Cognition, and Political Methodology.

Sandra González-Bailón

Sandra González-Bailón

Annenberg School, University of Pennsylvania
Carolyn Marvin Professor of Communication
Wednesday, March 20, 2024
12:00 PM - 1:00 PM EDT

TALK TITLE: Social and algorithmic choices in the transmission of information

ABSTRACT: The circulation of information requires prior exposure: we cannot disseminate what we do not see. In online platforms, information exposure results from a complex interaction between social and algorithmic forms of curation that shapes what people see and what they engage with. In this talk, I will discuss recent research investigating the social and technological factors that determine what information we encounter online. I will also discuss why it is important we adopt a systematic approach to mapping the information environment as a whole -- this approach is needed to capture aggregate characteristics that go unnoticed if we only focus on individuals. Networks, I will argue, offer measurement instruments that can help us map that landscape and identify pockets of problematic content as well as the types of audiences more likely to engage with it. Networks, I will argue, also help us compare (and differentiate) modes of exposure and the different layers that structure the current media environment.

BIO: Sandra González-Bailón is the Carolyn Marvin Professor of Communication at the Annenberg School for Communication, and Director of the Center for Information Networks and Democracy (cind.asc.upenn.edu). She also has a secondary appointment in the Department of Sociology at Penn. Prior to joining Penn, she was a Research Fellow at the Oxford Internet Institute (2008-2013). She completed her doctoral degree in Nuffield College (University of Oxford) and her undergraduate studies at the University of Barcelona. Her research agenda lies at the intersection of computational social science and political communication. Her applied research looks at how online networks shape exposure to information, with implications for how we think about political engagement, mobilization dynamics, information diffusion, and the consumption of news. Her articles have appeared in journals like PNAS, Science, Nature, Political Communication, the Journal of Communication, and Social Networks, among others. She is the author of the book Decoding the Social World (MIT Press, 2017) and co-editor of the Oxford Handbook of Networked Communication (OUP, 2020).

Andrew Guess

Andrew Guess

Princeton University
Assistant Professor of Politics and Public Affairs
Thursday, March 28, 2024
12:00 PM - 1:00 PM EDT

TALK TITLE: Algorithmic recommendations and polarization on YouTube

ABSTRACT: An enormous body of work argues that opaque recommendation algorithms contribute to political polarization by promoting increasingly extreme content. We challenge this dominant view, drawing on three large-scale, multi-wave experiments with a combined 7,851 human users, consistently showing that extremizing algorithmic recommendations has limited effects on opinions. Our experiments employ a custom-built video platform with a naturalistic, YouTube-like interface that presents real videos and recommendations drawn from YouTube. We experimentally manipulate YouTube's actual recommendation algorithm to create ideologically balanced and slanted variations. Our design allows us to intervene in a cyclical feedback loop that has long confounded the study of algorithmic polarization---the complex interplay between algorithmic supply of recommendations and user demand for consumption---to examine the downstream effects of recommendation-consumption cycles on policy attitudes. We use over 125,000 experimentally manipulated recommendations and 26,000 platform interactions to estimate how recommendation algorithms alter users' media consumption decisions and, indirectly, their political attitudes. Our work builds on recent observational studies showing that algorithm-driven "rabbit holes" of recommendations may be less prevalent than previously thought. We provide new experimental evidence casting further doubt on widely circulating theories of algorithmic polarization, showing that even large (short-term) perturbations of real-world recommendation systems that substantially modify consumption patterns have limited causal effects on policy attitudes. Our methodology, which captures and modifies the output of real-world recommendation algorithms, offers a path forward for future investigations of black-box artificial intelligence systems. However, our findings also reveal practical limits to effect sizes that are feasibly detectable in academic experiments.

BIO: Andrew Guess is an assistant professor of politics and public affairs at Princeton University. He uses a combination of experimental methods, large datasets, machine learning, and innovative measurement to study how people choose, process, spread, and respond to information about politics. He is also a founding co-editor of the Journal of Quantitative Description: Digital Media, with Kevin Munger and Eszter Hargittai. You can read their essay introducing the journal’s philosophy and goals here.

Kate Starbird

Kate Starbird

University of Washington
Associate Professor at the Department of Human Centered Design & Engineering
Thursday, April 04, 2024
12:00 PM - 1:00 PM EDT

TALK TITLE: Facts, frames, and (mis)interpretations: Understanding rumors as collective sensemaking

ABSTRACT: Polls show that people tend to agree that online misinformation is a critical problem. However, there is less agreement (even among researchers) about what exactly that problem consists of — or how we should address it. For example, we often see the problem described as an issue of "bad facts.” However, after studying online misinformation for over a decade, my team has concluded that people are more often misled not by faulty facts but by misinterpretations and mischaracterizations. Building from that observation, in this talk I'll present a framework for understanding the spread of misinformation — employed here as an umbrella term for several distinct, related phenomena — through the lens of collective sensemaking. Synthesizing prior theoretical work on rumoring, sensemaking and framing, I'll first describe collective sensemaking as a social process that takes place through interaction between evidence, frames, and interpretations. Next, drawing upon empirical case studies, I'll demonstrate how rumors are a natural byproduct of the sensemaking process, how disinformation campaigns attempt to manipulate that process, and how conspiracy theorizing is a type of patterned or corrupted sensemaking. I'll conclude with reflections about how this perspective could inform more nuanced understandings, new ways to operationalize and study these phenomena, and alternative approaches for addressing them.

BIO: Kate Starbird is an Associate Professor at the Department of Human Centered Design & Engineering (HCDE) at the University of Washington (UW). Dr. Starbird’s research sits at the intersection of human-computer interaction and the field of crisis informatics — i.e. the study of how social media and other information-communication technologies are used during crisis events. Currently, her work focuses on the production and spread of online rumors, misinformation, and disinformation during crises, including natural disasters and political disruptions. In particular, she investigates the participatory nature of online disinformation campaigns, exploring both top-down and bottom-up dynamics. Dr. Starbird received her BS in Computer Science from Stanford (1997) and her PhD in Technology, Media, and Society from the University of Colorado (2012). She is a co-founder and currently serves as director of the UW Center for an Informed Public.

Sinan Aral

Sinan Aral

Massachusetts Institute of Technology
David Austin Professor of Management, IT, Marketing and Data Science
Wednesday, April 10, 2024
12:15 PM - 1:15 PM EDT

TALK TITLE: A Causal Test of the Strength of Weak Ties

ABSTRACT: I'll present the results of multiple large-scale randomized experiments on LinkedIn’s People You May Know algorithm, which recommends new connections to LinkedIn members, to test the extent to which weak ties increased job mobility in the world’s largest professional social network. The experiments randomly varied the prevalence of weak ties in the networks of over 20 million people over a 5-year period, during which 2 billion new ties and 600,000 new jobs were created. The results provided experimental causal evidence supporting the strength of weak ties and suggested three revisions to the theory. First, the strength of weak ties was nonlinear. Statistical analysis found an inverted U-shaped relationship between tie strength and job transmission such that weaker ties increased job transmission but only to a point, after which there were diminishing marginal returns to tie weakness. Second, weak ties measured by interaction intensity and the number of mutual connections displayed varying effects. Moderately weak ties (measured by mutual connections) and the weakest ties (measured by interaction intensity) created the most job mobility. Third, the strength of weak ties varied by industry. Whereas weak ties increased job mobility in more digital industries, strong ties increased job mobility in less digital industries.

BIO: Sinan Aral is the David Austin Professor of Management, IT, Marketing and Data Science at MIT, Director of the MIT Initiative on the Digital Economy (IDE) and a founding partner at Manifest Capital. He is currently on the Advisory Boards of the Alan Turing Institute, the British National Institute for Data Science in London, the Centre for Responsible Media Technology and Innovation in Bergen, Norway and C6 Bank, one of the first all-digital banks of Brazil. Sinan’s research and teaching have won numerous awards including the Microsoft Faculty Fellowship, the PopTech Science Fellowship, an NSF CAREER Award, a Fulbright Scholarship, the Jamieson Award for Teaching Excellence (MIT Sloan’s highest teaching honor) and more than ten best paper awards conferred by his colleagues in research. In 2014, he was named one of the “World’s Top 40 Business School Professors Under 40” and, in 2018, became the youngest ever recipient of the Herbert Simon Award of Rajk László College in Budapest, Hungary. In 2018, Sinan published what Altmetrics called the second most influential scientific publication of the year on the “Spread of False News Online,” on the cover of Science. He is also the author of the upcoming book The Hype Machine, about how social media is disrupting our elections, our economies and our lives.