• Ideas

How Search Engines Amplify Hate — in Parkland and Beyond

3 minute read
Ideas
Dr. Safiya U. Noble is the author of Algorithms of Oppression: How Search Engines Reinforce Racism and is an assistant professor of communication at the University of Southern California, Annenberg School of Communication & Journalism. She is a partner in Stratelligence and co-founder of the Information Ethics & Equity Institute.

Tech companies have been slow to respond to the way their platforms have been used to amplify hate. Anonymity on social media platforms often makes it difficult to identify the right-wing radicalization that is happening to some Americans online, exposing users to violent and often racist disinformation.

We need new business practices and policies that address public harm propagated in media-technology platforms, particularly as bad actors use these platforms to enact violence on others. Important developments are under way in terms of commercial content moderation, which allow humans to flag threats and other forms of dangerous content, but have yet to reach the level of impact needed, given the volume of media that traffics through these platforms. Algorithms and automated decision-making technologies are not yet sophisticated enough to recognize certain types of online threats before mass violence occurs.

We know that anti-government, anti-Black, anti-Muslim, anti-gay and anti-immigrant hate crime is a massive presence, and it’s important to note how white nationalist trolls attempted to take credit for Nikolas Cruz in the immediate aftermath of the mass shooting he carried out at Marjory Stoneman Douglas High School in Parkland, Florida, on Feb. 14, as part of their desire to amplify and enact hate.

It’s easy to see why they would use this opportunity to manipulate the media: In 2009, the United States Department of Homeland Security reported that right-wing extremists used the election of the first African-American President, Barack Obama, to increase recruitment. Racist propaganda and disinformation litters our online media landscape, and, sometimes, incites action. We know people use online platforms to find information and make sense of their social experiences of marginalization, and members of a sector of society feel their identities as white Americans are threatened or under attack, despite evidence to the contrary. This was the case for Dylann Roof, whom we know to have murdered nine African Americans at the Emanuel African Methodist Episcopal Church in Charleston, S.C., in 2015. Roof said he developed the views that fueled his mass shooting online, where he searched for information about violence perpetrated by black Americans on whites.

It’s time we think about making our online platforms more transparent, rather than expecting consumers of social media and search engines to easily distinguish propaganda from fact. Rather than thinking of these media-tech companies as neutral, objective news and information companies, we could think of them for what they are – advertising engines to help companies and organizations know more about us and better target their products and services. We need greater transparency in how these systems work, more collaborations between the tech sector and their critics, and public policy to protect us all from harm.

More Must-Reads from TIME

Contact us at letters@time.com

TIME Ideas hosts the world's leading voices, providing commentary on events in news, society, and culture. We welcome outside contributions. Opinions expressed do not necessarily reflect the views of TIME editors.