"These viewers typically subscribe to these channels (causing YouTube to recommend their videos more often) and often follow external links to them. "Using paired behavioral and survey data provided by participants recruited from a representative sample (n=1,181), we show that exposure to alternative and extremist channel videos on YouTube is heavily concentrated among a small group of people with high prior levels of gender and racial resentment," write the researchers.
Robertson, and Northeastern's Christo Wilson complicates these popular narratives. Chen, Dartmouth's Brendan Nyhan, University of Exeter's Jason Reifler, Stanford's Ronald E. Never one to dial back alarmism, The Daily Beast put out a headline in 2018 calling YouTube's algorithm a "far-right radicalization factory" and claimed that an "unofficial network of fringe channels is pulling YouTubers down the rabbit hole of extremism." Even MIT Technology Review sounded the alarm in 2020 about how "YouTube's algorithm seems to be funneling people to alt-right videos."Ī new study by City University of New York's Annie Y. Some travel all the way to neo-Nazism, while others stop at milder forms of bigotry." "Some young men discover far-right videos by accident, while others seek them out. "Over years of reporting on internet culture, I've heard countless versions of story: an aimless young man-usually white, frequently interested in video games-visits YouTube looking for direction or distraction and is seduced by a community of far-right creators," wrote Kevin Roose for The New York Times back in 2019.