UW News

March 18, 2020

How people investigate — or don’t — fake news on Twitter and Facebook

UW News

The back of a person wearing a hat and a jacket. Drawn around the person's head are four speech bubbles: Each one contains a logo for either Skype, Facebook, Twitter or WhatsApp

UW researchers studied how people investigated potentially suspicious posts on their own Facebook and Twitter feeds.

Social media platforms, such as Facebook and Twitter, provide people with a lot of information, but it’s getting harder and harder to tell what’s real and what’s not.

Researchers at the University of Washington wanted to know how people investigated potentially suspicious posts on their own feeds. The team watched 25 participants scroll through their Facebook or Twitter feeds while, unbeknownst to them, a Google Chrome extension randomly added debunked content on top of some of the real posts. Participants had various reactions to encountering a fake post: Some outright ignored it, some took it at face value, some investigated whether it was true, and some were suspicious of it but then chose to ignore it. These results have been accepted to the 2020 ACM CHI conference on Human Factors in Computing Systems.

“We wanted to understand what people do when they encounter fake news or misinformation in their feeds. Do they notice it? What do they do about it?” said senior author Franziska Roesner, a UW associate professor in the Paul G. Allen School of Computer Science & Engineering. “There are a lot of people who are trying to be good consumers of information and they’re struggling. If we can understand what these people are doing, we might be able to design tools that can help them.”

First row says "when someone sees a post," second row has three options -- they can choose to ignore it, be skeptical or take it face value. If they're skeptical, they can choose to investigate it or not. If they take it at face value, they can choose to like it or share it.

Participants had various reactions to encountering a fake post: Some outright ignored it, some took it at face value, some investigated whether it was true, and some were suspicious of it but then chose to ignore it.Franziska Roesner/University of Washington

Previous research on how people interact with misinformation asked participants to examine content from a researcher-created account, not from someone they chose to follow.

“That might make people automatically suspicious,” said lead author Christine Geeng, a UW doctoral student in the Allen School. “We made sure that all the posts looked like they came from people that our participants followed.”

The researchers recruited participants ages 18 to 74 from across the Seattle area, explaining that the team was interested in seeing how people use social media. Participants used Twitter or Facebook at least once a week and often used the social media platforms on a laptop.

Then the team developed a Chrome extension that would randomly add fake posts or memes that had been debunked by the fact-checking website Snopes.com on top of real posts to make it temporarily appear they were being shared by people on participants’ feeds. So instead of seeing a cousin’s post about a recent vacation, a participant would see their cousin share one of the fake stories instead.

A Facebook post that shows a picture of a crazy cloud formation over Sydney. The text above the picture says "A friend posted this pic of last night's storm in Syndey. Think there might be a craft in there somewhere?"

An example of a fake post that a participant might see on their Facebook feed during the study.Geeng et al./2020 ACM CHI conference on Human Factors in Computing Systems

The researchers either installed the extension on the participant’s laptop or the participant logged into their accounts on the researcher’s laptop, which had the extension enabled. The team told the participants that the extension would modify their feeds — the researchers did not say how — and would track their likes and shares during the study — though, in fact, it wasn’t tracking anything. The extension was removed from participants’ laptops at the end of the study.

A Twitter post that says "New Zealand pulls Fox News off the air, we should have done this years ago." and links to a thinkprogress.org article

An example of a fake post that a participant might see on their Twitter feed during the study.Geeng et al./2020 ACM CHI conference on Human Factors in Computing Systems

“We’d have them scroll through their feeds with the extension active,” Geeng said. “I told them to think aloud about what they were doing or what they would do if they were in a situation without me in the room. So then people would talk about ‘Oh yeah, I would read this article,’ or ‘I would skip this.’ Sometimes I would ask questions like, ‘Why are you skipping this? Why would you like that?'”

Participants could not actually like or share the fake posts. On Twitter, a “retweet” would share the real content beneath the fake post. The one time a participant did retweet content under the fake post, the researchers helped them undo it after the study was over. On Facebook, the like and share buttons didn’t work at all.

A Facebook post with a picture of a dishwasher that has important photos on the bottom shelf and important documents on the top shelf. The text says "FLORIDA PEEPS! GOOD ADVICE! EMPTY YOUR DISHWASHER AND PUT ANYTHING YOU WANT TO PRESERVE IN THERE. It's waterproof and secured to cabinets so it's more likely to survive a storm. TELL YOUR FRIENDS.

An example of a fake post that a participant might see on their Facebook feed during the study. A participant mentioned skipping this post because they saw the word “Florida” and decided it didn’t pertain to them.Geeng et al./2020 ACM CHI conference on Human Factors in Computing Systems

After the participants encountered all the fake posts — nine for Facebook and seven for Twitter — the researchers stopped the study and explained what was going on.

“It wasn’t like we said, ‘Hey, there were some fake posts in there.’ We said, ‘It’s hard to spot misinformation. Here were all the fake posts you just saw. These were fake, and your friends did not really post them,'” Geeng said. “Our goal was not to trick participants or to make them feel exposed. We wanted to normalize the difficulty of determining what’s fake and what’s not.”

The researchers concluded the interview by asking participants to share what types of strategies they use to detect misinformation.

In general, the researchers found that participants ignored many posts, especially those they deemed too long, overly political or not relevant to them.

But certain types of posts made participants skeptical. For example, people noticed when a post didn’t match someone’s usual content. Sometimes participants investigated suspicious posts — by looking at who posted it, evaluating the content’s source or reading the comments below the post — and other times, people just scrolled past them.

“I am interested in the times that people are skeptical but then choose not to investigate. Do they still incorporate it into their worldviews somehow?” Roesner said. “At the time someone might say, ‘That’s an ad. I’m going to ignore it.’ But then later do they remember something about the content, and forget that it was from an ad they skipped? That’s something we’re trying to study more now.”

A Twitter post that says "Last year lettuce killed more Americans than undocumented immigrants so it's a good thing we're halting food inspections over a wall that won't work."

An example of a fake post that a participant might see on their Twitter feed during the study.Geeng et al./2020 ACM CHI conference on Human Factors in Computing Systems

While this study was small, it does provide a framework for how people react to misinformation on social media, the team said. Now researchers can use this as a starting point to seek interventions to help people resist misinformation in their feeds.

“Participants had these strong models of what their feeds and the people in their social network were normally like. They noticed when it was weird. And that surprised me a little,” Roesner said. “It’s easy to say we need to build these social media platforms so that people don’t get confused by fake posts. But I think there are opportunities for designers to incorporate people and their understanding of their own networks to design better social media platforms.”

Savanna Yee, a UW master’s student in the Allen School, is also a co-author on this paper. This research was funded by the National Science Foundation.

For more information, contact Roesner at franzi@cs.washington.edu and Geeng at cgeeng@cs.washington.edu.

Grant number: CNS-1651230

Tag(s):