Researchers have created and tested a prototype browser extension called Viblio that lets viewers and creators add Wikipedia-like citations to YouTube videos.
While Google has long been synonymous with search, people are increasingly seeking information directly through video platforms such as YouTube. Videos can be dense with information: text, audio, and image after image.
Yet each of these layers presents a potential source of error or deceit. And when people search for videos directly on a site like YouTube, sussing out which videos are credible sources can be tricky.
The new browser extension prototype could help people vet videos with Wikipedia-like citations.
The prototype offers users an alternate timeline, studded with notes and links to sources that support, refute, or expand on the information presented in the video. Those links also appear in a list view, like the “References” section at the end of Wikipedia articles.
In tests, 12 participants found the tool useful for gauging the credibility of videos on topics ranging from biology to political news to COVID-19 vaccines.
“We wanted to come up with a method to encourage people watching videos to do what’s called ‘lateral reading,’ which is that you go look at other places on the web to establish whether something is credible or true, as opposed to diving deep into the thing itself,” says senior author Amy X. Zhang, an assistant professor in the Paul G. Allen School of Computer Science & Engineering at the University of Washington.
“In previous research, I’d worked with the people at X’s Community Notes and with Wikipedia and seen that crowdsourcing citations and judgments can be a useful way to call out misinformation on platforms.”
To inform Viblio’s design, the team studied how 12 participants—mostly college students under 30—gauged the credibility of YouTube videos when searching for them on the platform and while watching them. All said familiarity with the video’s source and the name of the channel were important. But many cited signs of a video’s potentially faulty credibility: the quality of the video, the user’s degree of interest in it, its ranking in search results, its length, and the number of views or subscribers.
The team also found that in one case a participant misinterpreted a YouTube information panel as an endorsement of the video from the Centers for Disease Control and Prevention. But these panels are actually links to supplemental information that the site attaches to videos on “topics prone to misinformation.”
“The trouble is that a lot of YouTube videos, especially more educational ones, don’t offer a great way for people to prove they’re presenting good information,” says Emelia Hughes, a doctoral student at the University of Notre Dame who completed this research as a UW undergraduate student in the Information School.
“I’ve stumbled across a couple of YouTubers who were coming up with their own ways to cite sources within videos. There’s also not a great way to fight bad information. People can report a whole video, but that’s a pretty extreme measure when someone makes one or two mistakes.”
The researchers designed Viblio so users can better understand videos’ content while also avoiding things like users misinterpreting the additional information. To add a citation, users click a button on the extension. They can then add a link, select the timespan their citation references, and add optional comments. They can also select the type of citation, which marks it with a colored dot in the timeline: “refutes the video clip’s claim” (red), “supports the video clip’s claim” (green), or “provides further explanation” (blue).
To test the system, the team had the study participants use Viblio for two weeks on a range of videos, including clips from Good Morning America, Fox News, and ASAPScience. Participants could add citations as well as watch videos with other participants’ citations. For many, the added citations changed their opinion of certain videos’ credibility. But the participants also highlighted potential difficulties with deploying Viblio at a larger scale, such as the conflicts that arise in highly political videos or those on controversial topics that don’t fall into true-false binaries.
“What happens when people with different value systems add conflicting citations?” says coauthor Tanu Mitra, an assistant professor in the Information School.
“We of course have the issue with bad actors potentially adding misinformation and incorrect citations, but even when the users are acting in good faith, but have conflicting options, whose citation should be prioritized? Or should we be showing both conflicting citations? These are big challenges at scale.”
The researchers highlight a few areas for further study, such as expanding Viblio to other video platforms such as TikTok or Instagram; studying its usability at a greater scale to see whether users are motivated enough to continue adding citations; and exploring ways to create citations for videos that don’t get as much traffic and thus have fewer citations.
“Once we get past this initial question of how to add citations to videos, then the community vetting question remains very challenging,” Zhang says. “It can work. At X, Community Notes is working on ways to prevent people from ‘gaming’ voting by looking at whether someone always takes the same political side. And Wikipedia has standards for what should be considered a good citation. So it’s possible. It just takes resources.”
The team will present its findings May 14 in Honolulu at the ACM CHI Conference on Human Factors in Computing Systems.
Viblio is not available to the public.
This research was funded by the WikiCred Grants Initiative.
Source: University of Washington