Sep 3, 2021

On Facebook

When we get busy blaming "the media", let's be clear - and inclusive: Who all we talkin' 'bout here?


My Facebook feed has changed dramatically just over the last year or so. That could be at least partly due to people dropping out of the thing because it's turned into such a fuckin' cess pit, but I think it has plenty to do with the massive increase in the volume of targeted advertising, and the even more annoying attempts to push me into adding friends - people I don't know from Adam's off ox.

Anyway, somebody threw a study into it, and gee, what a whole big buncha surprises they came up with.

WaPo: (pay wall)

Misinformation on Facebook got six times more clicks than factual news during the 2020 election, study says

Right-leaning pages also produce more misinformation, the forthcoming study found.


A new study of user behavior on Facebook around the 2020 election is likely to bolster critics’ long-standing arguments that the company’s algorithms fuel the spread of misinformation over more trustworthy sources.

The forthcoming peer-reviewed study by researchers at New York University and the Université Grenoble Alpes in France has found that from August 2020 to January 2021, news publishers known for putting out misinformation got six times the amount of likes, shares, and interactions on the platform as did trustworthy news sources, such as CNN or the World Health Organization.

Ever since “fake news” on Facebook became a public concern following the 2016 presidential election, publishers who traffic in misinformation have been repeatedly shown to be able to gain major audiences on the platform. But the NYU study is one of the few comprehensive attempts to measure and isolate the misinformation effect across a wide group of publishers on Facebook, experts said, and its conclusions support the criticism that Facebook’s platform rewards publishers that put out misleading accounts.

The study “helps add to the growing body of evidence that, despite a variety of mitigation efforts, misinformation has found a comfortable home — and an engaged audience — on Facebook,” said Rebekah Tromble, director of the Institute for Data, Democracy and Politics at George Washington University, who reviewed the study’s findings.

In response, Facebook said that the report measured the number of people who engage with content, but that is not a measure of the number of people that actually view it (Facebook does not make the latter number, called impressions, publicly available to researchers).

“This report looks mostly at how people engage with content, which should not be confused with how many people actually see it on Facebook,” said Facebook spokesman Joe Osborne. "When you look at the content that gets the most reach across Facebook, it is not at all like what this study suggests.”

He added that the company has 80 fact checking partners covering over 60 languages that work to label and reduce the distribution of false information.

The study’s authors relied on categorizations from two nonprofit organizations that study misinformation, NewsGuard and Media Bias/Fact Check. Both groups have categorized thousands of Facebook publishers by their political leanings, ranging from far left to far right, and by their propensity to share trustworthy or untrustworthy news. The team then took 2,551 of these pages and compared the interactions on posts on pages by publishers known for misinformation, such as the left-leaning Occupy Democrats and the right-leaning Dan Bongino and Breitbart, to interactions on posts from factual publishers.

The researchers also found that the statistically significant misinformation boost is politically neutral — misinformation-trafficking pages on both the far left and the far right generated much more engagement from Facebook users than factual pages of any political slant. But publishers on the right have a much higher propensity to share misleading information than publishers in other political categories, the study found. The latter finding echoes the conclusions of other researchers, as well as Facebook’s own internal findings ahead of the 2018 midterm elections, according to Washington Post reporting.


Occupy Democrats, Bongino and Breitbart did not immediately respond to requests for comment.

Facebook’s critics have long charged that misleading, inflammatory content that often reinforces the viewpoints of its viewers generates significantly more attention and clicks than mainstream news.

That claim — which has been reiterated by members of Congress as well as by Silicon Valley engineers in films such as “The Social Dilemma” — had gained significant traction during the pandemic. Conspiracy theories about covid-19 and vaccines, along with misleading information about treatments and cures, have gone viral, and may have influenced the views of large numbers of Americans. A recent survey by the COVID States Project found that U.S. Facebook users were less likely to be vaccinated any other type of news consumer, even consumers of right-leaning Fox News.

President Biden upped the ante in July when he said covid-related misinformation on platforms such as Facebook was “killing people,” a comment he later walked back.

But there has been little hard data to back up the assertions about the harm caused by Facebook’s algorithms, in part because Facebook has limited the data that researchers can access, Tromble said.

In 2018, an MIT study of misleading stories on Twitter — a platform whose content, unlike Facebook’s, is largely public — found that they performed better among Twitter users than factual stories. Other studies have found that engagement with misinformation is not as widespread as people might think, and that the people who consume and spread misinformation tend to be small numbers of highly motivated partisans.


Facebook is also increasingly restricting access to outside groups that make attempts to mine the company’s data. In the past several months, the White House has repeatedly asked Facebook for information about the extent of covid misinformation on the platform, but the company did not provide it.

One of the researchers Facebook has clamped down on was the NYU researcher, Laura Edelson, who conducted the study. The company cut off Edelson and her colleagues’ accounts last month, arguing that her data collection — which relied on users voluntarily downloading a software widget that allows researchers to track the ads that they see — put Facebook potentially in violation of a 2019 U.S. Federal Trade Commission privacy settlement.

The commission, in a rare rebuttal, shot back that the settlement makes exceptions for researchers and that Facebook should not use it as an excuse to deny the public the ability to understand people’s behavior on social networks.

Edelson noted that because Facebook stopped her project, called the NYU Ad Observatory, last month, she would not be able to continue to study the reach and impact of misinformation on the platform.

In response to criticism that it is becoming less transparent, Facebook recently published a new transparency report that shows the most popular content on the platform every quarter. But the report is highly curated, and Facebook censored an earlier version of the report out of concerns that it would generate bad press, according to a person familiar with the discussions who spoke on the condition of anonymity to describe sensitive conversations. That led critics to argue that the company was not being transparent.

One of the reasons it is hard to tell how much exposure people have to misinformation on Facebook in particular is because so much content is shared in private groups, Tromble said.


To conduct the study, Edelson’s team used a Facebook-owned business analytics tool called CrowdTangle to conduct the analysis. The tool is often used by journalists and researchers to track the popularity of posts. But CrowdTangle has limitations as well: The tool shares how many likes and shares a particular post received, but does not disclose what are known as impressions, or how many people saw the post.

Edelson said the study showed that Facebook algorithms were not rewarding partisanship or bias, or favoring sites on one side of the political spectrum, as some critics have claimed. She said that Facebook amplifies misinformation because it does well with users, and the sites that happen to have more misinformation are on the right. Among publishers categorized as on the far right, those that share misinformation get a majority — or 68 percent — of all engagement from users.

No comments:

Post a Comment