New York: Ever wondered how Facebook determines what you see in your news feed every time you log in? The social networking site has actually devised different algorithms for it.
Apart from things like where you live and the pages you follow, it primarily looks for two broader signals — Topics that are being mentioned a lot and topics that suddenly seen an increase in mentions, Re/Code reported.
For example, singer Justin Bieber is mentioned often on Facebook. So the total volume of mentions is always high and is not a good indicator of whether or not he is part of a trending topic.
So, Facebook would look for a hike in mentions relative to the normal prattle around Bieber.
“This means that things that trend are not just the most highly mentioned people or topics, they have to be tied to some kind of relevant event,” the report said.
Once a topic is identified as trending, it is approved by a human controller, who also writes a short description for the story.
These people don’t get to pick what Facebook adds to the trending section.
“That’s done automatically by the algorithm. They just get to pick the headline,” the report added.
Facebook has dismissed a media report that claimed journalists working as factcheckers for the social media giant are frustrated and are ending partnerships as the company failed to use their expertise to combat misinformation.
A report in The Guardian on Thursday said outside reporters have lost trust in Facebook, “which has repeatedly refused to release meaningful data about the impacts of their work”.
Reacting to the report, Meredith Carden, Head of News Integrity Partnerships at Facebook, said the Guardian story presents several inaccuracies.
“Contrary to a claim in the story, we absolutely do not ask fact-checkers to prioritise debunking content about our advertisers,” Carden said in statement.
The report, she added, is based primarily on the account of a single fact-checker who has not been involved with the Facebook fact-checking program for six months.
“We have been committed to fighting misinformation for years now and have strong relationships with our third-party fact-checking partners — we now have 35 partners in 24 countries around the world,” said Facebook.
The report quoted Brooke Binkowski, former managing editor of Snopes, a factchecking site that has partnered with Facebook for two years, as saying that the social network is using journalists for handling crisis PR.
“They’re not taking anything seriously. They are more interested in making themselves look good and passing the buck… They clearly don’t care,” said Binkowski, who now runs her own fact-checking site which does not partner with Facebook.
According to Facebook, it values the ongoing partnerships and the work that these journalists do.
The third-party fact checking programme was launched in 2016 after the US Presidential election.
“We’re planning to expand the programme to even more countries in 2019,” said Carden.
According to Facebook, three separate researches have found that the overall volume of false news on Facebook is decreasing since it put up third-party fact-checking programme and other anti-misinformation measures in place.
However, The Guardian report said the company has ignored journalists’ concerns.
Some newsroom leaders said “they had grown increasingly resentful of Facebook, especially following revelations that the company had paid a consulting firm to go after opponents by publicising their association with billionaire Jewish philanthropist George Soros”.
A New York Times investigation in November suggested that the social network hired a Republican-owned political consulting and PR firm that “dug up dirt on its competitors” including Soros.