24 June 2022

FACEBOOK >> TAKE-THE-REVENUE-AND-RUN-THE ADS! ...'Strategic Transparency' to Deceive Users and Obscure How Little The Public Knows

Intro: Sarah Kay Wiley, a researcher at the Tow Center for Digital Journalism at Columbia University, said that Meta’s reports identifying CIB networks of fake accounts and pages that aim to deceive users obscure how little researchers and the public still know about what goes on within the company, and on its platforms.
Between July 2018 and April 2022, Meta made at least $30.3 million in ad revenue from networks it removed from its own platforms for engaging in coordinated inauthentic behavior (CIB), data compiled by WIRED shows

Meta Made Millions in Ads From Networks of Fake Accounts

The social media giant banned accounts promoting disinformation, spam, or propaganda—and kept the money it made from ads.

<div class=__reading__mode__extracted__imagecaption>ILLUSTRATION: SIMON ABRANOWICZ

"When Meta’s Mark Zuckerberg was called to testify before Congress in 2018, he was asked by Senator Orin Hatch how Facebook made money. Zuckerberg’s answer has since become something of a meme: “Senator, we run ads.”

Between July 2018 and April 2022, Meta made at least $30.3 million in ad revenue from networks it removed from its own platforms for engaging in coordinated inauthentic behavior (CIB), data compiled by WIRED shows. Margarita Franklin, head of security communications at Meta, confirmed to WIRED that the company does not return the ad money if a network is taken down. Franklin clarified that some of the money came from adverts that didn't break the company's rules, but were published by the same public relations or marketing organizations later banned for participating in CIB operations.

A report from The Wall Street Journal estimates that by the end of 2021, Meta absorbed 17 percent of the money in the global ad market and made $114 billion from advertising. At least some of the money came from ads purchased by networks that violated Meta’s policies and that the company itself has flagged and removed.

“The advertising industry globally is estimated to be about $400 billion to $700 billion,” said Claire Atkin, cofounder of the independent watchdog Check My Ads Institute. “That is a large brush, but nobody knows how big the industry is. Nobody knows what goes on inside of it.”

But Atkin says that part of what makes information, including ads, feel legitimate on social media is the context they appear in. “Facebook, Instagram, WhatsApp, this entire network within our internet experience, is where we connect with our closest friends and family. This is a place on the internet where we share our most intimate emotions about what’s happening in our lives,” says Atkin. “It is our trusted location for connection.”

For nearly four years, Meta has released periodic reports identifying CIB networks of fake accounts and pages that aim to deceive users and, in many cases, push propaganda or disinformation in ways that are designed to look organic and change public opinion. These networks can be run by governments, independent groups, or public relations and marketing companies.

Last year, the company also began addressing what it dubbed “coordinated social harm,” where networks used real accounts as part of their information operations. Nathaniel Gleicher, head of security policy at Meta, announced the changes in a blog post, noting that “threat actors deliberately blur the lines between authentic and inauthentic activities, making enforcement more challenging across our industry.”

This change, however, demonstrates how specific the company’s criteria for CIB is, which means that Meta may not have documented some networks that used other tactics at all. Information operations can sometimes use real accounts, or be run on behalf of a political action committee or LLC, making it more difficult to categorize their behavior as “inauthentic.”

“One tactic that's been used more frequently, at least since 2016, has been not bots, but actual people that go out and post things,” says Sarah Kay Wiley, a researcher at the Tow Center for Digital Journalism at Columbia University. “The CIB reports from Facebook, they kind of get at it, but it's really hard to spot.”

[    ] (Meta’s public earnings documents do not break down how much the company earns by country, only by region.)

. . .

Many of the largest networks that Meta removed were run by public relations or marketing firms, like the Archimedes Group in Israel and Pragmatico in Ukraine. When this happens, Meta will remove and ban every account and page associated with that firm, whether or not it is involved in a particular CIB campaign, in an effort to discourage businesses from selling “disinformation for hire” services.

CIB campaigns and disinformation are not limited to Facebook and Instagram. Twitter, which labels such activity “information operations,” has identified and removed thousands of accounts on its own platform. Though researchers have identified disinformation campaigns on TikTok, the company’s Community Guidelines Enforcement Reports do not indicate whether or how the platform deals with artificially boosted content.

Wiley says that Meta’s reports obscure how little researchers and the public still know about what goes on within the company, and on its platforms. In a January report, Meta said that due to evolving threats against its teams, it would “prioritize enforcement and the safety of our teams over publishing our findings,” which could make transparency worse.

“Is this the tip of the iceberg? Unfortunately, I think it is,” says Wiley. . .

>

No comments:

QOD: You can dig it