Comment: Social media may be losing its algorithm camouflage

Regulators in Europe are looking at ways to understand AI through its effects rather than its data.

By Parmy Olson / Bloomberg Opinion

There’s a perfectly good reason to break open the secrets of social media giants. Over the past decade, governments have watched helplessly as their democratic processes were disrupted by misinformation and hate speech on sites like Meta Platforms’ Facebook, Alphabet’s YouTube and Twitter. Now some governments are gearing up for a comeuppance.

In the next two years, Europe and the United Kingdom are preparing laws that will rein in the troublesome content that social media firms have allowed to go viral. There has been much skepticism over their ability to look under the hood of companies like Facebook. Regulators, after all, lack the technical expertise, manpower and salaries that Big Tech boasts. And there’s another technical snag: The artificial-intelligence systems tech firms use are notoriously difficult to decipher.

But naysayers should keep an open mind. New techniques are developing that will make probing those systems easier. AI’s “black box” problem isn’t as impenetrable as many think.

AI powers most of the action we see on Facebook or YouTube and, in particular, the recommendation systems that line up which posts go into your newsfeed, or what videos you should watch next; all to keep you scrolling. Millions of pieces of data are used to train AI software, allowing it to make predictions loosely similar to humans’. The hard part, for engineers, is understanding how AI makes a decision in the first place. Hence the black-box concept.

Consider two pictures, one of a fox and the other of a dog:

You can probably tell within a few milliseconds which animal is the fox and which is the dog. But can you explain how you know? Most people would find it hard to articulate what it is about the nose, ears or shape of the head that tells them which is which. But they know for sure which picture shows the fox.

A similar paradox affects machine-learning models. It will often give the right answer, but its designers often can’t explain how. That doesn’t make them completely inscrutable. A small but growing industry is emerging that monitors how these systems work. Their most popular task: Improve an AI model’s performance. Companies that use them also want to make sure their AI isn’t making biased decisions when, for example, sifting through job applications or granting loans.

Here’s an example of how one of these start-ups works. A financial firm recently used Israeli start-up Aporia to check whether a campaign to attract students was working. Aporia, which employs software and human auditors, found that the company’s AI system was actually making errors, granting loans to some young people it shouldn’t have, or withholding loans from others unnecessarily. When Aporia looked closer, it found out why: Students made up less than 1 percent of the data the firm’s AI had been trained on.

In a lot of ways, the reputation of AI’s black box for impenetrability has been exaggerated, according to Aporia’s chief executive officer, Liran Hosan. With the right technology, you can even — potentially — unpick the ultra-complicated language models that underpin social media firms, in part because in computing, even language can be represented by numerical code. Finding out how an algorithm might be spreading hate speech, or failing to tackle it, is certainly harder than spotting mistakes in the numerical data that represent loans, but it’s possible. And European regulators are going to try.

According to a spokesman for the European Commission, the upcoming Digital Services Act will require online platforms to undergo audits once a year to assess how “risky” their algorithms are to citizens. That may sometimes force firms to provide unprecedented access to information that many consider trade secrets: code, training data and process logs. (The commission said its auditors would be bound by confidentiality rules.)

But let’s suppose Europe’s watchdogs couldn’t delve into Facebook or YouTube code. Suppose they couldn’t probe the algorithms that decide what videos or posts to recommend. There would still be much they could do.

Manoel Ribeiro, a doctoral student at the Swiss Federal Institute of Technology in Lausanne, Switzerland, published a study in 2019 in which he and his co-authors tracked how certain visitors to YouTube were being radicalized by far-right content. He didn’t need to access any of YouTube’s code to do this. The researchers simply looked at comments on the site to see what channels users went to over time. It was like tracking digital footprints; painstaking work, but it ultimately revealed how a fraction of YouTube users were being lured into white-supremacist channels by way of influencers who acted like a gateway drug.

Ribeiro’s study is part of a broader array of research that has tracked the psychological side effects of Facebook or YouTube without needing to understand their algorithms. While offering relatively superficial perspectives of how social media platforms work, they can still help regulators impose broader obligations on the platforms. These can range from hiring compliance officers to ensure a company is following the rules, or giving accurate, random samples to auditors about the kinds of content people are being driven toward.

That is a radically different prospect to the secrecy that Big Tech has been able to operate under till now. And it’ll involve both new technology and new policies. For regulators, that could well be a winning combination.

Parmy Olson is a Bloomberg Opinion columnist covering technology.

Talk to us

More in Opinion

Editorial cartoons for Sunday, June 4

A sketchy look at the news of the day.… Continue reading

Lummi Tribal members Ellie Kinley, left, and Raynell Morris, president and vice president of the non-profit Sacred Lands Conservancy known as Sacred Sea, lead a prayer for the repatriation of southern resident orca Sk’aliCh’elh-tenaut — who has lived and performed at the Miami Seaquarium for over 50 years — to her home waters of the Salish Sea at a gathering Sunday, March 20, 2022, at the sacred site of Cherry Point in Whatcom County, Wash.

The Bellingham Herald
Editorial: What it will require to bring Tokitae home

Bringing home the last captive orca requires expanded efforts to restore the killer whales’ habitat.

AI ethics or AI Law concept. Developing AI codes of ethics. Compliance, regulation, standard , business policy and responsibility for guarding against unintended bias in machine learning algorithms.
Comment: What Congress can do to keep an eye on AI

It needs to establish guardrails, ensure accountability and keep the technology monopolies honest.

County auditor: Fell best suited for reelection to post

Garth Fell is the best candidate to continue to serve the Snohomish… Continue reading

Work zone speed cameras a money grab

Regarding the editorial about work zone speed cameras (“Set your muscle memory… Continue reading

Comment: What capital gains tax’s court win means for so many

The state Supreme Court’s decision makes the state’s taxes more fair and provides revenue to aid many.

Comment: State’s high court ignores precedent in writing its rules

In seeking to end ‘systemic racial injustice,’ court’s justices ignore constitutional constraints.

Comment: Public safety lost ground in this year’s Legislature

Legislation that would have better addressed racism’s effects on communities was not adopted by lawmakers.

A map of the I-5/SR 529 Interchange project on Tuesday, May 23, 2023 in Marysville, Washington. (Olivia Vanni / The Herald)
Editorial: Set your muscle memory for work zone speed cameras

Starting next summer, not slowing down in highway work zones can result in a $500 fine.

Most Read