Comment: Social media may be losing its algorithm camouflage

Regulators in Europe are looking at ways to understand AI through its effects rather than its data.

By Parmy Olson / Bloomberg Opinion

There’s a perfectly good reason to break open the secrets of social media giants. Over the past decade, governments have watched helplessly as their democratic processes were disrupted by misinformation and hate speech on sites like Meta Platforms’ Facebook, Alphabet’s YouTube and Twitter. Now some governments are gearing up for a comeuppance.

In the next two years, Europe and the United Kingdom are preparing laws that will rein in the troublesome content that social media firms have allowed to go viral. There has been much skepticism over their ability to look under the hood of companies like Facebook. Regulators, after all, lack the technical expertise, manpower and salaries that Big Tech boasts. And there’s another technical snag: The artificial-intelligence systems tech firms use are notoriously difficult to decipher.

But naysayers should keep an open mind. New techniques are developing that will make probing those systems easier. AI’s “black box” problem isn’t as impenetrable as many think.

AI powers most of the action we see on Facebook or YouTube and, in particular, the recommendation systems that line up which posts go into your newsfeed, or what videos you should watch next; all to keep you scrolling. Millions of pieces of data are used to train AI software, allowing it to make predictions loosely similar to humans’. The hard part, for engineers, is understanding how AI makes a decision in the first place. Hence the black-box concept.

Consider two pictures, one of a fox and the other of a dog:

You can probably tell within a few milliseconds which animal is the fox and which is the dog. But can you explain how you know? Most people would find it hard to articulate what it is about the nose, ears or shape of the head that tells them which is which. But they know for sure which picture shows the fox.

A similar paradox affects machine-learning models. It will often give the right answer, but its designers often can’t explain how. That doesn’t make them completely inscrutable. A small but growing industry is emerging that monitors how these systems work. Their most popular task: Improve an AI model’s performance. Companies that use them also want to make sure their AI isn’t making biased decisions when, for example, sifting through job applications or granting loans.

Here’s an example of how one of these start-ups works. A financial firm recently used Israeli start-up Aporia to check whether a campaign to attract students was working. Aporia, which employs software and human auditors, found that the company’s AI system was actually making errors, granting loans to some young people it shouldn’t have, or withholding loans from others unnecessarily. When Aporia looked closer, it found out why: Students made up less than 1 percent of the data the firm’s AI had been trained on.

In a lot of ways, the reputation of AI’s black box for impenetrability has been exaggerated, according to Aporia’s chief executive officer, Liran Hosan. With the right technology, you can even — potentially — unpick the ultra-complicated language models that underpin social media firms, in part because in computing, even language can be represented by numerical code. Finding out how an algorithm might be spreading hate speech, or failing to tackle it, is certainly harder than spotting mistakes in the numerical data that represent loans, but it’s possible. And European regulators are going to try.

According to a spokesman for the European Commission, the upcoming Digital Services Act will require online platforms to undergo audits once a year to assess how “risky” their algorithms are to citizens. That may sometimes force firms to provide unprecedented access to information that many consider trade secrets: code, training data and process logs. (The commission said its auditors would be bound by confidentiality rules.)

But let’s suppose Europe’s watchdogs couldn’t delve into Facebook or YouTube code. Suppose they couldn’t probe the algorithms that decide what videos or posts to recommend. There would still be much they could do.

Manoel Ribeiro, a doctoral student at the Swiss Federal Institute of Technology in Lausanne, Switzerland, published a study in 2019 in which he and his co-authors tracked how certain visitors to YouTube were being radicalized by far-right content. He didn’t need to access any of YouTube’s code to do this. The researchers simply looked at comments on the site to see what channels users went to over time. It was like tracking digital footprints; painstaking work, but it ultimately revealed how a fraction of YouTube users were being lured into white-supremacist channels by way of influencers who acted like a gateway drug.

Ribeiro’s study is part of a broader array of research that has tracked the psychological side effects of Facebook or YouTube without needing to understand their algorithms. While offering relatively superficial perspectives of how social media platforms work, they can still help regulators impose broader obligations on the platforms. These can range from hiring compliance officers to ensure a company is following the rules, or giving accurate, random samples to auditors about the kinds of content people are being driven toward.

That is a radically different prospect to the secrecy that Big Tech has been able to operate under till now. And it’ll involve both new technology and new policies. For regulators, that could well be a winning combination.

Parmy Olson is a Bloomberg Opinion columnist covering technology.

Talk to us

More in Opinion

Editorial cartoons for Friday, Oct. 7

A sketchy look at the news of the day.… Continue reading

Everett police responded to a double homicide in the 2000 block of Lexington Avenue on the afternoon of June 20, 2022. (Jake Goldstein-Street / The Herald)
Editorial: Cities team up in effort for better public safety

The group of 15 cities in the county wants attention on legislation and the criminal justice system.

Schwab: Must we boycott all Republicans to send GOP message?

If, nationwide, they won’t abandon Trumpism and its lies, all voters ought to withhold their support.

AquaSox need new stadium, but keep the current prices

I moved to Stanwood from Kirkland in 2004, I had several good… Continue reading

Does baseball matter more than seniors?

So, Everett wants to spend millions for a new baseball stadium but… Continue reading

Iranian woman’s death in police custody indefensible

From the perspective of an Ahmadi Muslim, the heinous actions committed by… Continue reading

More businesses should add health care coverage

We are in odd times when it comes to the job market.… Continue reading

Comment: What Trump could gain with filing to Supreme Court

His arguments may find favor with justices willing to expand the executive privilege of former presidents.

Comment: What campaign ad buys say about parties and issues

Ads for Democrats are looking to persuade on abortion and more, where GOP candidates are looking to motivate.

Most Read