FreeSpeech and Censorship: Thread

grarpamp grarpamp at gmail.com
Mon Mar 29 00:39:25 PDT 2021


https://taibbi.substack.com/p/alternatives-to-censorship-interview
https://taibbi.substack.com/p/meet-the-censored-status-coup

Alternatives To Censorship: Interview With Matt Stoller By Matt Taibbi

Authored by Matt Taibbi via TK News,

Led by Chairman Frank Pallone, the House Energy and Commerce Committee
Thursday held a five-hour interrogation of Silicon Valley CEOs
entitled, “Disinformation Nation: Social Media's Role in Promoting
Extremism and Misinformation.”

As Glenn Greenwald wrote yesterday, the hearing was at once
agonizingly boring and frightening to speech advocates, filled with
scenes of members of Congress demanding that monopolist companies
engage in draconian crackdowns.

[Click and drag to move]

Again, as Greenwald pointed out, one of the craziest exchanges
involved Texas Democrat Lizzie Fletcher:

Fletcher brought up the State Department’s maintenance of a list of
Foreign Terrorist Organizations. She praised the CEOs of Twitter,
Facebook, and Google, saying that “by all accounts, your platforms do
a better job with terrorist organizations, where that post is
automatically removed with keywords or phrases and those are
designated by the state department.”

Then she went further, chiding the firms for not doing the same
domestically. asking, “Would a federal standard for defining a
domestic terror organization similar to [Foreign Terrorist
Organizations] help your platforms better track and remove harmful
content?”

At another point, Fletcher noted that material from the January 6th
protests had been taken down (for TK interviews of several of the
videographers affected, click here) and said, “I think we can all
understand some of the reasons for this.” Then she complained about a
lack of transparency, asking the members, “Will you commit to sharing
the removed content with Congress?” so that they can continue their
“investigation” of the incident.

Questions like Fletcher’s suggest Congress wants to create a
multi-tiered informational system, one in which “data transparency”
means sharing content with Congress but not the public.

Worse, they’re seeking systems of “responsible” curation that might
mean private companies like Google enforcing government-created lists
of bannable domestic organizations, which is pretty much the opposite
of what the First Amendment intended.

Under the system favored by Fletcher and others, these monopolistic
firms would target speakers as well as speech, a major departure from
our current legal framework, which focuses on speech connected to
provable harm.

As detailed in an earlier article about NEC appointee Timothy Wu,
these solutions presuppose that the media landscape will remain highly
concentrated, the power of these firms just deployed in a direction
more to the liking of House members like Fletcher, Pallone,
Minnesota’s Angie Craig, and New York’s Alexandria Ocasio-Cortez, as
well as Senators like Ed Markey of Massachusetts. Remember this quote
from Markey: “The issue isn’t that the companies before us today are
taking too many posts down. The issue is that they’re leaving too many
dangerous posts up.”

    Remember: the last time Congress summed social media CEOs to be
interrogated, Sen. @EdMarkey left no doubt about what the demand of
Democrats is when doing this: we want you to censor more (and
obviously, the content they want censored is from their political
adversaries): pic.twitter.com/j64rCdZ82L
    — Glenn Greenwald (@ggreenwald) February 20, 2021

These ideas are infected by the same fundamental reasoning error that
drove the Hill’s previous drive for tech censorship in the Russian
misinformation panic. Do countries like Russia (and Saudi Arabia,
Israel, the United Arab Emirates, China, Venezuela, and others)
promote division, misinformation, and the dreaded “societal discord”
in the United State? Sure. Of course.

But the sum total of the divisive efforts of those other countries
makes up at most a tiny fraction of the divisive content we ourselves
produce in the United States, as an intentional component of our
commercial media system, which uses advanced analytics and engagement
strategies to get us upset with each other.

As Matt Stoller, Director of Research at the American Economic
Liberties Project puts it, describing how companies like Facebook make
money:

    It's like if you were in a bar and there was a guy in the corner
that was constantly egging people onto getting into fights, and he got
paid whenever somebody got into a fight? That's the business model
here.

As Stoller points out in a recent interview with Useful Idiots, the
calls for Silicon Valley to crack down on “misinformation” and
“extremism” is rooted in a basic misunderstanding of how these firms
make money. Even as a cynical or draconian method for clamping down on
speech, getting Facebook or Google to eliminate lists of taboo
speakers wouldn’t work, because it wouldn’t change the core function
of these companies: selling ads through surveillance-based herding of
users into silos of sensational content.

These utility-like firms take in data from everything you do on the
Web, whether you’re on their sites or not, and use that information to
create a methodology that allows a vendor to buy the most effective
possible ad, in the cheapest possible location. If Joe Schmo Motors
wants to sell you a car, it can either pay premium prices to advertise
in a place like Car and Driver, or it can go to Facebook and Google,
who will match that car dealership to a list of men aged 55 and up who
looked at an ad for a car in the last week, and target them at some
other, cheaper site.

In this system, bogus news “content” has the same role as porn or cat
videos — it’s a cheap method of sucking in a predictable group of
users and keeping them engaged long enough to see an ad. The salient
issue with conspiracy theories or content that inspires “societal
discord” isn’t that they achieve a political end, it’s that they’re
effective as attention-grabbing devices.

The companies’ use of these ad methods undermines factuality and
journalism in multiple ways. One, as Stoller points out, is that the
firms are literally “stealing” from legitimate news organizations.
“What Google and Facebook are doing is they're getting the proprietary
subscriber and reader information from the New York Times and Wall
Street Journal, and then they're advertising to them on other
properties.”

As he points out, if a company did this through physical means —
breaking into offices, taking subscriber lists, and targeting the
names for ads — “We would all be like, ‘Wow! That's outrageous. That's
crazy. That's stealing.’” But it’s what they do.

Secondly, the companies’ model depends upon keeping attention siloed.
If users are regularly exposed to different points of view, if they
develop healthy habits for weighing fact versus fiction, they will be
tougher targets for engagement.

So the system of push notifications and surveillance-inspired news
feeds stresses feeding users content that’s in the middle of the
middle of their historical areas of interest: the more efficient the
firms are in delivering content that aligns with your opinions, the
better their chance at keeping you engaged.

Rope people in, show them ads in spaces that in a vacuum are cheap but
which Facebook or Google can sell at a premium because of the intel
they have, and you can turn anything from QAnon to Pizzagate into cash
machines.

After the January 6th riots, Stoller’s organization wrote a piece
called, “How To Prevent the Next Social Media-Driven Attack On
Democracy—and Avoid a Big Tech Censorship Regime” that said:

    While the world is a better place without Donald Trump’s Twitter
feed or Facebook page inciting his followers to violently overturn an
election, keeping him or other arbitrarily chosen malignant actors off
these platforms doesn’t change the incentive for Facebook or other
social networks to continue pumping misinformation into users’ feeds
to continue profiting off of ads.

In other words, until you deal with the underlying profit model, no
amount of censoring will change a thing. Pallone hinted that he
understood this a little on Thursday, when he asked Zuckerberg if it
were true, as the Wall Street Journal reported last year, that in an
analysis done in Germany, researchers found that “Facebook’s own
engagement tools were tied to a significant rise in membership in
extremist organizations.” But most of the questions went in the other
direction.

“The question isn't whether Alex Jones should have a platform,”
Stoller explains. “The question is, should YouTube have recommended
Alex Jones 15 billion times through its algorithms so that YouTube
could make money selling ads?”

Below is an excerpted transcript from the Stoller interview at Useful
Idiots, part of which is already up here. When the full video is
released, I’ll update and include it.

Stoller is one of the leading experts on tech monopolies. He wrote the
Simon and Schuster book, Goliath: The Hundred Year War Between
Monopoly Power and Democracy, and is a former policy advisor to the
Senate Budget Committee. His writing has appeared in the Washington
Post, the New York Times, Fast Company, Foreign Policy, the Guardian,
Vice, The American Conservative, and the Baffler, among others.
Excerpts from his responses to questions from myself and Katie Halper
are below, edited for clarity:

Matt Taibbi: There's a debate going on within the Democratic
Party-aligned activist world about approaches to dealing with problems
in the speech world. Could you summarize?

Matt Stoller: There are two sides. One bunch of people has been
saying, “Hey, these firms are really powerful…” This is the
anti-monopoly wing. Google and Facebook, let’s break them up, regulate
them. They're really powerful and big, and that's scary. So, without
getting in too deep, there's the Antitrust subcommittee, that's been
saying, “Hey, these firms are really powerful, and they're picking and
choosing winners.” Usually, they talk about small businesses, but the
issue with speech is the same thing.

Then there’s another side, which is, I think, noisier and has more of
the MSNBC/CNN vibe. This is the disinformation/misinformation world.
This is the Russiagate people, the “We don't like that Trump can
speak” type of people. What their argument is, effectively, is that
firms haven't sufficiently curated their platforms to present what
they think is a legitimate form of public debate. They're thinking,
“Well, we need to figure out how to get them to filter differently,
and organize discourse differently.”

Ideologically, they just accept the dominance of these firms, and
they're just saying, “What's the best way for these firms to organize
discourse?”

Taibbi: By conceding the inevitability of these firms, they’re making
that concession, but saying they want to direct that power in a
direction that they'd like better.

Stoller: That's right. I mean, there's a lot of different reasons for
that. Some of them are neoliberal. A lot of the law professors are
like, “Oh, this is just the way of technology, and this is more
efficient.” Therefore, the question is, “How do you manage these large
platforms?” They're just inevitable.

Then there are people who are actually socialists who think, “Well,
the internet killed newspapers. The internet does all of these things.
Also, there's a bunch of them that never liked commercial press in the
first place. A lot of well-meaning people were like, “We never liked
advertising models to begin with. We think everything should be like
the BBC.”

So, those are the two groups that accept the inevitability thesis.
It's really deep-rooted in political philosophy. It's not just a
simple disagreement. Then there are people like us who are like, “No,
no. Actually, technology is deployed according to law and regulation,
and this specific regulatory model that we have, the business
structures of these firms, the way they make money from advertising,
those are specific policy choices, and we can make different ones if
we want.”

Katie Halper: When you say socialist, some may identify as socialists,
but that there's a general group of people who just believe, “We
oppose hate speech and White supremacy,” and so we have to make these
companies that are evil, and give them moral authority and a content
moderation authority, which is an inherent contradiction/wishful
thinking/inconsistent paradox.

In other words, you're saying leftists, right? Leftist, not liberals,
not neo-liberals, not even liberals, but people who are really would
identify as left.

Stoller: Yes. There's a part of the socialist world that's like, “What
we really want is egalitarianism in the form of a giant HR compliance
department that tells everyone to be tolerant.” Right? Then there are
most people who are like, “No. I just don't like wall street and I
want people to be equal and everyone should have a little bit over
something,” and they both call themselves socialists.

Taibbi: You and the American Economic Liberties Project have said,
there's a reason why taking Trump off Twitter isn't going to fix the
problem, because you're not fixing those incentives. Can you talk
about what those incentives are, and why they cause the problems?

Stoller: Google and Facebook, they sell ads, right? They collect lots
of information on you and they sell ads and ads are valuable for two
reasons. One, you're looking at them. Two, if they know who you are
and they know information about you, then they can make the ad more
valuable. A random ad space isn't worth very much, if you're showing
it to some undefined person. An ad space you're showing to a
55-year-old man who's thinking of buying a luxury car, somebody will
pay a lot for that ad space, if you know who that person is and you
know that that person has actually been browsing luxury car websites
and reading the Wall Street Journal about how best to liquidate their
portfolio or something to buy a luxury item.

Google and Facebook want to sell that advertising particularly on
their properties, where they get to keep 100% of the profits. If
Google sells an ad on YouTube, they get to keep the money. Facebook
sells an ad on Instagram or Facebook, they get to keep the money. So,
their goal is to keep you using their sites and to collect information
on you.

Taibbi: What methods do they use to keep you on the sites?

Stoller: They have all sorts of psychological tricks. Engagement is
the way that they talk about it, but it's like if you go and you look
for something on YouTube, they're going to send you something that's a
little bit more extreme. It's not necessarily just political. It's
like if you're a vegetarian, they'll say, or if you look at stuff
that's like, “Here's how to become a vegetarian,” they'll say, “Well,
about becoming a vegan?” If you look at stuff that suggests you’re a
little bit scared of whether this vaccine will work, if you search
for, “I would want to find safety data on this vaccine,” eventually,
they'll move you to like serious anti-vax world.

So, the question that we have to ask is whether you should block crazy
people from saying things, or do something else… Like Alex Jones, for
example, is crazy person or an entertainer, he says things that I
don't particularly like or agree with. The question, though, isn't
whether Alex Jones should have a platform. We've always allowed people
to go to the corner of the street and say whatever they want or to
write pamphlets or whatever.

The question is, should YouTube have recommended Alex Jones 15 billion
times through its algorithms so that YouTube could make money selling
ads? That's a different question than censorship.

Taibbi: Conversely they’re not recommending other material that might
discourage you from believing Alex Jones.

Stoller: Right. The other thing is, it's not just that they want to
create more inventory so they can sell ads. It's also the kinds of ads
that they're selling. So, you can sell an ad based on trust. The New
York Times or the Wall Street Journal — I hate using them as examples
— they have an audience and people. They built that audience by
investing in content, and then they sell ads to that audience, and the
advertiser knows where that advertising is going and it's based on
trust.

The alternative model which we have now is simply based on clickbait.
It's just, “Generate as many impressions as possible, and then sell
those impressions wherever the person is on the web.” That creates a
kind of journalism, which is designed to get clicks or not even
journalism. It's just you're creating content just to get engagement
and not actually to build trust.

So, what this business model does, we call it surveillance
advertising, but it's an infrastructure player, a communications
player manipulating you so that they could put content, engage content
in front of you. What that does is it incentivizes a low trust form of
content production. It both kills trusted content producers, a.k.a.
local newspapers, because you no longer need to be able to put
advertising in the Pittsburgh Post-Gazette, or whatever. You can just
geotarget those people through Google and Facebook. You can get some
Eastern European to falsify stories and people will click on that.

So, it kills legitimate newspapers and it creates an incentive for low
trust content, fraudulent content, defamatory content, whatever it is
that will keep people engaged and is often fraudulent. It hits really
local newspapers and niche newspapers the most, so Black-owned
newspapers and also newspapers having to do with hobbies. The actual
issue is more about niche audiences themselves, and the kind of
low-trust content that we're encouraging with our policy framework,
versus what we used to do, which we would encourage higher trust forms
of content.

Taibbi: How would you fix this problem, from a regulatory perspective?

Stoller: The House Antitrust Subcommittee had a report where they
recommended what we call regulated competition. That would say, “Okay.
You break up these platforms in ways that wouldn't interfere with the
efficiency of that particular network system.” So, Google and YouTube
don't need to be in the same company, you could separate them out.

There are ways that you'd have to handle the search engine. You
couldn't split Facebook into five Facebooks, because then you wouldn't
necessarily be able to talk to your friends and family, but you could
separate Instagram and Facebook easily. You could force
interoperability and then split up Facebook if you want to do that.
So, you could separate those things out and then ban surveillance
advertising for a starter.

Taibbi: What would that do to content if you ban surveillance
advertising? ANd how would that work?

Stoller: It would force advertisers to go back to advertising to
audiences. So, they would no longer be able to track you as an
individual and say, “We know this is what you're interested in.” They
would go to what's called contextual advertising, and they would say,
“Okay. If you're on a site that has to do with tennis, then we'll
advertise tennis rackets on that site because we assume that that
people are interested in tennis rackets.”

That's called contextual advertising, versus the current system: you
read an article about tennis in a tennis magazine and the platforms
say, “Oh, that's expensive to buy an ad there, so we'll track you
around the web and when you're on Candy Crush, we'll show you a tennis
racket ad.” That's the surveillance advertising model we have. That
pulls all the power to Google and Facebook who are doing all the
tracking, versus the contextual ad where the power is actually with
the tennis racket site that has the relationship with the people
interested in tennis.

Taibbi: So, the idea would be you would create a sort of a firewall
between the utilitarian functions of a site like Facebook or Google,
that provide a service where either you're searching for something or
you're communicating with somebody, and they wouldn't be allowed to
take that data from that utility-like function to sell you an ad?

Stoller: That's right. Germany is hearing a court case saying that you
can't combine advertising from Facebook, Instagram, and WhatsApp, and
then third parties to create a super profile of someone and show them
ads. They were saying that's an antitrust violation. There's a court
hearing on that, but, more broadly, that's what you have to do.

Ultimately, what we would want is we would want to have
subscription-based communication networks paying for services. This is
something that's worked for thousands of years. I give you something
in value, you give me money. It's an honest way of doing business. If
I don't value it enough to give you money, then I won't get it.

If people are like, “Oh, I don't want to pay for Facebook, or I don't
want to pay for YouTube,” or whatever it is, that makes no sense.
You're already paying. You're either paying with a Friday night you
spend surfing YouTube, where they sell a bunch of ads and you give up
your Friday night — or you pay with money, and it's an honest
transaction, and it's in the long run, a lot cheaper and more honest
method of payment.


More information about the cypherpunks mailing list