Most Information About Disinformation Is Misinformation
from the disinfo-about-disinfo dept
Reporter Joseph Bernstein recently published a fantastic cover story in Harpers all about "disinformation" and "fake news" but not in the way you normally think about it. It's not like most such articles -- often decrying just how much disinformation is flowing out there, but rather taking a very critical eye about how we (especially the media) talk about such things. The piece is thought-provoking and well worth reading, and I've spent the last week or so letting it sit and percolate in my head before writing up this post about it.
Right after the 2016 election, there was a flurry of hand-wringing from the media, trying to understand how what they were positive would happen (a Hillary Clinton victory and a Donald Trump loss), didn't actually happen. A convenient scapegoat for this surprising turn of events was... Facebook. The narrative took over that it was "fake news" on Facebook that convinced a bunch of gullible people to support a clearly unqualified candidate. That this convenient scapegoat also happened to be successfully siphoning advertising dollars away from some traditional media organizations was mostly just made pointing fingers at it feel even better. However, as we warned at the time, focusing in on social media and "fake news" was not just silly, but potentially counterproductive. Indeed, within weeks, authoritarians around the world started adopting the term "fake news" as a convenient excuse for censoring the media. And, obviously, it became a key part of Donald Trump's stump speech as well.
It wasn't long until "fake news" was used against any content someone in power didn't like, and it became a key tool to push for censorship of those who were actually exposing malfeasance.
Bernstein's article highlights how this same sort of thinking is happening with the term "disinformation." Of course, disinformation doesn't have a clear definition, and often it's in the eye of the beholder (like "fake news" before it). But, the media (and many politicians) have become so obsessed with "disinformation" that, once again, we've turned it into a kind of moral panic -- and a convenient one for censors around the globe. As Bernstein notes, the hue and cry over "disinformation" has made many Americans think it's one of the biggest threats around:
Everyone scrounges this wasteland for tainted morsels of content, and it’s impossible to know exactly what anyone else has found, in what condition, and in what order. Nevertheless, our American is sure that what her fellow citizens are reading and watching is bad. According to a 2019 Pew survey, half of Americans think that “made-up news/info” is “a very big problem in the country today,” about on par with the “U.S. political system,” the “gap between rich and poor,” and “violent crime.” But she is most worried about disinformation, because it seems so new, and because so new, so isolable, and because so isolable, so fixable. It has something to do, she knows, with the algorithm.
An important (and often overlooked!) point that Bernstein makes in the piece is that the big internet companies were pretty quick to embrace this idea that "disinformation" is a problem. It is true that Mark Zuckerberg initially pushed back on the idea, but after basically everyone attacked him for that, he quickly began his apology tour. And, why not? Even as it makes the company look bad in the narrative, at its core, the idea that disinformation on Facebook impacted the American election can be spun to be positive for Facebook. After all, if Facebook is so powerful, shouldn't you be advertising on it, Mr. Toilet Paper maker? If Facebook can help get Trump elected because of some memes posted by idiots, just think how much beer and nachos it can sell as well? Literally, Facebook has a vested interest in having people believe that disinformation works on its platform, because that's literally something the company can profit from.
Denial was always untenable, for Zuckerberg in particular. The so-called techlash, a season of belatedly brutal media coverage and political pressure in the aftermath of Brexit and Trump’s win, made it difficult. But Facebook’s basic business pitch made denial impossible. Zuckerberg’s company profits by convincing advertisers that it can standardize its audience for commercial persuasion. How could it simultaneously claim that people aren’t persuaded by its content? Ironically, it turned out that the big social-media platforms shared a foundational premise with their strongest critics in the disinformation field: that platforms have a unique power to influence users, in profound and measurable ways. Over the past five years, these critics helped shatter Silicon Valley’s myth of civic benevolence, while burnishing its image as the ultra-rational overseer of a consumerist future.
Behold, the platforms and their most prominent critics both proclaim: hundreds of millions of Americans in an endless grid, ready for manipulation, ready for activation. Want to change an output—say, an insurrection, or a culture of vaccine skepticism? Change your input. Want to solve the “crisis of faith in key institutions” and the “loss of faith in evidence-based reality”? Adopt a better content-moderation policy. The fix, you see, has something to do with the algorithm.
But... is it actually true? Are most Americans actually gullible suckers who will fall for any piece of made up nonsense that gives them a dopamine hit? As Bernstein's piece explores, the American advertising industry has spent decades pushing this very notion -- even putting forth various industry-supported studies to "prove" it. Whether or not those studies are accurate portrayals of reality is another question, but if they're not, well, wouldn't that create quite a paradox? The disinformation from questionable studies about how it's possible to manipulate people into believing things worked -- so even if the studies are inaccurate, there's at least one example of how their own "disinformation" had an impact. And, as Bernstein does highlight, much of that early research on how easy it is to manipulate people with ads was... highly questionable.
The profitable relationship between the ad industry and the soft sciences took on a dark cast in 1957, when the journalist Vance Packard published The Hidden Persuaders, his exposé of “motivation research”—then the bleeding edge of collaboration between Madison Avenue and research psychology. The alarming public image Packard’s bestseller created—ad men wielding some unholy concoction of Pavlov and Freud to manipulate the American public into buying toothpaste—is still with us today. And the idea of the manipulability of the public is, as Arendt noted, an indispensable part of the product. Advertising is targeted at consumers, but sold to businesses.
Packard’s reporting was based on what motivation researchers told him. Among their own motivations, hardly hidden, was a desire to appear clairvoyant. In a late chapter, Packard admits as much:
Some of the researchers were sometimes prone to oversell themselves—or in a sense to exploit the exploiters. John Dollard, [a] Yale psychologist doing consulting work for industry, chided some of his colleagues by saying that those who promise advertisers “a mild form of omnipotence are well received.”
So, the public (and the media and politicians) have been primed -- with disinformation -- to believe that disinformation works. And the big internet companies have their own vested interest in continuing that belief, even if it's not quite true.
The media narrative of sinister digital mind control has obscured a body of research that is skeptical about the effects of political advertising and disinformation. A 2019 examination of thousands of Facebook users by political scientists at Princeton and NYU found that “sharing articles from fake news domains was a rare activity”—more than 90 percent of users had never shared any. A 2017 Stanford and NYU study concluded that
if one fake news article were about as persuasive as one TV campaign ad, the fake news in our database would have changed vote shares by an amount on the order of hundredths of a percentage point. This is much smaller than Trump’s margin of victory in the pivotal states on which the outcome depended.
But, of course, part of the problem is that -- like "fake news" before it -- there is no easy definition of "disinformation," and thus the research on it isn't always talking about the same thing.
The most comprehensive survey of the field to date, a 2018 scientific literature review titled “Social Media, Political Polarization, and Political Disinformation,” reveals some gobsmacking deficits. The authors fault disinformation research for failing to explain why opinions change; lacking solid data on the prevalence and reach of disinformation; and declining to establish common definitions for the most important terms in the field, including disinformation, misinformation, online propaganda, hyperpartisan news, fake news, clickbait, rumors, and conspiracy theories. The sense prevails that no two people who research disinformation are talking about quite the same thing.
This will ring true to anyone who follows the current media discussion around online propaganda. “Misinformation” and “disinformation” are used casually and interchangeably to refer to an enormous range of content, ranging from well-worn scams to viral news aggregation; from foreign-intelligence operations to trolling; from opposition research to harassment. In their crudest use, the terms are simply jargon for “things I disagree with.” Attempts to define “disinformation” broadly enough as to rinse it of political perspective or ideology leave us in territory so abstract as to be absurd.
Another key point that Bernstein highlights is why "disinformation" is such a convenient term for the traditional media. After all, for decades they've pushed the idea that they're the ones pushing "objective truth" and taking a view from nowhere. So they're the ones who are supposed to save us from disinformation. When you've positioned yourself as the gatekeeper for truth, it helps to play up the constant threat of untruths. As Bernstein notes:
A quick scan of the institutions that publish most frequently and influentially about disinformation: Harvard University, the New York Times, Stanford University, MIT, NBC, the Atlantic Council, the Council on Foreign Relations, etc. That the most prestigious liberal institutions of the pre-digital age are the most invested in fighting disinformation reveals a lot about what they stand to lose, or hope to regain. Whatever the brilliance of the individual disinformation researchers and reporters, the nature of the project inevitably places them in a regrettably defensive position in the contemporary debate about media representation, objectivity, image-making, and public knowledge. However well-intentioned these professionals are, they don’t have special access to the fabric of reality.
And they sure do seem eager to put in place official arbiters of truth:
Still, Big Disinfo can barely contain its desire to hand the power of disseminating knowledge back to a set of “objective” gatekeepers. In February, the tech news website Recode reported on a planned $65 million nonpartisan news initiative called the Project for Good Information. Its creator, Tara McGowan, is a veteran Democratic operative and the CEO of Acronym, a center-left digital-advertising and voter-mobilization nonprofit whose PAC is funded by, among others, Steven Spielberg, the LinkedIn co-founder Reid Hoffman, and the venture capitalist Michael Moritz. The former Obama campaign manager David Plouffe, currently a strategist at the Chan Zuckerberg Initiative, is an official Acronym adviser. Meanwhile, a February New York Times article humbly suggested the appointment of a “reality czar” who could “become the tip of the spear for the federal government’s response to the reality crisis.”
If you step back, you can absolutely see the thinking here, though it's difficult to see any path by which it is actually effective. It only does make sense if it really is true that false info on Facebook really fried many people's brains -- rather than a much larger, and much more complex series of variables that have all contributed to large segments of the society being open to crazy ideas and conspiracy theories. What's notable in all of this is how ready many people are to believe that a meme on Facebook magically convinced their aunt to believe in the dumbest conspiracy theories, but somehow they, themselves, are miraculously immune to such things.
None of this is to argue that conspiracy theories and those who push them are not a problem. Clearly it's a part of the problem, but we're so focused on the symptoms that we see, rather than the underlying causes, that we get wrapped up in bad ideas that sound good from a narrative perspective, but will have no real impact on the solution.
For years I've tried to highlight that what we see -- what Facebook and other social media has exposed -- is often the consequences of huge societal failings. There are problems with education, with social safety nets, with healthcare (especially mental healthcare). There are problems with income inequality and corruption. There are tons of problems out there, and many of these manifest themselves through false information that people share online. But saying that the "disinformation" is the problem -- rather than a way in which the underlying problem shows itself -- misses the point entirely.
And Bernstein's article really does a great job calling this out. Is disinformation real? Sure, though what it actually is remains amorphous. But the focus on treating disinformation as the problem, rather than simply an exposed symptom of a much deeper, much more complex, much more troubling underlying societal ill, is missing the point.
There's a lot more in Bernstein's piece, and I really recommend reading the whole thing. He doesn't necessarily come to the same conclusion I come to here, but it really includes a lot of useful thinking about how we've over-indexed on disinformation as the problem when it's not at all clear that it really is the problem. Indeed, it's easy to come away from the article and realize that there's an awful lot of, er... disinformation about disinformation. The problem is not Facebook. The problem is that Facebook is shining a light on a whole bunch of other terrible shit.
Only certain types of people respond to certain types of propaganda in certain situations. The best reporting on QAnon, for example, has taken into account the conspiracy movement’s popularity among white evangelicals. The best reporting about vaccine and mask skepticism has taken into account the mosaic of experiences that form the American attitude toward the expertise of public-health authorities. There is nothing magically persuasive about social-media platforms; they are a new and important part of the picture, but far from the whole thing. Facebook, however much Mark Zuckerberg and Sheryl Sandberg might wish us to think so, is not the unmoved mover.
For anyone who has used Facebook recently, that should be obvious. Facebook is full of ugly memes and boring groups, ignorant arguments, sensational clickbait, products no one wants, and vestigial features no one cares about. And yet the people most alarmed about Facebook’s negative influence are those who complain the most about how bad a product Facebook is. The question is: Why do disinformation workers think they are the only ones who have noticed that Facebook stinks? Why should we suppose the rest of the world has been hypnotized by it? Why have we been so eager to accept Silicon Valley’s story about how easy we are to manipulate?
None of this means that manipulation never works. Or that people are not persuadable. Of course people are persuadable. But not in the way that many people think. It's not because of a random meme that make people embrace crazy ideas about vaccines or 5G. There needs to be much more infrastructure and issues for people to be susceptible. And, as Bernstein notes, perhaps the most manipulated are those who feel they need to believe in this story of "disinformation" to explain away all of their other failures to build a better society:
Indeed, it’s possible that the Establishment needs the theater of social-media persuasion to build a political world that still makes sense, to explain Brexit and Trump and the loss of faith in the decaying institutions of the West. The ruptures that emerged across much of the democratic world five years ago called into question the basic assumptions of so many of the participants in this debate—the social-media executives, the scholars, the journalists, the think tankers, the pollsters. A common account of social media’s persuasive effects provides a convenient explanation for how so many people thought so wrongly at more or less the same time. More than that, it creates a world of persuasion that is legible and useful to capital—to advertisers, political consultants, media companies, and of course, to the tech platforms themselves. It is a model of cause and effect in which the information circulated by a few corporations has the total power to justify the beliefs and behaviors of the demos. In a way, this world is a kind of comfort. Easy to explain, easy to tweak, and easy to sell, it is a worthy successor to the unified vision of American life produced by twentieth-century television.
But I'd take that even a step further. Many of the "Establishment" that are pushing this are also the people whose previous policies failed. The reason we have so many societal problems are because their beliefs about their own policy powers and what they could accomplish in the past did not work the way they expected. The world did not progress the way they planned. And thus, pushing the "oh it's social media and disinformation" button not only gives them a convenient story, it also absolves themselves of their own failures.
I recognize that this can be read to be saying that Facebook isn't a problem. Or that "disinformation" and false info isn't a problem. But that's not what I am saying at all (and I'm pretty sure it's not what Bernstein is saying either). It's just that the world is a lot more complicated than that. What we're seeing on Facebook and the flow of disinformation is a problem -- but it's not a problem you solve by sweeping it under the rug. It's a mirror on the real underlying societal problems the world faces -- which we should be talking about and trying to come up with better solutions for, rather than insisting that Facebook can make it all go away if only they had a better algorithm or better employees.
Filed Under: cable news, content moderation, disinformation, fake news, finger pointing, media literacy, misinformation, newspapers, propaganda, social media