Skip to main content

Wired - June 2019

"The internet is becoming a low-trust society, where an assumption of pervasive fraud is built into the way things function."

Online fakery runs wide and deep, but you don’t need me to tell you that. New species of digital fraud and deception come to light almost every week, if not every day: Russian bots that pretend to be American humans. American bots that pretend to be human trolls. Even humans that pretend to be bots. Yep, some “intelligent assistants,” promoted as advanced conversational AIs, have turned out to be little more than digital puppets operated by poorly paid people.

The internet was supposed to not only democratize information but also rationalize it—to create markets where impartial metrics would automatically surface the truest ideas and best products, at a vast and incorruptible scale. But deception and corruption, as we’ve all seen by now, scale pretty fantastically too.

According to ReviewMeta, an independent site that tracks the veracity of online feedback, there’s recently been a tremendous increase in Amazon reviews reviews written by users who have not made a verified purchase of the item they’re reviewing.

Surprise, surprise: Almost all of these unverified purchasers (98.2 percent) give the product five stars. Claims of fakery might also be fake. On Amazon, you can hardly shop for a simple sunscreen without encountering reviews claiming the product is counterfeit. Relieved to have been warned, you might be tempted to click away. But maybe that review itself was fake, planted by a competitor.

Big platforms like Google and Facebook earn their living mainly by taking money from advertisers and then delivering them eyeballs, with online ad brokers serving as middlemen. The idea is that these ads are precisely targeted and precisely measured, so that a brand is paying only for the eyeballs it wants to target and can actually surveil its audience to see exactly how long they’re watching ads. This advertising model is essentially the economic premise of the modern internet. But it’s one that has proven deeply susceptible to fraud—rife with fake views, fake clicks, and fake eyeballs.

In 2016, Facebook fessed up that, for two years, it had vastly overstated how long, on average, people were watching videos on the platform. The company characterized this as an “error” that hadn’t affected billings. But in 2018 a class action lawsuit brought by several small advertisers alleged that the social network had been inflating its figures by even more than it acknowledged—and that the company had known about it for longer than it let on.

At some point, the typical response to this onslaught of falsehood is to say, lol, nothing matters. But when so many of us are reaching this point, it really does matter. Social scientists distinguish high-trust societies (ones where you can expect most interactions to work) from low-trust societies (ones where you have to be on your guard at all times). People break rules in high-trust societies, of course, but laws, regulations, and norms help to keep most abuses in check; if you have to go to court, you expect a reasonable process. In low-trust societies, you never know. You expect to be cheated, often without recourse. You expect things not to be what they seem and for promises to be broken, and you don’t expect a reasonable and transparent process for recourse. It’s harder for markets to function and economies to develop in low-trust societies. It’s harder to find or extend credit, and it’s risky to pay in advance.

The internet is increasingly a low-trust society—one where an assumption of pervasive fraud is simply built into the way many things function.

People do adapt to low-trust societies, of course. Word-of-mouth recommendations from familiar sources become more important. Doing business with family and local networks starts taking precedence, as reciprocal, lifelong bonds bring a measure of predictability. Mafia-like organizations also spring up, imposing a kind of accountability at a brutal cost.

Ultimately, people in low-trust societies may welcome an authoritarian ruler, someone who will impose order and consequences from on high. Sure, the tyrant is also corrupt and cruel; but the alternative is the tiring, immiserating absence of everyday safety and security. During the reign of Kublai Khan, it was said that “a maiden bearing a nugget of gold on her head could wander safely throughout the realm.” The Great Khan required absolute submission, but even repression has some seeming perks.

In the digital world, the big platforms— Apple, Facebook, Google—sometimes exercise a Khan-like role in the way they impose order within their realms. Google nukes content farms; Apple rules its App Store with an iron grip; Amazon’s return policy—generous to customers but stringent to vendors—serves as a check against fraud; Facebook and Twitter have been pressured to de-platform the most noxious purveyors of conspiracy theories and fake news. And when they crack down, people cheer. ...
Read full article at Wired