Why Facebook feeds division and stress

Mark Zuckerberg’s manifesto, penned clearly in response to accusations leveled in the wake of the US election, is a scary, dystopian document. It shows that Facebook — launched, in Zuckerberg’s own words, to “extend people’s capacity to build and...

Why Facebook feeds division and stress

Mark Zuckerberg’s manifesto, penned clearly in response to accusations leveled in the wake of the US election, is a scary, dystopian document. It shows that Facebook — launched, in Zuckerberg’s own words, to “extend people’s capacity to build and maintain relationships” — is turning into something of an extraterritorial state run by a small, unelected government that relies extensively on algorithms for social engineering.

In 2012, Zuckerberg addressed future Facebook investors in a letter: “People sharing more — even if just with their close friends or families — creates a more open culture and leads to a better understanding of the lives and perspectives of others. We believe that this . . . helps people get exposed to a greater number of diverse perspectives. By helping people form these connections, we hope to rewire the way people spread and consume information.”

Whatever those beliefs were based on, they have largely failed the test of time. Instead of creating stronger relationships, Facebook has spawned anxieties and addictions that are the subject of academic studies from Portugal to Australia. Some studies have determined that using Facebook detracts from a user’s life satisfaction.

A Danish experiment in 2015, involving people weaned from Facebook for a week and a control group that kept using it, showed that people on the social network are 55 percent more likely to feel stressed; one source of that stress is envy of the glossified lives reported by other users. Users’ well-being, research has showed, only tends to increase when they have meaningful interactions — such as long message exchanges — with those who are already close to them.

In his latest manifesto, Zuckerberg uses parenting groups as an example of something his company does right. But recent research shows that some new mothers use Facebook to obtain validation of their self-perception as good parents, and failing to get enough such validation causes depressive symptoms.

As for the “rewired” information infrastructure, it has helped to chase people into ideological silos and feed them content that reinforces confirmation biases. Facebook actively created the silos by fine-tuning the algorithm that lies at its center — the one that forms a user’s news feed.

The algorithm prioritizes what it shows a user based, in large measure, on how many times the user has recently interacted with the poster and on the number of “likes” and comments the post has garnered.

In other words, it stresses the most emotionally engaging posts from the people to whom you are drawn — during an election campaign, a recipe for a filter bubble and, what’s more, for amplifying emotional rather than rational arguments.

Zuckerberg writes: “In recent campaigns around the world — from India and Indonesia across Europe to the United States — we’ve seen the candidate with the largest and most engaged following on Facebook usually wins.”

In the Netherlands today, liberal Prime Minister Mark Rutte’s page has 17,527 likes; that of fiery nationalist Geert Wilders, 174,188. In France, rationalist Emmanuel Macron has 165,850 likes, while far-right Marine Le Pen boasts 1.2 million. Helping them win is hardly something that would make Zuckerberg, a liberal, proud.

Zuckerberg doesn’t want to correct this mistake and stop messing with what people see. Instead, the manifesto describes how Facebook sorts groups into “meaningful” and, presumably, meaningless ones. Instead of facilitating communication among people who are already part of social support groups offline, he wants to project Facebook relationships into the real world: Clearly, that’s a more effective way of keeping competitors at bay.

The Facebook chief says his team is working on artificial intelligence that will be able to flag posts containing offensive information — nudity, violence, hate speech — and pass them on for final decisions by humans. If past experience is any indication, the overtaxed humans will merely rubber-stamp most decisions made by the technology, which Zuckerberg admits is still highly imperfect.

Zuckerberg casts Facebook as a global community that needs better policing, governance, nudging toward better social practices. He’s willing to allow some democracy and “referendums,” but the company will make the ultimate decision on the types of content people should see based on their behavior on Facebook.

Ultimately, this kind of social engineering affects people’s moods and behaviors. It can drive them toward commercial interactions or stimulate giving to good causes but it can also spill out into the real world in more troubling ways.

It’s absurd to expect humility from Silicon Valley heroes. But Zuckerberg should realize that by trying to shape how people use Facebook, he may be creating a monster.

© 2017, Bloomberg View

Our editors found this article on this site using Google and regenerated it for our readers.

NEXT NEWS