How to Fix Facebook

Total
0
Shares

This article is part of the On Tech newsletter. Here is a collection of past columns.[1]

This is the most important moment in the history of Facebook. Hyperbole, perhaps, but only a little.

A former product manager at Facebook, Frances Haugen, captivated U.S. senators at a hearing on Tuesday[2] with a nuanced diagnosis that the company needs to be saved from itself — for the good of all of us.

What felt different than Facebook’s 4 million previous scandals and congressional scoldings was Haugen’s focus on what she sees as the company’s foundational flaws of technical designs and corporate organization, and the messy but sophisticated discussions happening outside Facebook to improve the company.

Haugen said that Facebook stretched itself too thin to effectively confront harms like ethnic violence and human trafficking[3] that had been tied to activity on its apps. She dissected the ways that Facebook’s fixation on getting us to spend more time online aggravated our worst impulses. And she hammered the message that the public shouldn’t be kept in the dark about what Facebook knew about its influence on us and our world.

The picture that emerged from recent Wall Street Journal reporting[4] and Haugen’s media interviews[5] was not of Facebook as a cartoonish James Bond villain. It was of a company that can’t control the machines that it built, but refuses to accept that reality.

“Facebook is stuck in a feedback loop that they can’t get out of,” Haugen told senators.

Some of what Haugen and Facebook critics have said about the company is probably overstated. And a lot of what Haugen said wasn’t new. But she is a laser-focused messenger at a time when people in power are ready to stop bickering and ask: What now? What should be done to maximize the good of Facebook and minimize the harm?

There are no magic fixes, but Haugen and many others have offered sound suggestions on what to try.

The most compelling idea from Haugen was that “engagement-based ranking[6]” is an original sin of Facebook, YouTube, TikTok, Pinterest and other popular apps. When computers prioritize what we see online based on what is likely to captivate us and keep us around longer, they tend to fan the most salacious or extreme views, and subtly nudge people to post more of the same.

Haugen suggested, essentially, turning off the computer algorithms and making more of the internet gravitate toward designs like those of iMessage or past versions of Facebook and Instagram that showed posts in chronological order.

Kate Klonick, who has researched policies on online expression at internet companies, wrote[7] in The New York Times that Facebook could redesign its websites to optimize holistic measures of the good things that it offers. Rather than focusing on metrics such as which posts are likely to get a ton of shares or likes, it could look at what is likely to lead you to attend a protest or give to a charitable cause.

Haugen and others[8] have recommended changing U.S. law to hold Facebook responsible for real-world harms, including terrorist acts[9], resulting from posts that the company’s computer systems distributed to people’s feeds.

In a recent interview[10], Haugen also mentioned the idea of public representatives to oversee Facebook from the inside, similar to Federal Reserve examiners for large banks. She also backed the idea of regulations to force Facebook to work with researchers[11] who want to study the company’s effects on users.

And Haugen suggested that many of Facebook’s worst moments, including its social network being used to fan ethnic violence, may be the result of having too few people to manage its ambitions. Should Facebook be forced to do less, like quitting countries unless the company devotes more resources to them and establishes cultural competence[12]?

There are plenty of reasons to feel pessimistic. Facebook essentially told[13] Congress — “YOU tell us what to do.” Yet, U.S. lawmakers and regulators have done little to tell Facebook how to better govern apps used by billions of humans.

Facebook has said[14], correctly, that it strives to continually improve its apps and that doing so is a tricky exercise in trade-offs. Mark Zuckerberg on Tuesday rejected[15] the (oversimplified) notion that his company chooses profits over people’s lives and well-being, and that the company ignores ideas for improvement.

Maybe none of the ideas tossed around to fix Facebook will be better than the status quo. But what felt fresh from Haugen was a message of hope: We need the best of Facebook, and we must work together to make it better.

More reading and listening:

References

  1. ^ past columns (www.nytimes.com)
  2. ^ at a hearing on Tuesday (www.nytimes.com)
  3. ^ ethnic violence and human trafficking (www.wsj.com)
  4. ^ Wall Street Journal reporting (www.wsj.com)
  5. ^ media interviews (www.nytimes.com)
  6. ^ engagement-based ranking (www.nytimes.com)
  7. ^ wrote (www.nytimes.com)
  8. ^ others (www.nytimes.com)
  9. ^ terrorist acts (www.nytimes.com)
  10. ^ a recent interview (www.wsj.com)
  11. ^ regulations to force Facebook to work with researchers (www.washingtonpost.com)
  12. ^ company devotes more resources to them and establishes cultural competence (www.nytimes.com)
  13. ^ told (twitter.com)
  14. ^ has said (about.fb.com)
  15. ^ rejected (www.nytimes.com)
Leave a Reply

Your email address will not be published. Required fields are marked *

You May Also Like