Facebook Mocks its Users with New Research Policy
In any modern, first-world country, the government requires legitimate university researchers to go through an independent review board (IRB) when conducting research on human beings. This is due to past abuses by both governments and organizations who have used the guise of “seeking knowledge” to cover up their efforts to manipulate people for their own means and ends.
But you know what? Facebook isn’t a first-world country. So in an effort to better understand how to best monetize your use of their service, they don’t need an IRB’s approval to conduct research on you.
And now, in my opinion, they are outright mocking their users with their latest update to their research practices.
The latest controversy started when Facebook “data scientist” researcher Adam D.I. Kramer published a study that manipulated people’s Facebook newsfeeds without their knowledge or consent. Kramer issued a non-apology after the incident, and then what followed was three months of total silence from Facebook on the issue.
Today, Facebook finally broke their silence and wrote this blog entry about their updated research policies.
Nowhere in the entry is the word “ethics” mentioned. This alone demonstrates that Facebook still doesn’t understand what it did wrong with previous research and has little insight into how social science research is completely different — for a reason — from computer science. Facebook still has no ethics officer.
So what did Facebook actually change? Will new research have to go through an IRB (or whatever an equivalent would be in a for-profit company looking to maximize the monetization of its users)?
As it turns out, not much.
Now, some research that is subjectively determined that may need “additional review” will undergo… additional review. By others at Facebook.
I guess nobody at Facebook has heard the phrase, “conflict of interest”?
Well, at least I’m sure they’re now going to require informed consent of its users to participate in human subject experiments, right?
Nope. Facebook knows that few would willingly give their consent to be manipulated by their algorithms and at the whims of their “data scientists,” so they still won’t bother to ask you for consent.
So what will? Facebook change? Well, apparently, they’ve added “clearer guidelines” because apparently Facebook had few (or unclear?) guidelines previously. Apparently anyone at Facebook could do anything they wanted research-wise, manipulating its users’ experiences whenever and however they wanted. They’ve also added “training,” because, again, apparently Facebook let people with no or little training in human subjects research to conduct experiments on its users. That means you, you poor unsuspecting datapoint.
The new internal review panel has a bunch of people on it, but again, it’s not clear anyone there is actually familiar or has a background in social sciences research and ethics. Since no mention of ethics was found, it’s likely Facebook still doesn’t quite “get” what it did wrong. At least that’s what I take away from this wonderful marketing-speak blog entry.
And you have to love this last swipe at every Facebook user today:
Like most companies today, our products are built based on extensive research, experimentation and testing. […]
We want to do this research in a way that honors the trust you put in us by using Facebook every day.
What they don’t mention is that most companies today who build their products on “extensive research, experimentation and testing” do so under explicit test conditions clearly communicated to its users.
Do you think P&G goes out and changes the formulation of their most popular products on unsuspecting consumers without getting their consent first? Do you think GM or Ford tests changes to their vehicle’s suspension or ergonomics on people who’ve just plunked down $30,000 for one of their cars?? Do you think that Target changes all of its pricing just for you when you walk into the store, just to see how you might react??
The point being, companies do indeed conduct testing and research on their products — but only with the customer’s explicit and informed consent. They don’t do it just to test hypotheses at random.
Nothing in Facebook’s new research policies asks users for their explicit, informed consent to participate in a human subject experiment. And there remains no independent review of their human subject research — nothing even close to a university IRB.
In sum, the new Facebook research policy mocks its users. You, apparently, are nothing more than a marketing datapoint to Facebook, not deserving of the same informed consent any normal research participant would be entitled to. Facebook has made it clear, at least to me, that they are not interested in becoming a legitimate member of the research community. Instead, they’re appear to be happy to remain “data scientists” — and you are simply the data.1
Read the new research policy: Research at Facebook
- And naturally, there’s nowhere to comment about this blog entry on Facebook. For a social networking company, Facebook ironically appears to dislike any actual social networking with itself. [↩]
Grohol, J. (2018). Facebook Mocks its Users with New Research Policy. Psych Central. Retrieved on August 12, 2020, from https://psychcentral.com/blog/facebook-mocks-its-users-with-new-research-policy/