I just finished listening to +TWiT from this weekend with +Leo Laporte, +Natali Morris, +Tim Stevens, and Denise Howell.  I don't find myself saying this about TWiT that often, but I disagree with the panel's thoughts on the recent revelations about Facebook's experimentation on their users news feeds.  (http://www.contriving.net/link/fr)
So many people seem to be in such a tizzy over this, but I just don't get it.
The very nature of human interaction is to influence others. Â In today's world, we influence other people's emotions via advertisements. Â We influence other people's emotions via "addictive" video games like Farmville or Simpson's Tapped Out. Â We influence other people's emotions via how we choose to dress. Â We influence other people's emotions via the type of car we drive. Â All of this is intentional.
The more organized examples of influencing other people's emotions is done for a profit! Â Heck, Zynga employs researchers whose sole purpose is to game your emotions and get you to spend more money. Â The same holds true for advertising agencies, corporations, and political campaigns.
On the show +Tim Stevens brings up the point that websites do A/B testing all of the time to see what kind of content or presentation causes more user engagement.  That is exactly the same sort of thing.  Emotions drive user engagement.  *Websites manipulate your emotions to make more money.*  This is normal.  This is part of being human.
When you talk to the lady at the DMV you slap a smile on your face so you can manipulate her emotions in your favor so she'll hurry you through the process.  When you take your date to a fancy restaurant, you're manipulating their emotions for  some desired outcome. Â
The majority of the time you're consciously or unconsciously partaking in a behavior to manipulate the emotions of those around you in your favor.
Facebook's experiment isn't any worse than any of these. Â In fact, if you want to make a value judgement, it's probably the least "bad" out of them all, as the results of their experiment were published.
Finally, Leo brings up the point that the study was stupid because the results were obvious. Â I have several things to say about that. Â
First of all, I don't think the results are completely obvious. Â Yes, I would have predicted the results they got with a fair amount of confidence. However, it's not out of the realm of possibility that people just don't care about their news feeds enough to let it affect them, or to see the opposite happen…namely for people to post more positive stories in "retaliation" against negative news feeds.
Secondly, you could extract a lot of nuance out of such a study. Â Maybe, for each of the types of responses I detail in the last paragraph, there are groups of people who consistently respond in each manner. Â That would be an interesting and useful result.
Lastly, just because you think the results of an experiment are obvious, doesn't mean the experiment is not worth doing. Â You don't know for sure the results until you do them, and how wonderful is it when an experiment dis-confirms your beliefs? Â Basically, using "obvious result" as the determining factor in choosing whether to do an experiment is a bad idea.
I'm no fan of Facebook. Â I think it's a time wasting cesspool out of which most can extract very minimal amounts of value, and I only visit it a few times a year. Â I just can't get excited about this relatively (if not absolutely) small experiment they did amidst the vast and turbulent sea of emotional manipulation we live in…particularly when the most pernicious examples are for profit and we accept those with barely a peep.
A lot of people have some sort of intuitive, it-just-feels-wrong, response to this story and I just don't get it.
Google+: View post on Google+