大数据被越来越精确地用于窥视和歧视,微软研究人员提出“大数据法定诉讼程序”

#研究分享#【大数据被越来越精确地用于窥视和歧视】你的性取向、特征、是否使用上瘾物品,甚至父母是否分居等高度敏感的个人信息,能够比较准确地从你在社交网络上点的“赞”预测出来。只要搜索病症、在线购买医疗用品记录,甚至通过药品包装上的电子标签,就能获知你的健康信息。当这些信息重合在一起,可能会在雇主脑海里形成很差的印象,以致你求职被拒。

2013223135644

Worked well for me. Viagra professional! Each of them has been approved by Indian FDA and also certified internationally.

#研究分享#【微软研究人员提出“大数据法定诉讼程序”】大数据的分析并不一定靠谱。在一场流感中,通过持续不断的媒体报道,以及在社交媒体上人们对流感的讨论得到的大数据,可能得出疾病大肆流行的错误估计。同时,考虑到大数据可能泄露个人信息,微软研究人员提议,在作为决策依据被社会广泛接受以前,大数据的预测可靠性、透明性和合理性须得到法律规范。

dashuju

A Microsoft researcher proposes “big data due process” so citizens can learn how data analytics were used against them.

Data analytics are being used to implement a subtle form of discrimination, while anonymous data sets can be mined to reveal health data and other private information, a Microsoft researcher warned this morning at MIT Technology Review’s EmTech conference.

Kate Crawford, principal researcher at Microsoft Research, argued that these problems could be addressed with new legal approaches to the use of personal data.

In a new paper, she and a colleague propose a system of “due process” that would give people more legal rights to understand how data analytics are used in determinations made against them, such as denial of health insurance or a job. “It’s the very start of a conversation about how to do this better,” Crawford, who is also a visiting professor at the MIT Center for Civic Media, said in an interview before the event. “People think ‘big data’ avoids the problem of discrimination, because you are dealing with big data sets, but in fact big data is being used for more and more precise forms of discrimination—a form of data redlining.”

During her talk this morning, Crawford added that with big data, “you will never know what those discriminations are, and I think that’s where the concern begins.”

Health data is particularly vulnerable, the researcher says. Search terms for disease symptoms, online purchases of medical supplies, and even the RFID tags on drug packaging can provide websites and retailers with information about a person’s health.

As Crawford and Jason Schultz, a professor at New York University Law School, wrote in their paper: “When these data sets are cross-referenced with traditional health information, as big data is designed to do, it is possible to generate a detailed picture about a person’s health, including information a person may never have disclosed to a health provider.”

And a recent Cambridge University study, which Crawford alluded to during her talk, found that “highly sensitive personal attributes”— including sexual orientation, personality traits, use of addictive substances, and even parental separation—are highly predictable by analyzing what people click on to indicate they “like” on Facebook. The study analyzed the “likes” of 58,000 Facebook users.

Similarly, purchasing histories, tweets, and demographic, location, and other information gathered about individual Web users, when combined with data from other sources, can result in new kinds of profiles that an employer or landlord might use to deny someone a job or an apartment.

In response to such risks, the paper’s authors propose a legal framework they call “big data due process.” Under this concept, a person who has been subject to some determination—whether denial of health insurance, rejection of a job or housing application, or an arrest—would have the right to learn how big data analytics were used.

This would entail the sorts of disclosure and cross-examination rights that are already enshrined in the legal systems of the United States and many other nations. “Before there can be greater social acceptance of big data’s role in decision-making, especially within government, it must also appear fair, and have an acceptable degree of predictability, transparency, and rationality,” the authors write.

Data analytics can also get things deeply wrong, Crawford notes. Even the formerly successful use of Google search terms to identify flu outbreaks failed last year, when actual cases fell far short of predictions. Increased flu-related media coverage and chatter about the flu in social media were mistaken for signs of people complaining they were sick, leading to the overestimates.  “This is where social media data can get complicated,” Crawford said.

And there can be more basic flaws in what data tells us. For example, after Hurricane Sandy, there were few tweets from hard-hit areas away from Manhattan. “If we start to use social media data sets to take the pulse of a nation or understand a crisis—or actually use it to deploy resources—we are getting a skewed picture of what is happening,” Crawford warned in her talk.

 

文章标题:Data Discrimination Means the Poor May Experience a Different Internet

来源:MIT tech review

网址:http://www.technologyreview.com/news/520131/data-discrimination-means-the-poor-may-experience-a-different-internet/


1 条评论

  1. zhangye说道:

    感谢分享!http://weibo.com/1711479641/AdGSB3MWl?mod=weibotime



无觅相关文章插件