The biggest issues of bias/fairness in ML are not to do with the algorithms or results, but the underlying data.<p>A trivial example would be: what if you trained a classifier to predict whether a person would be re-arrested before they went to trial? Some communities are policed more heavily so you would tend towards reinforcing the bias that exists and provide more ammunition to those arguing for further bias in the system, a feedback loop if you would.<p>Or what if some protected group needs a higher down payment because the group is not well understood enough so that you can't distinguish between those who will repay your loans and who won't? Maybe educational achievement is a really good predictor on one group, but less effective on another. Is it fair to use the protected class (or any information correlated with it) when it is essentially machine-enabled stereotyping?<p>Recently it has been noted that NLP systems trained on large corpuses of text tend to exhibit society's biases where they assume that nurses are women and programmers are men. From a statistical perspective this correlation is there, but we tend to be more careful about how we use this information than a machine. We wouldn't want to use this information to constrain our search for people to hire to just those that fulfil our stereotypes, but a machine would. This paper has some details on such issues: <a href="http://arxiv.org/abs/1606.06121" rel="nofollow">http://arxiv.org/abs/1606.06121</a><p>I don't think there are any easy solutions here, but I think it's important to be aware that data is only a proxy for reality and fitting the data perfectly doesn't mean you have achieved fair outcomes.
Another recent paper on this topic: <a href="http://arxiv.org/pdf/1606.08813v3.pdf" rel="nofollow">http://arxiv.org/pdf/1606.08813v3.pdf</a>. It shows how naive lending algorithms can skew against minority groups simply because there is less data available about them, even if their expected repayment rate is the same.<p>It can be self-reinforcing. Imagine some new demographic group of customers appears, and without any data you make some loans to them. The actual repayment rate will be low, not because that group has a worse distribution than other groups, but simply because you couldn't identify the lowest-risk members. A simplistic ML model would conclude that the new group is more risky.<p>Of course, smart lenders understand that in order to develop a new customer demographic they need to experiment by lending, with the expectation that their first loans will have high losses, but that in the long run learning about how to identify the low-risk people from that demographic is worthwhile. And they correct for the fact that the first cohort was accepted blind when estimating overall risk for the group.
What is fairness but political accountability?<p>There is an old joke about how people use statistics like a drunk uses a lamp post: for support and not for illumination. Given this, we can expect people to use AI like everything else in statistics, to support the agenda of whoever is operating it while defraying negative personal accountability for the results, because artificial intelligence. It's just an obfuscated and sophisticated version of, "Computer says no."<p>The alternative is the near future headline, "AI confirms racists, sexists, on to something."
Everyone suggesting that we ought to legislate that machines must be illogical/suboptimal is missing the point.<p>If machine learning algorithms are unfairly discriminating against some group, then they are making sub-optimal decisions and costing their users money. This is a self-righting problem.<p>However, a good machine learning algorithm may uncover statistical relationships <i>that people don't like</i>; for example, perhaps some nationalities have higher loan repayment rates. In these cases, the algorithm is not at odds with reality; the angsty humans are. If some people want to force machines to be irrational, they should at least be honest about their motivations and stop pretending it has a thing to do with "fairness".
After studying this issue, and learning a lot more about learning and optimization, I've come to the conclusion that the best solution [1] is probably explicit racial/sexual/other special interest group quotas.<p>Specifically, we should train a classifier on non-Asian minorities. We should train a different classifier on everyone else. Then we should fill our quotas from the non-Asian minority pool and draw from the primary pool for the rest of the students.<p>As this blog post describes, no matter what you do you'll reduce accuracy. But every other fairness method I've seen reduces accuracy both <i>across</i> special interest groups and also <i>within</i> them. Quotas at least give you the best non-Asian minorities and also the best white/Asian students.<p>Quotas also have the benefit of being simple and transparent - any average Joe can figure out exactly what "fair" means, and it's also pretty transparent that some groups won't perform as well as others and why. In contrast, most of the more complex solutions obscure this fact.<p>[1] Here "best" is within the framework of requiring a corporatist spoils system. I don't actually favor such a system, but I'm taking the existence of such a spoils system as given.
Either you allow an algorithm to be ruthlessly fair, or you introduce bias and never get the problem solved correctly, because someone, somewhere, will still find a way to gripe about the amount of bias when, inevitably, it goes against them, or is perceived to be against them due to lack of knowledge. Then you wind up bikeshedding over the bias and not the actual problem.
I am actually optimistic on Big Data's effect in equality.<p>Small data is actually kind of the problem. When you have limited ability to process data or limited data density then your segmentation ability is limited to small data like state, county, zip code, credit score, whether you own a home, etc.<p>Big data processing, big bad ML algorithms and the ubiquity of data is making advanced segmentation available that allows us to make arguably more equitable outcomes.
> As a result, the advertiser might have a much better understanding of who to target in the majority group, while essentially random guessing within the minority.<p>If this is the case, then it should be detected and ML should NOT be used for the minority class. There are many classifiers out there which work on one-class problems.