Disclaimer: Worth noting that this paper is apparently pre-print [1] and was submitted to the Lancet, who are not particularly reliable in recent times (flaws hydroxychloroquine study [2], stamped out early investigation in COVID lab-leak possibility [3]).<p>Regarding the blog and paper:<p>> AI can trivially learn to identify the self-reported racial identity of patients to an absurdly high degree of accuracy<p>If it is 'trivial' for a model to learn race, then what they are detecting is likely pretty obvious.<p>In their paper they use 'private and public datasets' [4]. There is probably exactly where the problem occurs. People of the similar race tend to live in groups together, if you collect data in some area, there's a good chance you source data from the same ethnic pool. What it appears they are detecting is the difference between different machines and practices. To test this, they should collect their own medical data on the same machine. They also need to figure out how to remove the bias of the medical staff themselves.<p>Another assumption is that just because the medical doctors cannot see differences in the anatomy of races, it doesn't mean there isn't one. One of the commentators on this blog indicated that dark skin may affect the absorption of x-rays, for example.<p>> AI does learn to do this when trained for clinical tasks<p>Just because AI models <i>can</i> detect race from medical data, doesn't mean they <i>will</i>. They may in the early stages learn such simplistic relationships, but as accuracy increases, it will be forced to look for more complex classifiers/indicators.<p>Finally, this blog really skirts around the important thing - how does this affect patient care? Is it beneficial to actually have race-based medical plans? We know that some medicines affect different races - for example in pain and pain management [5].<p>I think the point to be made here isn't 'racism' but the <i>possibility</i> of bias that people training models should be aware of. Some of this bias will have a positive outcome, some will have a negative outcome.<p>[1] <a href="https://www.bibsonomy.org/bibtex/c3e2ede2f835d35c08c54a670951276b#citation_BibTeX" rel="nofollow">https://www.bibsonomy.org/bibtex/c3e2ede2f835d35c08c54a67095...</a><p>[2] <a href="https://www.nbcnews.com/health/health-news/lancet-retracts-large-study-hydroxychloroquine-n1225091" rel="nofollow">https://www.nbcnews.com/health/health-news/lancet-retracts-l...</a><p>[3] <a href="https://www.washingtonexaminer.com/news/wuhan-lab-collaborator-peter-daszak-recused-from-lancets-covid-19-origins-investigation" rel="nofollow">https://www.washingtonexaminer.com/news/wuhan-lab-collaborat...</a><p>[4] <a href="https://arxiv.org/pdf/2107.10356.pdf" rel="nofollow">https://arxiv.org/pdf/2107.10356.pdf</a><p>[5] <a href="https://www.ncbi.nlm.nih.gov/pmc/articles/PMC3654683/" rel="nofollow">https://www.ncbi.nlm.nih.gov/pmc/articles/PMC3654683/</a>