TE
TechEcho
Home24h TopNewestBestAskShowJobs
GitHubTwitter
Home

TechEcho

A tech news platform built with Next.js, providing global tech news and discussions.

GitHubTwitter

Home

HomeNewestBestAskShowJobs

Resources

HackerNews APIOriginal HackerNewsNext.js

© 2025 TechEcho. All rights reserved.

Bloomberg's analysis didn't show that ChatGPT is racist

66 pointsby leenyabout 1 year ago

9 comments

cjk2about 1 year ago
Fairly obvious. Is a parrot racist because it heard someone being racist and repeats it without being able to reason about it?<p>It lacks intent and understanding so it can&#x27;t be racist. It might make racist sounding noises though.<p>A fine example ... <a href="https:&#x2F;&#x2F;www.youtube.com&#x2F;watch?v=2hUS73VbyOE" rel="nofollow">https:&#x2F;&#x2F;www.youtube.com&#x2F;watch?v=2hUS73VbyOE</a>
评论 #40061569 未加载
评论 #40058801 未加载
SpaceManNabsabout 1 year ago
This article does the same stat 101 mistakes that the Bloomberg article does with p-values.<p>All this article can say it is that it cannot reject the null hypothesis (chatgpt does not produce statistical discrepancies).<p>It certainly cannot state that chatgpt is definitively not racist. The article moves the discussion in the right direction though.<p>Also, I didn&#x27;t look too closely, but their table under &quot;Where the Bloomberg study went wrong&quot; has unreasonable expected frequencies. But then I noticed it was because it was measuring &quot;name-based discrimination.&quot; This is a terrible proxy to determine racism in the resume review process, but that is what Bloomberg decided on so wtv lol. Not faulting the article for this, but this discussion seems to be focused on the wrong metric.<p>If you are going to argue people over stats, then don&#x27;t make the same mistakes...
评论 #40056855 未加载
observationistabout 1 year ago
Any naive use of an LLM is not likely to produce good results, even with the best models. You need a process - a sequence of steps, and appropriately safeguarded prompts at each step. AI will eventually reach a point when you can get all the subtle nuance and quality in task performance you might desire, but right now, you have to dumb things down and be very explicit. Assumptions will bite you in the ass.<p>Naive, superficial one shot prompting, even with CoT or other clever techniques, or using big context, is insufficient to achieve quality, predictable results.<p>Dropping the resume into a prompt with few-shot examples can get you a little consistency, but what really needs to be done is repeated discrete operations, that link the relevant information to the relevant decisions. You&#x27;d want to do something like tracking years of experience, age, work history, certifications, and so on, completely discarding any information not specifically relevant to the decision of whether to proceed in the hiring process. Once you have that information separated out, you consider each in isolation, scoring from 1 to 10, with a short justification for each scoring based on many-shot examples. Then you build a process iteratively with the bot, asking it which variables should be considered in context of the others, and incorporate a -5 to 5 modifier based on each clustering of variables (8 companies in the last 2 years might be a significant negative score, but maybe there&#x27;s an interesting success story involved, so you hold off on scoring until after the interview.)<p>And so on, down the line, through the whole hiring process. Any time a judgment or decision has to be made, break it down into component parts, and process each of the parts with their own prompts and processes, until you have a cohesive whole, any part of which you can interrogate and inspect for justifiable reasoning.<p>The output can then be handled by a human, adjusted where it might be reasonable to do so, and you avoid the endless maze of mode collapse pits and hallucinated dragons.<p>LLMs are not minds - they&#x27;re incapable of acting like minds, unless you build a mind-like process around them. If you want a reasonable, rational, coherent, explainable process, you can&#x27;t achieve that with zero or one shot prompting. Complex and impactful decisions like hiring and resume processing isn&#x27;t a task current models are equipped to handle naively.
评论 #40056933 未加载
评论 #40057592 未加载
评论 #40057201 未加载
评论 #40057390 未加载
评论 #40070668 未加载
benaabout 1 year ago
As someone who read enough of the article before it became a full-blown ad for their services: neat.<p>They do have a point with regards to Bloomberg&#x27;s analysis.<p>Bloomberg&#x27;s analysis have white women being selected more often than all other groups for software developers, with the exception of hispanic women.<p>That&#x27;s a little weird. More often than not, when something is sexist or racist, it&#x27;s going to favor white men. But then you also see that the differences are all less than 2% from the expectation. Nothing super major and well within the bounds of &quot;sufficiently random&quot;.<p>Now, I also wouldn&#x27;t make the claim that ChatGPT isn&#x27;t racist based on this either. It&#x27;s fair to say that ChatGPT did not exhibit a racial preference in this task.<p>The best you can say is that the study says nothing.<p>What they should do is basically poison the well. Go in with predetermined answers. Give it 7 horrible resumes and 1 acceptable. It should favor the acceptable resume. You can also reverse it with 7 acceptable resumes and 1 horrible resume. It should hardly ever pick the loser. That way you can test if ChatGPT is even attempting to evaluate the resumes or is just picking one out of the group at random.
fwipabout 1 year ago
I hate headlines&#x2F;framings like this.<p>&gt; It’s convention that you want your p-value to be less than 0.05 to declare something statistically significant – in this case, that would mean less than 5% chance that the results were due to randomness. This p-value of 0.2442 is way higher than that.<p>You can&#x27;t get &quot;ChatGPT isn&#x27;t racist&quot; out of that. You can only get &quot;this study has not conclusively demonstrated that ChatGPT is racist&quot; (for the category in question).<p>And in fact, in half of the categories, ChatGPT3.5 does show very strong evidence of racism &#x2F; racial bias (p-value below 1e-4).
评论 #40056882 未加载
评论 #40055832 未加载
评论 #40055850 未加载
评论 #40055874 未加载
评论 #40055840 未加载
up2isomorphismabout 1 year ago
Trying to say a car is a murderer does not make sense. ChatGPT is a symbol generator, with local high probability of resembling to a person, so it is not a person, how can it be a racist?
Animatsabout 1 year ago
The big result is that ChatGPT is terrible at resume evaluation. Only slightly better than random.
评论 #40057790 未加载
gurumeditationsabout 1 year ago
In my experience image generators are anti-gay and refuse to create images featuring gay people many times.
tmoravecabout 1 year ago
If Bloomberg calculated the p-value, they couldn&#x27;t write a catchy article. It&#x27;s a conspiracy theory of course but this omission seems too big for a simple oversight.