Despite advancements in AI, new research reveals that large language models continue to perpetuate harmful racial biases, particularly against speakers of African American English.
We’re like teenaged trailer trash parents who just gave birth to a genius at the trailer park where we’re all dysfunctional alcoholics and meth addicts …
… now we’re acting surprised that our genius baby talks like an idiot after listening to us for ten years.
While it may be obvious to you, most people don’t have the data literacy to understand this, let alone use this information to decide where it can/should be implemented and how to counteract the baked in bias. Unfortunately, as is mentioned in the article, people believe the problem is going away when it is not.
“Wow Johnson, no matter how much biased data we feed this thing it just keeps repeating biases from human society.”
Sample input from a systematically racist society (the entire world), get systematically racist output.
No shit. Fix society or “tune” your model, whatever that entails…
Obviously only one of these is feasible from a developer perspective.
This is thing I keep pointing out about AI
We’re like teenaged trailer trash parents who just gave birth to a genius at the trailer park where we’re all dysfunctional alcoholics and meth addicts …
… now we’re acting surprised that our genius baby talks like an idiot after listening to us for ten years.
While it may be obvious to you, most people don’t have the data literacy to understand this, let alone use this information to decide where it can/should be implemented and how to counteract the baked in bias. Unfortunately, as is mentioned in the article, people believe the problem is going away when it is not.
.