this post was submitted on 23 Sep 2024
61 points (100.0% liked)
Technology
37717 readers
416 users here now
A nice place to discuss rumors, happenings, innovations, and challenges in the technology sphere. We also welcome discussions on the intersections of technology and society. If it’s technological news or discussion of technology, it probably belongs here.
Remember the overriding ethos on Beehaw: Be(e) Nice. Each user you encounter here is a person, and should be treated with kindness (even if they’re wrong, or use a Linux distro you don’t like). Personal attacks will not be tolerated.
Subcommunities on Beehaw:
This community's icon was made by Aaron Schneider, under the CC-BY-NC-SA 4.0 license.
founded 2 years ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
view the rest of the comments
"Wow Johnson, no matter how much biased data we feed this thing it just keeps repeating biases from human society."
Sample input from a systematically racist society (the entire world), get systematically racist output.
No shit. Fix society or "tune" your model, whatever that entails..
Obviously only one of these is feasible from a developer perspective.
While it may be obvious to you, most people don't have the data literacy to understand this, let alone use this information to decide where it can/should be implemented and how to counteract the baked in bias. Unfortunately, as is mentioned in the article, people believe the problem is going away when it is not.
The real problem are implicit biases. Like the kind of discrimination that a reasonable user of a system can't even see. How are you supposed to know, that applicants from "bad" neighborhoods are rejected at a higher rate, if the system is presented to you as objective? And since AI models don't really explain how they got to a solution, you can't even audit them.
I have a feeling that's the point with a lot of their use cases, like RealPage.
It's not a criminal act when an AI did it! (Except it is and should be.)
"It's not redlining when an algorithm does it!"
.
In my mind, this is the whole purpose of regulation. A strong governing body can put in restrictions to ensure people follow the relevant standards. Environmental protection agencies, for example, help ensure that people who understand waste are involved in corporate production processes. Regulation around AI implementation and transparency could enforce that people think about these or that it at the very least goes through a proper review process. Think international review boards for academic studies, but applied to the implementation or design of AI.
AI ethics is a field which very much exists- there are plenty of ways to measure and define how racist or biased a model is. The comparison groups are typically other demographics.... such as in this article, where they compare AAE to standard English.
This is thing I keep pointing out about AI
We're like teenaged trailer trash parents who just gave birth to a genius at the trailer park where we're all dysfunctional alcoholics and meth addicts ...
... now we're acting surprised that our genius baby talks like an idiot after listening to us for ten years.