Just the other day I asked AI to create an image showing a diverse group of people, it didn’t include a single black person. I asked it to rectify and it couldn’t do it. This went on a number of times before I gave up. There’s still a long way to go.
Inherent bias is going to get worse and worse if we let AI roam free.
I am instead thinking it will instead not be the case? Bigger models will be able to store more of the less common realities
They will, at best, replicate the data sets. They will learn racial discrimination and propagate it.
If you have a deterministic system, for example, to rate a CV, you can ensure that no obvious negative racial bias is included. If instead you have a LLM (or other AI) there is no supervision on which data element is used and how. The only thing we can check is if the predictions match the (potentially racist) data.
You may be able to prompt for the less common realities, but the default of the model is still going to see “doctor” as a white man.