AI-infused hiring programs have drawn scrutiny, most notably over whether they end up exhibiting biases based on the data they’re trained on.
AI-infused hiring programs have drawn scrutiny, most notably over whether they end up exhibiting biases based on the data they’re trained on.
Isn’t the whole point of AI decision making to provide plausible deniability for these sort of things?
Depends how the law is applied…
Kinda like if a self driving car kills someone, who is liable, driver, manufacturer, seller?
I guess you pay insurance and they take on liability is another option.