A tester’s thoughts on Automation and AI part 3

A tester’s thoughts on Automation and AI part 3

A lot has happened in the testing field when it comes to automation and AI. Therefore our test expert Eva Holmquist has written a blog series on the theme where she formulates her thoughts. This is part number three about if AI is unbiased or not.

There is a common belief that AI makes unbiased decisions. In reality, the kind of narrow artificial intelligence that exists today is far from unbiased, see for instance Richards Fall’s articles, When AI “goes bad”  and Algorithms and Bias in the Criminal Justice System which talks about this very issue. You also have a great article by Rahul Bhargava that talks about the need to shift focus from the learning to the teaching aspect of machine learning. Because that’s the thing, no machine is learning in a vacuum. They are learning in our world, that’s full of bias. It’s also humans that choose which training data to use and what criteria to use for decisions. There is a lot of examples of bias in systems with learning capabilities that are evident, for example, racist twitter bots, recruitment systems only choosing male applicants and systems that predict black defendants will have higher risks of recidivism than they actually do. I wonder how many biased systems there are that we haven’t discovered…

This means that when we test systems with learning capabilities, it’s important that we test for these aspects as well. Is the system behaving with unaccepting prejudice? And because it continues to learn, we also need to monitor to catch biased behaviour. How do you suggest we do this?

And if you want to discuss the subject further, don't be afraid to reach out to me.

Food for thought
  • Eva Holmquist
    Eva Holmquist
    Senior Test Specialist
    072-502 83 93