Ensuring Artificial Intelligence is Ethical? That’s Everyone’s Responsibility

Artificial intelligence has evolved from just being a concept and has now become a serious topic in society. In fact, some experts think that machine intelligence is being used to decide who is who in society. Claiming that a good example is how the technology was used to decide the president of the U.S. in the past presidential election.

Yes, the impact of AI in society is so serious that experts can’t help being stressed when they think of possible misuse of this tech. The unfortunate thing is that, even before Musk, the late Prof. Hawking, and other tech aficionados manage to pressure governments to control the coming robots that are said might harvest human organs, claims have it that already, AI is biased and in a horrible way.

Source: Santa Cruz Tech Beat

Okay, the original intention of machine intelligence was to solve society problems. And that goes without saying that AI is progressing quite well in solving medical issues. For sometimes now, the tech has also been seen as instrumental in tackling security issues, which has led the U.S., Israel, and China employ it to curb criminals activities.

What About When AI is Not Reliable

Last month during the AI Ethics and Society conference, which took place in New Orleans, a team of scientists presented an invention they thought would be welcomed. The team leader showed how a new algorithm could be used to predict any retaliatory activity after a crime. However, a realistic debate broke forth, critics pointing out how the same system may erroneously mark innocent people as criminals or create mistrust among the police.

The concerns were so viable and with a high likelihood of occurring in real-life that the presenter received a serious backlash from the audience. The astonishing fact was that the Harvard University engineer agreed that the concerns were feasible but tried to run from the responsibility. “I’m just an engineer,” he said, forgetting they are the people developing these algorithms for human use.

Source: Computer Weekly

In other words, if the audience didn’t speak out, that would have been considered a safe project, while it’s not. Another serious real-life occurrence: a Harvard professor, Dr. Latanya Sweeney, is on record saying that she once paid a company to have her presumably clean record of citizenship presented to a potential employer, but after a Google search, her name came up associated to targeted ads that suggest she’d been in detainment before.

The Dire Consequences of Biased AI

It’s obvious that she lost the opportunity because that was a major requirement — but after a serious follow-up Dr. Sweeney found out that a great percentage of the names that suggesting the owner is a“black” citizen, in fact, 80% of such names returned with the same targeted arrest ad.

But with the seemly white names, the results were amazingly clean. That simply suggests that the search engines -which are said to lead in using AI technology have serious bias issue. In the right language, an expert would say it’s the algorithm that is being used which is biased.

This doesn’t end there. Another study has revealed that women don’t get to see the high paying social media job posting ads as much as men do. And unfortunately, the whole thing is deliberate, the algorithms have been trained to limit female profiles from seeing the job ad. Not to mention the more serious allegations that now claim, the algorithms being used as admissible testimony in courtrooms seem to propagate systemic inequalities — something that is completely against the U.S. penal system.

Source: LinkedIn

Way Forward

Fortunately, there has been a raising force seeking to correct the ethical issues in AI. Several non-governmental organizations have taken it upon themselves to pressure top institutions and governments to handle the matter with the seriousness it deserves.

Dr. Kate Crawford one of the renowned researchers in data bias and algorithmic accountability, during the Neural Information Processing Systems, gave a moving speech against “biased AI,” turning heads in agreement. FATML, a group that seeks to ensure Accountability in Machine Learning is also another force that is fervently pushing to solve biases and discriminations in AI.

Nonetheless, you and I also have a responsibility to voice our concerns on all available platforms — to ensure artificial intelligence and any other emerging technologies remain ethical, safe and fair to all humanity.