AI and information crime

Artificial Intelligence (AI) is moving at speed into the mainstream. Almost every day we are learning about new uses for it, and discovering the ways in which it is already a part of our daily lives. Just this week medical researchers have revealed that the use of AI in testing for breast cancer is more effective than the work of expert radiologists. As one doctor interviewed said, this use of AI frees up radiologists to do other important work. And that’s a great benefit, as it takes roughly 10 years to train in radiology.

On the other hand, people are concerned that AI may take over almost every area of our lives, from self-driving cars to fighting wars. And it may do our laundry as well. Basically it comes down to this — will AI replace humans? That’s the great fear, but one which is largely exaggerated. As Kathleen Walch writes: “However, it’s becoming increasingly clear that AI is not a job killer, but rather, a job category killer.” I have also written about this aspect of AI before, pointing to the fact that “jobs are not destroyed, but rather employment shifts from one place to another and entirely new categories of employment are created.”

Indeed, as Walch says, “companies will be freed up to put their human resources to much better, higher value tasks instead of taking orders, fielding simple customer service requests or complaints, or data entry related tasks.” What businesses must do is have honest conversations with their employees about the use of AI and show how it can allow humans to be more creative by giving the dull, routine tasks to AI.

The one area where AI is causing us issue is in the generation of fake news in a range of formats. It is already almost impossible to tell if an image is real or AI-generated, or if you’re talking to a bot or a real person? AI-generated ‘disinformation’ is not necessarily generated by criminals: as we all now know, State actors are perhaps the worst offenders, and we have plenty of example to look at coming from Russia, the USA and the UK. Lies are fed to the citizens using social media accounts that appear to be reputable government sources, and the social media companies collude with these sources, as Facebook has shown us. Walch says, “Now all it takes is a few malicious actors spreading false claims to traumatically alter public opinion and quickly shift the public’s view.” Brexit and the election of Trump are good examples of this in play.

And it is in this area that we must consider the ethics of using AI most closely right now. As Walch says, “Governments and corporations alike will have to think about how they will reign in the potential damage done by AI-enabled content creation,” and she adds, “we must encourage companies and governments to consider fake content to be as malicious as cybersecurity threats and respond appropriately.”

What we are talking about is essentially propaganda. There are those of us who can see through the smoke and mirrors, but many can’t, and these citizens need protection from the malicious acts of the information criminals.

The problem with AI bias

Ai has come a long way. Just after WW2, there was a preconception that developing Artificial Intelligence would lead to something like an ‘Attack of the Zombie Robots’, and that AI could only be a bad thing for humanity. Fortunately we have come a long way from the old sci-fi view of AI, and we even have robotics used in surgery, but there is still a lingering feeling that AI and robotics are threatening in some way, and one of those ways is ‘bias’.

AI is very much part of the fourth industrial revolution, which also includes cyber-physical systems powered by technologies like machine learning, blockchain, genome editing and decentralised governance. The challenges that we face in developing our use of AI, are tricky ethical ones for the most part, which need a sensitive approach.

So, what is the issue? As James Warner writes in his article on AI and bias,

“AI is aiding the decision making from all walks of life. But, the point is that the foundation of AI is laid by the way humans function.” And as we all know — humans have bias, unfortunately. Warner says, “It is the result of a mandatory limited view of the world that any single person or group can achieve. This social bias just as it reflects in humans in sensitive areas can also reflect in artificial intelligence.”

And this human bias, as it cascades down into AI can be dangerous to you and me. For example, Warner writes: “the investigative news site Pro Publica found that a criminal justice algorithm used in Florida mislabelled African American defendants as high risk. This was at twice the rate it mislabelled white defendants.” Facial recognition has already been highlighted as an area with shocking ethnic bias, as well as recognition errors.

IBM suggests that researchers are quickly learning how to handle human bias as it infiltrates AI. It has said that researchers are learning to deal with this bias and coming up with new solutions that control and finally free AI systems out of bias.

The ways in which bias can creep in are numerous, but researchers are developing algorithms that can assist with detecting and mitigating hidden biases in the data. Furthermore, scientists at IBM have devised an independent bias rating system with the ability to determine the fairness of an AI system.

One outcome of all this may be that we discover more about how human biases are formed and how we apply them throughout our lives. Some biases are obvious to us, but others tend to sneak around unnoticed, until somebody else points it out. Perhaps we will find that AI can teach us how to handle a variety of biased opinions, and be more fair ourselves.

What is the point of a robot tax?

While browsing articles on Artificial Intelligence, I stumbled across a piece by Milton Ezrati at Forbes. Discussing the possibility of a robot tax? This idea had been proposed by Bill de Blasio before he gave up his bid to gain the Democratic presidential nomination. Ezrati thinks it is a dreadful idea, but he is aware that both Silicon Valley leaders and current government progressives are quite keen on it.

According to the article, a robot tax would have four parts: First, it would apply to any company introducing labour saving automation. Second, it would insist that this employer either find new jobs for the displaced workers at their same pay level or pay them a severance. Third, the tax would require a new federal agency, the Federal Automation and Worker Protection Agency (FAWPA) and fourth, it would require Washington eliminate all tax incentives for any innovation that leads to automation.

The assumption appears to be that workers displaced by automation will never again find work at a comparable wage. Elon Musk for one, Bill Gates and Mark Zuckerberg are amongst those who are worried about this aspect of it, as is Democratic candidate, Andrew Yang, who suggests the introduction of a universal basic income, “to substitute, he claims, for the incomes lost to robots and artificial intelligence generally.”

However, it is not proven that the introduction of AI and robots will disadvantage workers so substantially. As Ezrati say, “innovation, if it initially displaces some workers, always eventually creates many more new jobs even as it boosts overall productivity and increases output.”

And, as he also points out, “since the industrial revolution began more than 250 years ago, business and industry have actively applied wave after wave of innovation and yet economies have nonetheless continued to employ on average some 95 percent of those who want to work.”

In my opinion, and in this respect I am in agreement with Ezrati, we have focused far too much on what will be lost with the introduction of more robotics, and not sufficiently on what is to be gained. His analogy that uses the introduction of email and the Internet regarding typists’ jobs illustrates this. Whilst those working in admin, messenger departments and typing pools no longer had their current job, new forms of employment emerged for them.

Similarly, when the introduction of automatic teller machines threatened to throw thousands of bank clerks out of work, the machines created profits that meant they could employ more tellers, and these tellers, with the assistance of different technologies, could do more interesting, complex, and valuable jobs at higher pay than they received before the ATMs were put in place.

A robot tax would be counter-productive and stunt growth in innovation, hampering the possibility of finding new types of jobs and improving living standards. It’s a proposed tax that simply doesn’t make sense.

How AI can help you

Artificial Intelligence (AI) is operational in a lot of services these days, from automating processes to delivering a pleasing customer experience online. It is something we are living with and come to accept.

As Jim Sinur argues in his Forbes article, “How can AI help me?” we all have “to deal with living with AI,” and we also have to figure out how to make the best of AI in our lives. Sinur asks, “Will it be like other technology that I have to learn to for life success or will it be more like a person that I have to get along with to thrive in the future? Will AI make our lives better or should we be practicing for those apocalypse scenarios I keep reading about?” How many of us have even raised these questions?

AI can do the drudge work

Of course, responses depend on the individual’s life circumstances. For example, if you are an employee, AI has the potential to do the drudge work, leaving you free to do the more meaningful, creative tasks. And AI can assist you with that. This is a more positive view of Ai in employment and is the opposite of the scenario in which AI and robots will steal all the jobs. Sinur says, “I see AI giving us an edge in boosting our ability to consume and leverage knowledge on a grand scale even outside of our native skillsets, culture bases, and language capabilities,” and this view creates a vision of an exciting future relationship with AI.

A more personalised customer experience

Consumers can also benefit from an enhanced relationship with AI. Sinur points out an important factor that should encourage us as consumers to welcome AI: “With the advent of customer journeys, combined with AI, the complexities and company needs can fade in the background while customer needs are being represented within organizational systems, interfaces, and constraints.” The potential to offer a more personalised customer service is vital, as it is becoming evident that the consumer wants to come first more than ever, and is not prepared to kow-tow to the needs of the business. What is happening in the banking sector is a good example of this: younger customers are gravitating towards neobanks precisely because they offer a personalised service.

AI — your personal assistant

And AI can help us develop personal skills and become a more global citizen. In the case of multi-cultural interactions, AI can help us with language translation and to avoid subtle cultural errors. It may help us to be better communicators in any number of settings and it could even be a life coach and personal trainer.

I agree with Sinur that as it develops we will see that AI adds value to our lives, and that there is so much further to go with it. One day we will be amazed that we were ever afraid of this dynamic technology.