The Ethics of AI
This article discusses the importance of implementing ethical practices within your predictive models, data workflows, products and AI research.
Join the DZone community and get the full member experience.
Join For FreeWhether in daily mobility, in industrial applications, or in the form of assistance solutions at home, artificial intelligence permeates an ever wider range of our lives. It is associated with great hopes, but it also raises fears. Therefore, the call for ethical guidelines regarding the new technologies is becoming increasingly louder.
There are various questions we need to find answers to:
- How can the “right” values be integrated into technical action?
- Can and may AI determine what is right?
- How to ensure that AI is developed, used, and provided in the service of mankind?
- And how can the decision-making processes of AI, including wrong decisions, be made transparent, given the huge, complex amounts of data?
We organized a panel discussion on what is importance of implementing ethical practices within your predictive models, data workflows, products and AI research. I was the part of the panel along with Scott Haines, Lizzie Siegle, and Nick Walsh.
In this article, we will go through some of the points we discussed with the panel and their views on various topics along with my view on each topic.
For the sake of simplicity, we will discuss further details into 3 broad categories:
- Fairness and Bias
- Interpretability and Explainability
- Privacy
Let's jump into it!
1. Fairness and Bias
Bias is often identified as one of the major risks associated with artificial intelligence (AI) systems. As a survey on bias and fairness in machine learning suggests, there are clear benefits to algorithmic decision-making; unlike people, machines do not become tired or bored and can take into account orders of magnitude more factors than people can. However, like people, algorithms are vulnerable to biases that render their decisions “unfair.” In the context of decision-making, fairness is the absence of any prejudice or favoritism toward an individual or a group based on their inherent or acquired characteristics. Thus, an unfair algorithm is one whose decisions are skewed toward a particular group of people.
The public discussion about bias in such scenarios often assigns blame to the algorithm itself. The algorithm, it is said, has made the wrong decision, to the detriment of a particular group. But this claim fails to take into account the human component: People perceive bias through the subjective lens of fairness.
So, bias occurs when we discriminate against (or promote) a defined group consciously or unconsciously, and it can creep into an AI system as a result of skewed data or an algorithm that does not account for skewed data. Fairness, meanwhile, is a social construct. And in fact, when people judge an algorithm to be “biased,” they are often conflating bias and fairness: They are using a specific definition of fairness to pass judgment on the algorithm.
No AI system can be universally fair or unbiased as there are more than 20 definitions of fairness and it is impossible for every decision to be fair to all parties. But we can design these systems to meet specific fairness goals, thus mitigating some of the perceived unfairness and creating a more responsible system overall.
This discussion also brings us to a question if we can relate trolley problem here?
Let's first understand what is trolley problem:
A runaway trolley is heading down the tracks toward five workers who will all be killed if the trolley proceeds on its present course. Adam is standing next to a large switch that can divert the trolley onto a different track. The only way to save the lives of the five workers is to divert the trolley onto another track that only has one worker on it. If Adam diverts the trolley onto the other track, this one worker will die, but the other five workers will be saved.
Should Adam flip the switch, killing the one worker but saving the other five?
The original trolley problem came from a paper about the ethics of abortion, written by English philosopher Phillipa Foot. It’s the third of three increasingly complicated thought experiments she offers to help readers assess whether intentionally harming someone (e.g. choosing to hit them) is morally equivalent to merely allowing harm to occur (choosing not to intervene to stop them from getting hit).
Apart from the above version, there are various modern-day implications of this problem which are considered while social psychology studies conducted by different universities. You can try Moral Machine by MIT.
Who else is ready to accept this as a solution?
Jokes aside, is this applicable in the case of AI?
Simon Beard explains in his article that this philosophical issue is irrelevant to self-driving cars because they don’t have intentions. Those who think intentions are morally significant tend to hold strong views about our freedom of will. But machines don’t have free will (at least not yet). Thus, as far back as Isaac Asimov’s three laws of robotics, we’ve recognized that for machines to harm someone — or “through inaction, allow a human being to come to harm” — is morally equivalent.
This seems a logical explanation as AI because the trolley problem is very situational and AI works in real-world scenarios which are extremely uncertain and you can't certainly apply universal laws of ideal behavior to machines. AI tends to learn continuously from its mistakes and tries to achieve a state with better decision-making capabilities with each iteration.
This brings us to the second point of our discussion of how important explainability and interpretability of decisions made by AI is, considering the ever-increasing lack of transparency has only become more visible amidst techlash.
2. Interpretability and Explainability
When it comes to machine learning and artificial intelligence, explainability and interpretability are often used interchangeably.
According to Miller, interpretability is the degree to which a human can understand the cause of a decision. Interpretable predictions lead to better trust and provide insight into how the model may be improved. We can also define interpretability as the extent to which we are able to predict what is going to happen, given a change in input or algorithmic parameters.
Explainability, meanwhile, is the extent to which the internal mechanics of a machine or deep learning system can be explained in human terms. It’s easy to miss the subtle difference with interpretability but consider it like this: interpretability is about being able to discern the mechanics without necessarily knowing why. Explainability is being able to quite literally explain what is happening.
Let's understand using example, consider a chemistry experiment: Knowing that when you mix 1 gm. of blue color chemical with 100 ml of liquid you get pink color solution and this is interpretability but understanding the exact chemical reaction for this experiment means explainability.
Interpretability and explainability are not required if the model has no significant impact or when the problem is well studied. This might also enable people or programs to manipulate the system but if we’re unable to properly deliver improved interpretability, and ultimately explainability, in our algorithms, we’ll seriously be limiting the potential impact of artificial intelligence. Which would be a shame.
Let's discuss few methods to increase the interpretability of complex ML models.
LIME or Local Interpretable Model-Agnostic Explanations, is a method developed in the paper Why should I trust you? for interpreting individual model predictions based on locally approximating the model around a given prediction. The researchers explain that LIME can explain “the predictions of any classifier in an interpretable and faithful manner, by learning an interpretable model locally around the prediction.”
What this means in practice is that the LIME model develops an approximation of the model by testing it out to see what happens when certain aspects within the model are changed. Essentially it’s about trying to recreate the output from the same input through a process of experimentation.DeepLIFT (Deep Learning Important FeaTures) is another method that serves as a recursive prediction explanation method for deep learning. This method decomposes the output prediction of a neural network on a specific input by backpropagating the contributions of all neurons in the network to every feature of the input. DeepLIFT assigns contribution scores based on the difference between activation of each neuron and its ‘reference activation.’ DeepLIFT can also reveal dependencies that are missed by other approaches by optionally giving separate consideration to positive and negative contributions.
Layerwise Relevance Propagation (LRP) is a technique for determining which features in a particular input vector contribute most strongly to a neural network’s output. The technique was originally described in this paper. It defines a set of constraints to derive a number of different relevance propagation functions.
Algorithmic generalization: Improving generalization is not as easy as it sounds. Sometimes Models feel secondary when we realize that most ML engineering is applying algorithms in a very specific way to uncover a certain desired outcome. However, by shifting this attitude to consider the overall health of the algorithm, and the data on which it is running, you can begin to set a solid foundation for improved interpretability.
Pay attention to feature importance: Feature importance tries to answer question, what features have the biggest impact on predictions. Looking closely at the way the various features of your algorithm have been set is a practical way to actually engage with a diverse range of questions, from business alignment to ethics. Debate and discussion over how each feature should be set might be a little time-consuming, but having that tacit awareness that different features have been set in a certain way is nevertheless an important step in moving towards interpretability and explainability.
So, we explored the views around to open a black-box or not! Now, we have to think about the privacy aspect, which is one of the major concerns in today's world.
3. Privacy
When we look around for application areas of AI, we find that much of the most privacy-sensitive data analysis today–such as search algorithms, recommendation engines, and adtech networks–are driven by machine learning and decisions by algorithms. As artificial intelligence evolves, it magnifies the ability to use personal information in ways that can intrude on privacy interests by raising analysis of personal information to new levels of power and speed.
The discussion of AI in the context of the privacy debate often brings up the limitations and failures of AI systems. Bias in data and unfair systems have raised significant issues, but privacy legislation is complicated enough even without packing in all the social and political issues that can arise from uses of information. As Cameron F. Kerry suggested in one of his survey, to evaluate the effect of AI on privacy, it is necessary to distinguish between data issues that are endemic to all AI, like the incidence of false positives and negatives or overfitting to patterns, and those that are specific to use of personal information.
The privacy legislative proposals that involve these issues refer to decisions made by AI as “automated decisions” (borrowed from EU data protection law) or “algorithmic decisions.” In general, I believe that this helps to shift people’s focus from the use of AI as a tool to the use of personal data in AI and to the impact this use may have on individuals. This debate centers in particular on algorithmic bias and the potential for algorithms to produce unlawful or undesired discrimination in the decisions to which the algorithms relate. These are major concerns for civil rights and consumer organizations that represent populations that suffer undue discrimination.
There's a lot to think about and discuss about the scope of privacy legislation as some of them also address algorithmic discrimination.
Let's split our views into two separate points:
- to what extent can or should legislation address issues of algorithmic bias? According to Law, Discrimination is not self-evidently a privacy issue, since it presents broad social issues that persist even without the collection and use of personal information, and fall under various civil rights laws.
- When we want to ensure privacy protection in AI systems and regulate use of consumer data and try to stop this unfair and deceptive practices the consumer choice based notice-and-choice model for privacy policies become meaningless. We will require a change in the paradigm of privacy, to protect such privacy interests in the context of AI.
we need an approach that addresses risk more obliquely, with accountability measures designed to identify discrimination in the processing of personal data and regulate use of it. The number of organizations and companies, as well as several legislators, propose such accountability. Their proposals take various forms:
Transparency: Accountability is a major concern and being able to examine how a company is handling data makes it much easier to hold the company accountable for things. This mostly refers to disclosures relating to the uses of algorithmic decision-making. Replacing current lengthy, detailed privacy policies are not helpful to most consumers with “privacy disclosures” that require a complete description of what and how data is collected, used, and protected would enhance the benchmarking which is being done by regulators. In turn, requiring that these disclosures identify significant uses of personal information for algorithmic decisions would help watchdogs and consumers know where to look out for untoward outcomes.
Explainability: As we discussed above, this helps us understand how the data has been used to reach to a particular decision and which features have played a significant role in the conclusion. European Union’s General Data Protection Regulation (GDPR) also follows this approach to settle disputes because of automated decisions. The GDPR requires that, for any automated decision with “legal effects or similarly significant effects” such as employment, credit, or insurance coverage, the person affected has recourse to a human who can review the decision and explain its logic. This incorporates a “human-in-the-loop” component and an element of due process that provide a check on anomalous or unfair outcomes. The central problem with both explainability is that you’re adding an additional step in the development process along with significant regulatory burden. Indeed, you’re probably adding multiple steps. From one perspective, this looks like you’re trying to tackle complexity with even greater complexity.
Risk assessment: When we talk about risk assessment in this context we are mostly dealing will assessing the impact of algorithmic decision on individuals and potential bias in the design of a system. As Cameron F. Kerry rightly points out, for the regulatory burden to be proportionate, the level of risk assessment should be appropriate to the significance of the decision-making in question, which depends on the consequences of the decisions, the number of people and volume of data potentially affected, and the novelty and complexity of algorithmic processing.
Audits: There are some general accountability requirements as well as ways such as self-audits or third-party audits which help authorities to ensure companies comply with their privacy programs. When we couple auditing outcomes of AI decisions with proactive risk assessments it can help us match foresight with hindsight; although, like explainability, auditing machine-learning routines is difficult and still developing.
Because of the difficulties of foreseeing machine learning outcomes as well as reverse-engineering algorithmic decisions, no single measure can be completely effective in avoiding perverse effects. Thus, where algorithmic decisions are consequential, it makes sense to combine measures to work together. That's why it’s important for algorithm operators and developers to ensure that they won't leave some groups of people worse off as a result of the algorithm’s design or its unintended consequences. [AI debate, Nicol Turner Lee with Paul Resnick and Genie Barton]
On a Positive Note
Amid the dissonance of concern over unfair decisions made by artificial intelligence (AI) and use of personal data. The potential for AI to do good cannot be overlooked. Technology leaders such as Microsoft, IBM, Google, and many others have entire sections of their business focused on the topic and dedicate resources to build AI solutions for good and to support developers who do. In the fight to solve extraordinarily difficult challenges such as Accessibility, Climate Change, Conservation and the Environment, World hunger, Human rights, Fake News and so many others, we can surely use help by AI!
Note
: In case you're interested to watch this discussion, you can check out the video here:
References:
- Importance of Interpretability
- MIT's self-driving cars survey for Moral Machines
- can we open the Black Box?
- What do we do about the biases in AI?
- what is fair?
- A Survey on Bias and Fairness in Machine Learning
- Explainability & interpretability in AI
- protecting privacy in an AI driven world
- The trolley dilemma in AI
- The problem with the trolley problem
- Layerwise Relevance Propagation
- AI & Privacy
Published at DZone with permission of Jayesh Bapu Ahire, DZone MVB. See the original article here.
Opinions expressed by DZone contributors are their own.
Comments