Search

US-12626165-B2 - Reducing computational requirements for machine learning model explainability

US12626165B2US 12626165 B2US12626165 B2US 12626165B2US-12626165-B2

Abstract

A first input transaction is classified into a first input space cluster in a set of input space clusters. It is determined that the first input space cluster maps to a single explainability space cluster in a set of explainability space clusters. Using an interpretable model corresponding to the single explainability space cluster, a first machine learning model prediction is explained, the first machine learning model prediction resulting from processing, by a machine learning model, the first input transaction.

Inventors

  • Stefan A. G. Van Der Stockt
  • Erika Agostinelli
  • Edward James Biddle
  • Sourav Mazumder

Assignees

  • INTERNATIONAL BUSINESS MACHINES CORPORATION

Dates

Publication Date
20260512
Application Date
20220323

Claims (18)

  1. 1 . A computer-implemented method comprising: outputting, from a machine learning model and responsive to processing a first input transaction, a first machine learning prediction; classifying, into a first input space cluster in a set of input space clusters, the first input transaction, wherein the first input space cluster comprises a first set of member transactions in which each member transaction includes a first common set of weighted explainability features that is descriptive of the corresponding member transaction; mapping the first input space cluster to a single explainability space cluster in a set of explainability space clusters, wherein each second member transaction in the single explainability space cluster includes a second common set of weighted explainability features that is descriptive of a common result produced from each second member transaction, and wherein the mapping is stable when there is a complete overlap between the first common set of weighted explainability features and the second common set of weighted explainability features; and generating, responsive to the mapping being stable, by executing a sequence of tests at a set of nodes in a decision tree of an interpretable model corresponding to the single explainability space cluster, an explanation output corresponding to the first machine learning model prediction, wherein a node in the set of nodes comprises a test on a corresponding weighted explainability feature, wherein a branch from the node comprises an outcome of the test, and wherein the sequence is constructed a path from a root node to a leaf node in the decision tree, the sequence of tests in the path producing the explanation output.
  2. 2 . The computer-implemented method of claim 1 , wherein the explaining is performed responsive to determining that the first machine learning model prediction matches, within a threshold amount of similarity, a result provided by the interpretable model.
  3. 3 . The computer-implemented method of claim 1 , further comprising: classifying, into a second input space cluster in the set of input space clusters, a second input transaction; determining that the second input space cluster maps to more than one explainability space cluster; and explaining, using an explainability model, a second machine learning model prediction, the second machine learning model prediction resulting from processing, by the machine learning model, the second input transaction.
  4. 4 . The computer-implemented method of claim 1 , further comprising: determining, using feature importance data produced by using an explainability model to analyze a set of training transactions, a set of weighted explainability features, a weighted explainability feature in the set of weighted explainability features comprising a weight of a contribution of an explainability feature in explaining a machine learning model prediction, the machine learning model prediction resulting from processing, by the machine learning model, a training transaction in the set of training transactions; grouping, into the set of explainability space clusters according to the weighted set of explainability features, the set of training transactions; and constructing, for each explainability space cluster in the set of explainability space clusters, a corresponding interpretable model.
  5. 5 . The computer-implemented method of claim 4 , further comprising: constructing, by clustering the set of training transactions according to values of the set of weighted explainability features in the set of training transactions, the set of input space clusters.
  6. 6 . The computer-implemented method of claim 4 , wherein the weight of the contribution of the explainability feature in explaining the machine learning model prediction is above a threshold weight.
  7. 7 . A computer program product for machine learning model explainability, the computer program product comprising: one or more computer readable storage media, and program instructions collectively stored on the one or more computer readable storage media, the stored program instructions comprising: program instructions to output, from a machine learning model and responsive to processing a first input transaction, a first machine learning prediction; program instructions to classify, into a first input space cluster in a set of input space clusters, the first input transaction, wherein the first input space cluster comprises a first set of member transactions in which each member transaction includes a first common set of weighted explainability features that is descriptive of the corresponding member transaction; program instructions to perform a mapping the first input space cluster to a single explainability space cluster in a set of explainability space clusters, wherein each second member transaction in the single explainability space cluster includes a second common set of weighted explainability features that is descriptive of a common result produced from each second member transaction, and wherein the mapping is stable when there is a complete overlap between the first common set of weighted explainability features and the second common set of weighted explainability features; and program instructions to generate, responsive to the mapping being stable, by executing a sequence of tests at a set of nodes in a decision tree of an interpretable model corresponding to the single explainability space cluster, an explanation output corresponding to the first machine learning model prediction, wherein a node in the set of nodes comprises a test on a corresponding weighted explainability feature, wherein a branch from the node comprises an outcome of the test, and wherein the sequence is constructed a path from a root node to a leaf node in the decision tree, the sequence of tests in the path producing the explanation output.
  8. 8 . The computer program product of claim 7 , wherein the explaining is performed responsive to determining that the first machine learning model prediction matches, within a threshold amount of similarity, a result provided by the interpretable model.
  9. 9 . The computer program product of claim 7 , the stored program instructions further comprising: program instructions to classify, into a second input space cluster in the set of input space clusters, a second input transaction; program instructions to determine that the second input space cluster maps to more than one explainability space cluster; and program instructions to explain, using an explainability model, a second machine learning model prediction, the second machine learning model prediction resulting from processing, by the machine learning model, the second input transaction.
  10. 10 . The computer program product of claim 8 , the stored program instructions further comprising: program instructions to determine, using feature importance data produced by using an explainability model to analyze a set of training transactions, a set of weighted explainability features, a weighted explainability feature in the set of weighted explainability features comprising a weight of a contribution of an explainability feature in explaining a machine learning model prediction, the machine learning model prediction resulting from processing, by the machine learning model, a training transaction in the set of training transactions; program instructions to group, into the set of explainability space clusters according to the weighted set of explainability features, the set of training transactions; and program instructions to construct, for each explainability space cluster in the set of explainability space clusters, a corresponding interpretable model.
  11. 11 . The computer program product of claim 10 , the stored program instructions further comprising: program instructions to construct, by clustering the set of training transactions according to values of the set of weighted explainability features in the set of training transactions, the set of input space clusters.
  12. 12 . The computer program product of claim 10 , wherein the weight of the contribution of the explainability feature in explaining the machine learning model prediction is above a threshold weight.
  13. 13 . The computer program product of claim 7 , wherein the stored program instructions are stored in the at least one of the one or more storage media of a local data processing system, and wherein the stored program instructions are transferred over a network from a remote data processing system.
  14. 14 . The computer program product of claim 8 , wherein the stored program instructions are stored in the at least one of the one or more storage media of a server data processing system, and wherein the stored program instructions are downloaded over a network to a remote data processing system for use in a computer readable storage device associated with the remote data processing system.
  15. 15 . The computer program product of claim 7 , wherein the computer program product is provided as a service in a cloud environment.
  16. 16 . A computer system comprising one or more processors, one or more computer-readable memories, and one or more computer-readable storage media, and program instructions stored on at least one of the one or more storage media for execution by at least one of the one or more processors via at least one of the one or more memories, the stored program instructions comprising: program instructions to output, from a machine learning model and responsive to processing a first input transaction, a first machine learning prediction; program instructions to classify, into a first input space cluster in a set of input space clusters, the first input transaction, wherein the first input space cluster comprises a first set of member transactions in which each member transaction includes a first common set of weighted explainability features that is descriptive of the corresponding member transaction; program instructions to perform a mapping the first input space cluster to a single explainability space cluster in a set of explainability space clusters, wherein each second member transaction in the single explainability space cluster includes a second common set of weighted explainability features that is descriptive of a common result produced from each second member transaction, and wherein the mapping is stable when there is a complete overlap between the first common set of weighted explainability features and the second common set of weighted explainability features; and program instructions to generate, responsive to the mapping being stable, by executing a sequence of tests at a set of nodes in a decision tree of an interpretable model corresponding to the single explainability space cluster, an explanation output corresponding to the first machine learning model prediction, wherein a node in the set of nodes comprises a test on a corresponding weighted explainability feature, wherein a branch from the node comprises an outcome of the test, and wherein the sequence is constructed a path from a root node to a leaf node in the decision tree, the sequence of tests in the path producing the explanation output.
  17. 17 . The computer system of claim 16 , wherein the explaining is performed responsive to determining that the first machine learning model prediction matches, within a threshold amount of similarity, a result provided by the interpretable model.
  18. 18 . The computer system of claim 16 , the stored program instructions further comprising: program instructions to classify, into a second input space cluster in the set of input space clusters, a second input transaction; program instructions to determine that the second input space cluster maps to more than one explainability space cluster; and program instructions to explain, using an explainability model, a second machine learning model prediction, the second machine learning model prediction resulting from processing, by the machine learning model, the second input transaction.

Description

BACKGROUND The present invention relates generally to a method, system, and computer program product implementing machine learning model explainability. More particularly, the present invention relates to a method, system, and computer program product for reducing computational requirements for machine learning model explainability. A machine learning model is a model trained on training data to make predictions or decisions without being explicitly programmed with a set of rules. Instead, the model learns from the training data. In machine learning, a feature is an individual measurable property or characteristic of a phenomenon. A feature is also a data attribute. A machine learning model is said to produce a result, or prediction, from a set of input feature data. The combination of the set of input feature data and the corresponding model output is also called a transaction. For example, one well-known dataset used in machine learning experiments predicts whether or not a passenger survived the sinking of the Titanic using input features such as the passenger's gender, age, the class of the passenger's ticket, where the passenger embarked from, the fare paid, the deck where the passenger's cabin was, and the passenger's ticket number. The model's prediction can then be checked against the passenger's actual result. Because machine learning models learn from training data, the models typically provide results, but do not articulate how a model came to a specific result. However, understanding how a model came to a specific result—also called model explainability—helps ensure that the system continues to perform as expected, even if production data differs from the original training data. Model evaluation also helps a business compare model predictions, quantify model risk, and optimize model performance. Model explainability is also important in promoting user trust in the model's results, and helps those affected by a decision to challenge or change that outcome. Model explainability also helps mitigate compliance, legal, security, and reputational risks of model use, and might be necessary to meet regulatory standards or guidelines. For example, consider a machine learning model configured to predict whether or not to approve a loan to a borrower. Borrowers who are not approved will want to understand why. The lending institution will want to ensure that its model is accurate, so that borrowers who are approved actually pay their loans back. There may also be compliance, reputational, legal, and regulatory requirements associated with particular loan decisions. SUMMARY The illustrative embodiments provide a method, system, and computer program product. An embodiment includes a method that classifies, into a first input space cluster in a set of input space clusters, a first input transaction. An embodiment determines that the first input space cluster maps to a single explainability space cluster in a set of explainability space clusters. An embodiment explains, using an interpretable model corresponding to the single explainability space cluster, a first machine learning model prediction, the first machine learning model prediction resulting from processing, by a machine learning model, the first input transaction. An embodiment includes a computer usable program product. The computer usable program product includes one or more computer-readable storage devices, and program instructions stored on at least one of the one or more storage devices. An embodiment includes a computer system. The computer system includes one or more processors, one or more computer-readable memories, and one or more computer-readable storage devices, and program instructions stored on at least one of the one or more storage devices for execution by at least one of the one or more processors via at least one of the one or more memories. BRIEF DESCRIPTION OF THE DRAWINGS Certain novel features believed characteristic of the invention are set forth in the appended claims. The invention itself, however, as well as a preferred mode of use, further objectives and advantages thereof, will best be understood by reference to the following detailed description of the illustrative embodiments when read in conjunction with the accompanying drawings, wherein: FIG. 1 depicts a block diagram of a network of data processing systems in which illustrative embodiments may be implemented; FIG. 2 depicts a block diagram of a data processing system in which illustrative embodiments may be implemented; FIG. 3 depicts a block diagram of an example configuration for reducing computational requirements for machine learning model explainability in accordance with an illustrative embodiment; FIG. 4 depicts a block diagram of an example configuration for reducing computational requirements for machine learning model explainability in accordance with an illustrative embodiment; FIG. 5 depicts a block diagram of an example configuration for reducing compu