Test Suite#
Post-hoc Explanation#
Calculate Permutation Feature Importance (PFI) for model features. |
|
Calculate H-statistics for all feature pairs to measure feature interactions. |
|
Calculate and visualize Partial Dependence Plot (PDP) for specified model features. |
|
Calculate Accumulated Local Effects (ALE) plots for one or two features. |
|
Generate a LIME (Local Interpretable Model-agnostic Explanations) explanation for a specific sample. |
|
Generate SHAP (SHapley Additive exPlanations) values for local model explanation. |
Inherent Interpretation#
Extracts and visualizes the coefficients of linear model features. |
|
Calculate and visualize global feature importance for the model. |
|
Analyze and visualize how one or two features influence model predictions through main effects or interaction effects. |
|
Calculates and visualizes feature importance scores for a single sample. |
|
Calculate and visualize local feature importance for a specific data sample using linear approximation. |
|
This method generates a table populated with unwrapper summary statistics. |
|
Generate and visualize parallel coordinate plots for Local Linear Model (LLM) coefficients. |
|
Calculate local feature importance for a specific feature using LLM profiles. |
|
Generates the LLM coefficients and statistics. |
|
Generate a visualization of the complete decision tree model structure. |
|
Generate a visualization of the decision path for a specific sample through the decision tree. |
|
Calculate and visualize expert weights for a specific sample. |
|
Analyze feature effects averaged across all mixture-of-experts clusters. |
|
Analyze and summarize characteristics of mixture-of-experts clusters. |
Diagnostics#
Evaluate model performance on training and test datasets. |
|
Analyze the relationship between model residuals and a specified feature. |
|
Analyze model residuals by clustering data points and evaluating performance within clusters. |
|
Analyzes feature importance by examining their relationship with prediction residuals. |
|
Identify low-accuracy regions based on specified slicing features. |
|
Identify overfit regions based on one or two slicing features. |
|
Get unreliable regions based on one or two slicing features. |
|
Get unreliable regions based on one or two slicing features. |
|
Evaluate a model's slicing fairness metric across different protected-reference groups. |
|
Evaluate model robustness by measuring performance under feature perturbations. |
|
Evaluates model reliability using split conformal prediction. |
|
Evaluate model resilience by analyzing performance on challenging data subsets. |
|
Evaluate fairness metrics across different protected and reference groups. |
|
Attempts to mitigate model unfairness by applying feature value binning. |
|
Mitigate model unfairness through feature value binning. |
Model Comparison#
Compare predictive performance metrics across multiple models. |
|
Compares model performance across different data slices based on a specified feature. |
|
Compares model performance across different data slices to identify potential overfit regions. |
|
Compares reliability metrics across different slices of data for multiple models. |
|
Compares model robustness across different data slices by analyzing performance stability under perturbations. |
|
Evaluates fairness metrics across different protected and reference groups by slicing the data. |
|
Performs robustness testing by comparing model performance under different perturbation levels. |
|
Compares reliability performance of multiple models under data shifts by evaluating prediction intervals/sets. |
|
Compare model resilience performance under data shifts across multiple models. |
|
Compare model residuals by clustering data points and evaluating performance within clusters. |
|
Compares fairness metrics across multiple models. |
Utilities#
Return the dataset object. |
|
Set dataset for test suite. |
|
Return the model object. |
|
Set model for test suite. |
|
Return the list of main effects. |
|
Return the list of pairwise interactions. |
|
List all the experiments saved in database. |
|
Register a test into MLFlow. |
|
Load config and result of registered tests. |
|
Load config and result of registered tests. |
|
Return the list all registered tests. |
|
Get ValidationResult object of registered test. |
|
Export report to html |