Robustness of explainable AI algorithms for disease biomarker discovery from functional connectivity datasets
In this paper, we propose quantitative metrics to evaluate the robustness of salient features identified DL models for brain disease classification.
This repository contains the implementation of the following metrics:-
- Model Parameter Randomization Check
- Data Randomization Check
- Fidelity
- Stability
- Target Sensitivity
- Implementation Invariance
- Modular Ratio
- Hub Assortativity Coefficient
We package all these metrics in the RE-CONFIRM framework.
-
The
tasksfolder holds the function implementations for the RE-CONFIRM metrics. -
The
assetsfolder includes all necessary data to execute the scripts intasks. This data includes encoded labels for both the ABIDE and ADHD datasets, as well as the topological properties of functional connectivity, saved in.npyformat. -
The
utilsfolder contains wrapper functions and helper code for calculating the metrics. -
Finally,
run_experiments.pyandmain.pyprovide baseline code for training models, generating saliency scores, and calculating metrics based on these scores. -
The script
run_experiments.pyrequires four mandatory arguments at runtime: type of GNNs[gcn, cheb, gat, sage]to be utilized, explainability method to be applied[gnnexp, attnexp, ig, pgexp, gbackprop], dataset to be analyzed[abide, adhd]and choice of analysis site: either a single site (NYU) or a combination of all available sites[yes, no]. -
Additionally,
main.pyrequires an extra argument specifying the evaluation metric to be used:[scores, mprc, drc, stability, invariance, ambivert, hdr, fidelity_individual, homfid, mod] -
You can run the code by running the following command:
python run_experiments.py gcn gnnexp abide noHere, we are utilizing the GCN architecture with the GNNExplainer method, applied across all sites of the ABIDE dataset.
More updates to follow.