Testing
How do the out-of-the-box tests developed by ValidMind work?
All the default tests are developed using open-source Python and R libraries.
The ValidMind Library1 test interface is a light wrapper that defines utility functions to agnostically interact with different dataset and model backends, and contains functions to collect and post results to the ValidMind Platform2 using a generic results schema.
When do I use tests and tests suites?
While you have the flexibility to decide when to use which ValidMind tests, here are a few typical scenarios:3
- Dataset testing — To document and validate your dataset.
- Model testing — To document and validate your model.
- End-to-end testing — To document a binary classification model and the relevant dataset end-to-end.
Can we configure, customize, or add our own tests?
Yes, ValidMind allows tests to be manipulated at several levels:
- You can configure which tests are required to run programmatically depending on the model use case.4
- You can change the thresholds and parameters for default tests already available in the library — for instance, changing the threshold parameter for the class imbalance flag.5
- You can also connect your own custom tests with the ValidMind Library. These custom tests are configurable and are able to run programmatically, just like the rest of the library.6
Does ValidMind support using synthetic datasets?
- The ValidMind Library supports you bringing your own datasets, including synthetic datasets, for testing and benchmarking purposes such as for fair lending and bias testing.13
- If you are unable to share your real-world data with us, ValidMind is happy to work with you to generate custom synthetic datasets based on characteristics of your data, or provide scripts to assist with synthetic dataset generation if details cannot be shared.
Does ValidMind support monitoring models after deployment?
Yes, ValidMind offers ongoing monitoring support to help you regularly assess a model’s accuracy, stability, and robustness to ensure it remains reliable after deployment:
- You can enable monitoring for both new and existing models.14
- You use the ValidMind Library to automatically populate the monitoring template for your model with data, providing a comprehensive view of your model’s performance over time.
- You then access and examine these results within the ValidMind Platform, allowing you to identify any deviations from expected performance and take corrective actions as needed.15
- Once generated via the ValidMind Library, view and add metrics over time to your ongoing monitoring plans in the ValidMind Platform.16