Based on Ph.D. research by Padmaraj Nidagundi - Adapting Lean Canvas for Software Testing Strategies
Specialized approach for testing AI/ML systems and AI-augmented applications
Purpose: Address oracle problem in AI systems where expected outputs are unknown.
Approach: Define metamorphic relations between inputs and outputs. For example, if adding noise to an image shouldn't drastically change classification, test this property rather than exact outputs.
Tools: Custom frameworks, MetamorphicTesting.py libraries
Purpose: Identify vulnerabilities by intentionally trying to fool AI models.
Approach: Generate adversarial examples using techniques like FGSM, PGD, or C&W attacks. Test model robustness against perturbations.
Tools: CleverHans, Foolbox, Adversarial Robustness Toolbox (ART)
Purpose: Ensure AI models don't discriminate against protected groups.
Approach: Measure disparate impact, equal opportunity, demographic parity across sensitive attributes. Test with diverse datasets representing all user groups.
Tools: Fairlearn, AI Fairness 360, What-If Tool
Purpose: Verify that model decisions are interpretable and reasonable.
Approach: Use SHAP, LIME, or attention visualization to understand feature importance. Test if explanations align with domain expertise.
Tools: SHAP, LIME, InterpretML, Captum
Purpose: Ensure AI systems meet latency and throughput requirements.
Approach: Test inference time under varying loads, batch sizes, and concurrent requests. Profile GPU/CPU utilization.
Tools: Locust, JMeter, TensorFlow Profiler, PyTorch Profiler
Purpose: Ensure training and inference data meets quality standards.
Approach: Validate data schema, distributions, completeness, and consistency. Detect drift between training and production data.
Tools: Great Expectations, TensorFlow Data Validation, Pandera
Purpose: Detect when model performance degrades over time.
Approach: Continuously monitor prediction distributions, feature distributions, and performance metrics. Set up alerts for significant drift.
Tools: Evidently AI, WhyLabs, Fiddler AI, Arize
Purpose: Validate end-to-end ML pipelines from data ingestion to prediction serving.
Approach: Test data preprocessing, feature engineering, model inference, and post-processing steps. Verify API contracts and error handling.
Tools: Pytest, MLflow, Kubeflow Pipelines, Airflow
In the rapidly evolving landscape of 2026, where AI and advanced model protocols like MCP (Model Context Protocol) are transforming software development, the Lean Canvas Test Strategy can be leveraged in innovative ways:
Anyone can start by adapting this template: Download the canvas, input your project specifics, and use AI tools to analyze and optimize your test strategy iteratively.