AffectLog's Trustworthy AI (ALT-AI) - Design Document
AffectLog's Trustworthy AI (ALT-AI) provides a set of tools for explaining, visualizing, and understanding complex machine learning models. It aims to facilitate model transparency, interpretability, and aid compliance with emerging regulatory standards (e.g., GDPR, EU AI Act). ALT-AI helps data scientists, analysts, and stakeholders interpret model predictions, identify feature importance, assess fairness, and evaluate whether models align with ethical and legal requirements.
Technical Usage Scenarios & Features
ALT-AI supports both global (overall model behavior) and local (individual predictions) explanations. It helps users:
- Understand which features influence model outcomes the most
- Compare different models for performance and fairness
- Ensure no demographic group is disproportionately affected
The toolbox is designed to be flexible and scalable, while prioritizing privacy, security, and compliance.
Features/Main Functionalities
- Model Explanation: Provides global and local explanations to clarify how models derive their predictions.
- Feature Importance: Quantifies each feature's impact, aiding interpretability and better feature selection/engineering decisions.
- Model Comparison: Compares different models based on predefined metrics and segments, helping select the most suitable model.
- Fairness Analysis: Evaluates models for fairness and potential biases across sensitive groups.
Technical Usage Scenarios
- Model Development & Validation: Data scientists use ALT-AI to ensure models meet business, ethical, and regulatory standards before deployment.
- Audit & Compliance: ALT-AI facilitates documentation and transparency required for audits, potentially aiding with GDPR adherence and upcoming EU AI Act considerations.
- Feature Engineering: By understanding feature importance, practitioners can refine and enhance their model inputs.
Requirements
- R1: MUST support integration with popular Python-based ML frameworks, including scikit-learn, and, where feasible, TensorFlow and PyTorch models via wrappers. Also supports numpy, pandas for data handling, and onnxruntime for ONNX models.
- R2: MUST provide APIs for generating explanations, feature importance scores, and model comparisons.
- R3: MUST ensure data privacy, security, and must not require access to raw personal data for explanation generation.
- R4: SHOULD leverage partner infrastructure for scalability and handle large datasets and complex models efficiently.
Timeline: Feasibility discussions (e.g., integration with Decentralized AI Training BB) are tentatively planned for Q1 2025. After these discussions, a more precise project timeline and roadmap will be established. A high-level work plan has been shared with the relevant Building Block (BB) and Work Package leader for consideration.
Integrations
Direct Integrations with Other BBs
- Decentralized AI Training BB: ALT-AI may integrate with models produced by the Decentralized AI Training BB to provide post-training analyses. Integration feasibility is subject to discussion with the Decentralized AI Training BB team. The intent is to securely retrieve the resulting trained models—under appropriate consent and policy enforcement via the Prometheus-X Data Space Connector (PDC)—and then perform AI risk assessment and explainability tasks.
Integrations via Connector
- Decentralized AI Training BB: If deemed feasible, ALT-AI could access trained models through the PDC with secure connectors, enforcing ODRL policies and user consent frameworks. ALT-AI would operate on aggregated model artifacts rather than raw data, preserving privacy.
Relevant Standards
- Data Format Standards: Adheres to common data exchange formats (JSON, CSV).
- Model Cards: Follows Mitchell et al. (2019) to document model purpose, performance, and ethical considerations.
- Data Cards: Follows Gebru et al. (2018) to document dataset provenance, collection methods, and known biases.
- GDPR Compliance: ALT-AI operates on aggregated model artifacts and anonymized datasets.
- EU AI Act Alignment: ALT-AI emphasizes fairness, transparency, and explainability.
- PDC Integration (If Feasible): Respects ODRL policies and user consent frameworks in a data-minimized manner.
- Interoperability Standards: Follows DSSC guidelines for data space interoperability.
Supported Model Types
- Scikit-learn Models: Directly supported via joblib or pickle serialization.
- ONNX Models: Supported via onnxruntime wrappers.
- TensorFlow/Keras & PyTorch Models: Supported by scikit-learn-like wrappers.
- Tabular Data: Pandas DataFrames or NumPy arrays are the primary input formats.
- CSV, JSON, Parquet: For ingestion, with conversion to DataFrame/NumPy as needed.
Architecture
ALT-AI comprises several components:
- Model Adapter: Adapts various types of ML models to a standardized format.
- Explanation Generator: Generates explanations, feature importances, and comparisons.
- Results Processor: Organizes explanation results for easy consumption.
- Security Layer: Ensures privacy and security during explanation processes.
(See classDiagram-v1.1.png
for a class diagram, sequenceDiagram-v1.1.png
for dynamic behavior.)
Configuration and Deployment Settings
- Configuration options: model type, explanation type (global/local), resource allocation.
- Logging: includes process tracking, error handling, and performance metrics.
- Usage limits: possible maximum on features, records, or model complexity.
Third Party Components & Licenses
- Pandas/Numpy: BSD-3-Clause
- scikit-learn: BSD-3-Clause
- TensorFlow: Apache-2.0
- PyTorch: BSD-style
- ONNX & onnxruntime: MIT License
Implementation Details
Built with flexibility, compliance, and scalability in mind. Integration feasibility with the Decentralized AI Training BB will be assessed in Q1 2025, after which a detailed roadmap will be provided.
Partners & Roles
- Prometheus-X Organization: Governance and infrastructure frameworks.
- AffectLog: Develops and maintains ALT-AI.
- Data Providers & Model Developers: Supply data and models plus Data/Model Cards.
- End Users: Data scientists, analysts, regulators for interpretability insights.
Usage In The Dataspace
- Interoperability: Standard formats and documentation templates.
- Data Governance: If integrated with Decentralized AI Training BB, leverages privacy-preserving infrastructure.
- Scalability & Regulatory Readiness: Handles large datasets, monitors compliance with regulations (GDPR, EU AI Act).
Leveraging AffectLog for Organizational Skill Gap Analysis
ALT-AI can interpret models for skill gap analysis, clarifying key features driving skill shortages and verifying fairness. If combined with decentralized training, privacy is enhanced.
OpenAPI Specification
Future iterations may provide an OpenAPI spec for model submission, explanation retrieval, and compliance reporting.
Test Specification
Newly Added Test Definitions
The following table covers the core endpoints tested for acceptance, focusing on verifying correct functionality of the /predict
and /explain
routes. These definitions are technology independent; they can be executed manually (e.g., using curl
) or via the provided unittest
harness.
Test Case |
Test Description |
Prerequisites |
Inputs |
Expected Outcome |
Test #1 |
/predict – 14 Features Ensures correct binary classification for exactly 14 features. |
1. ALT-AI app running (local or Docker). 2. Model preloaded (the EBM class). 3. No authentication required. |
JSON with 14 feature values, e.g.:
<br>{<br> \"features\": [30, \"State-gov\", 141297, \"Bachelors\", 13, \"Married-civ-spouse\", \"Prof-specialty\", \"Husband\", \"Asian-Pac-Islander\", \"Male\", 0, 0, 40, \"India\"]<br>}<br> |
1. HTTP 200 response. 2. Body: {\"prediction\": [\"some_class\"]} (a list with a single item). 3. Verifiable by checking the JSON structure and ensuring only one prediction is returned (e.g. \">50K\" or 0/1 ). |
Test #2 |
/predict – 15 Features Verifies extra feature is discarded, leaving 14 for the model. |
1. ALT-AI app running. 2. Same environment as Test #1. 3. The code is expected to log a warning upon receiving 15 features. |
JSON with 15 feature values, the 15th typically \">50K\" , e.g.:
<br>{<br> \"features\": [30, \"State-gov\", 141297, \"Bachelors\", 13, \"Married-civ-spouse\", \"Prof-specialty\", \"Husband\", \"Asian-Pac-Islander\", \"Male\", 0, 0, 40, \"India\", \">50K\"]<br>}<br> |
1. HTTP 200 response. 2. Body: {\"prediction\": [\"some_class\"]} with a single item. 3. Console (expected) to warn “Received 15 features; ... removing it.” 4. Verifiable by checking that the model used only the first 14. |
Test #3 |
/explain Verifies the returned global explanation includes names and scores . |
1. ALT-AI app running. 2. Model is trained in memory. |
No input (GET). |
1. HTTP 200 response. 2. JSON body includes {\"explanation\": {\"names\": [...], \"scores\": [...]}} . 3. Verifiable by checking if names and scores are arrays of length > 0. |
Manual Execution Example
-
Set up the ALT-AI application:
pip install -r requirements.txt
python app.py
(listens on port 5002)
-
Test #1 (14 features):
curl -X POST http://localhost:5002/predict \
-H "Content-Type: application/json" \
-d '{"features":[30,"State-gov",141297,"Bachelors",13,"Married-civ-spouse","Prof-specialty","Husband","Asian-Pac-Islander","Male",0,0,40,"India"]}'
- Expect
{"prediction":["..."]}
with one item.
-
Test #2 (15 features):
curl -X POST http://localhost:5002/predict \
-H "Content-Type: application/json" \
-d '{"features":[30,"State-gov",141297,"Bachelors",13,"Married-civ-spouse","Prof-specialty","Husband","Asian-Pac-Islander","Male",0,0,40,"India",">50K"]}'
- Expect HTTP 200 and a single‐item prediction.
- Also expect a console warning “Received 15 features...”.
-
Test #3 (/explain
):
curl -X GET http://localhost:5002/explain
- Expect
{"explanation":{"names":[...],"scores":[...]}}
.
Disclaimers
- Integration with the Decentralized AI Training BB will be revisited in Q1 2025 after feasibility discussions.
- The above tests assume the system is running locally and that standard Python logging is enabled to display warnings.
- Additional unit tests may be needed for broader coverage (e.g., error handling, large dataset performance).