Lisa Anderson, CEO
Former policy director at White House OSTP, co-author of the AI Bill of Rights
Making AI Understandable and Trustworthy
AIClarum is an AI transparency and compliance automation platform founded in Austin, Texas. We believe organizations should be able to explain every AI decision — clearly, accurately, and in language regulators and end users can trust.
Our platform integrates SHAP and LIME explainability directly into your existing AI workflows, automating compliance documentation for the EU AI Act, ISO 42001, and NIST AI RMF frameworks.
EU AI Act Ready
Pre-built compliance templates for every risk tier
Plain-Language Explanations
AI decisions in language anyone can understand
Six Reasons Organizations Choose Us
From explainability to audit trails, AIClarum delivers end-to-end AI transparency.
SHAP & LIME Explainability
Industry-leading AI decision explainability with SHAP and LIME integration — understand which features drive every model prediction.
Automated Compliance Auditing
Automated compliance auditing for EU AI Act, ISO 42001, and NIST AI RMF. Generate documentation at the push of a button.
Real-Time Bias Monitoring
Real-time model monitoring detects bias drift and fairness degradation before they become compliance violations or reputational risks.
Plain-Language Engine
Our plain-language explanation engine makes AI decisions understandable to non-technical stakeholders, regulators, and customers.
Comprehensive Audit Trail
A comprehensive audit trail for every AI decision, accessible to internal teams and external regulators in structured, exportable formats.
Industry Compliance Templates
Pre-built compliance templates for healthcare, finance, and HR AI use cases — reduce implementation time from months to days.
Transparency Tools for Every AI Use Case
SHAP and LIME at Production Scale
AIClarum integrates natively with scikit-learn, PyTorch, TensorFlow, and XGBoost. Our explainability engine processes thousands of predictions per minute and stores feature attribution data for retrospective audits.
- Global and local SHAP explanations
- LIME-based perturbation analysis
- Counterfactual explanation generation
From Model Card to Audit Report
Our compliance automation layer maps your model's characteristics to regulatory requirements automatically. For each framework, AIClarum pre-fills compliance questionnaires using live telemetry from your production models.
- EU AI Act Article 9–15 documentation
- ISO 42001 control mapping
- NIST AI RMF Govern–Manage cycle support
Catch Bias Before It Becomes a Liability
AIClarum's monitor tracks demographic parity, equalized odds, and calibration metrics across protected attributes. Alerts trigger when fairness thresholds breach configured tolerance bands.
- Demographic parity and equalized odds
- Data drift detection with PSI and KL divergence
- Automated remediation recommendations
Built by Experts in AI Policy and Engineering
Lisa Anderson
Chief Executive Officer
Former policy director at White House OSTP, co-authored AI Bill of Rights. JD Yale, MS CS Stanford.
Thomas Berg
Chief Technology Officer
Former IBM Research XAI scientist. Developed LIME/SHAP integrations for 2,000+ data science teams.
Diana Okonkwo
Chief Compliance Officer
15 years AI regulation, drafted EU AI Act guidelines. Expert in GDPR, ISO 42001, NIST AI RMF.
Ready to Make Your AI Explainable?
Join the organizations trusting AIClarum to keep their AI systems transparent, fair, and compliant.
Request a Demo View Solutions