In today’s cloud-native and AI-driven enterprise landscape, system failures are no longer caused by simple outages but by complex interactions between microservices, automation, and machine-learning models. To understand how explainable AI can transform reliability engineering, we spoke with Adithya Jakkaraju who authored the IEEE International Conference on Advances in Next-Generation Computer Science (ICANCS) 2025 Best Paper, “Explainable AI for Resilient Microservices: A Transparency-Driven Approach,” which presents a practical framework for building trustworthy, auditable AI-driven resilience in large-scale systems.

Q: Can you summarize the core idea behind your research?

Adithya: The central idea of the paper is that AI-driven resilience systems fail not because they lack intelligence, but because they lack transparency. Modern microservices platforms increasingly rely on AI for anomaly detection, predictive scaling, and automated recovery. However, these decisions often operate as black boxes. When incidents occur, engineers are left without clarity on why an action was taken. This research introduces a Transparency-Driven Resilience Framework that embeds explainable AI directly into the resilience lifecycle so every AI-driven decision is interpretable, auditable, and operationally actionable.

Q: What specific problems do black-box AI systems create in production environments?

Adithya: Black-box AI introduces three major problems during high-severity incidents:

  1. Unclear causality: Engineers cannot determine which service or metric triggered an action.
  2. Delayed root cause analysis: Time is lost validating whether an AI decision was correct.
  3. Reduced trust: Teams hesitate to rely on automation when they cannot explain it to stakeholders or regulators.

In large microservices environments, these issues compound quickly, leading to cascading failures and longer recovery times.

Q: How does your framework address these challenges?

Adithya: The framework integrates explainability as a first-class architectural requirement. It maps specific explainable AI techniques to resilience scenarios such as anomaly detection, failure propagation, and predictive scaling.

For example:

This ensures that every AI action is accompanied by a clear and technically grounded explanation.

Q: Was this approach validated with real system data?

Adithya: Yes. The framework was validated using a production-like microservices environment with over 38 services deployed across Kubernetes clusters. Faults such as latency spikes, memory leaks, and cascading dependency failures were intentionally injected.

The results showed:

These results demonstrate that transparency directly improves operational outcomes.

Q: Many engineers worry that explainability adds performance overhead. How does your work address this?

Adithya: That concern is valid. The study measured computational overhead carefully. Real-time explanations introduced approximately 15–20% additional compute cost, primarily due to SHAP calculations. However, this trade-off was justified by the substantial reductions in downtime and escalation rates. The framework also supports tiered explainability, using lightweight explanations for routine events and deeper analysis only during critical incidents, keeping overhead controlled.

Q: How does this research translate to regulated industries like finance and insurance?

Adithya: Regulated industries require not only resilience, but accountability. AI systems must explain their decisions to auditors, regulators, and executive stakeholders. By producing cryptographically auditable explanation logs and trace-aligned diagnostics, the framework enables organizations to meet governance requirements while still benefiting from automation. This is especially critical in financial services, where unexplained system behavior can have regulatory and economic consequences.

Q: Did the explainability layer change how engineers interacted with incidents?

Adithya: Yes, significantly. In controlled evaluations with site reliability engineers, explainable diagnostics reduced uncertainty during outages. Engineers were able to identify root causes faster and make confident remediation decisions without second-guessing the AI. Incident resolution confidence scores increased from 3.1 to 4.6 out of 5, and escalation tickets dropped by nearly 40% in complex failure scenarios.

Q: What makes this work different from existing AIOps approaches?

Adithya: Great question. Most AIOps solutions focus on prediction accuracy but ignore interpretability. This work treats explainability as a resilience property, not a visualization afterthought. It provides architectural patterns, performance benchmarks, and measurable outcomes that show how explainable AI can be deployed safely at scale, rather than remaining a research concept.

Q: What is the broader takeaway for system architects and engineering leaders?

Adithya: The key takeaway is that reliable AI systems must be understandable systems. Automation without transparency increases risk rather than reducing it. By embedding explainability into AI-driven resilience, organizations can achieve faster recovery, fewer escalations, and greater trust in autonomous systems. Transparency is not a cost; it is a force multiplier for reliability.

Q: Last question - What’s next for this area of research?

Adithya: Future work will focus on cross-cloud explainability, reinforcement learning transparency, and standardizing explanation formats for enterprise observability tools. As AI becomes more deeply embedded into critical infrastructure, explainability will be essential for building systems that are not only intelligent, but dependable.

This story was published under HackerNoon’s Business Blogging Program.