Data Science / Artificial Intelligence
Breaking open the Agentic AI “Black Box”
Large Language Models (LLMs) are transforming processes with capabilities in natural language understanding and generation.
Data Science / Artificial Intelligence
Large Language Models (LLMs) are transforming processes with capabilities in natural language understanding and generation.
Now, new “Agentic AI”, or "AI agents", systems designed for goal-directed autonomy, promise to tackle complex tasks with minimal human supervision and transform how software and automation are conceived. [1],[2],[3],[4],[5]
While the hype around fully autonomous AI is strong, the reality is tempered by technical limitations, data quality issues, and compliance needs. The core tension lies in balancing the value of autonomy by offloading complex tasks and achieving results at scale with the inherent risks of “black box” decision-making. The probabilistic nature of LLMs means outputs can be unpredictable or erroneous, demanding robust guardrails. [6]
Security, privacy, and regulatory frameworks like revDSG, GDPR and the EU AI Act further necessitate human oversight and traceability, especially in critical sectors. [7],[8],[9] True value from Agentic AI, therefore, emerges when it is anchored in controlled workflows and human oversight. We explored the promises and challenges of Agentic AI in a Proof of Concept (PoC) involving a joint team of Swisscom's Data C AI Consulting and Noumena Digital engineers(opens in new tab).
Agentic AI systems represent a shift from reactive AI models to proactive systems capable of multi-step workflows and decision-making to achieve goals. These LLM-powered autonomous programs perceive their environment, plan, invoke tools e.g. through APIs and Model Context Protocol (MCP) servers, and interact with digital surroundings and other agents, using e.g.Agent-to-Agent (A2A) implementations. [10]
Agentic AI operates on a continuous iterative cycle, described as a Perception-Reasoning- Action-Feedback loop or, similarly, a Thought-Action-Observation cycle: [1],[10],[11],[12],[26]
1. Perception (Observation): The system ingests and interprets diverse data, structured (databases, APIs) and unstructured (images, emails).
2. Reasoning (Thought): LLMs or multimodal models determine the next best action based on the interpreted data and the overall goal, selecting from available tools.
3. Action: The agent executes the chosen action, such as calling an external API or an MCPserver, under predefined guardrails.
4. Feedback (Observation): The action's outcome is returned to the system, informing the next reasoning step or plan adjustment.
Humans can be integrated at any step for guidance or approval ("human-in-the-loop"). Agents are envisioned to exist on a spectrum of autonomy, from single LLM calls within human-coded workflows to highly autonomous systems that dynamically choose their actions. [4]
Agentic AI offers various benefits but also introduces significant risks that businesses mustprevent. In the following, find a list of successful use cases:
However, AI also has its prominent failure cases. In a highly publicized case in 2024, an Air Canada chatbot provided incorrect information about bereavement fare policies. The airline was held liable for it, showing that businesses can be responsible for errors made by their AI systems. [23],[24]
For the PoC, we utilized Noumena’s technology stack, which emphasizes enterprise-grade security, fine-grained access control, audit trails, and permissioned orchestration. To put this to the test, Swisscom and Noumena partnered in a proof-of-concept (PoC) project to explore secure Agentic AI in insurance claims processing.
Traditionally, claims are classified, reviewed, and adjusted. Classification directs the claim to the correct department. The review assesses the completeness, trustworthiness, and validity of the information, determining eligibility and compensation. Denied claims require clear explanations. A department-specific adjustor cross-checks reviews for quality and fairness.
AI can automate classification and review. In this PoC, an LLM classified claims and provided access to department-specific LLM reviewers. To ensure human oversight, the adjustor typically remains human. However, to illustrate the flexible implementation of policies, the PoC included a rule allowing reviewer recommendations for small amounts to bypass human adjustment.The PoC focused on integrating services seamlessly, orchestrating AI and human interactions along traceable workflows, restricting agent (whether AI or human) access to data on a need-to-know basis, and implementing transparent AI delegation policies.
Processing ffow for insurance claims as implemented in the PoC. Claims processing (health, car, household) used Al for classification and department-specific review, leveraging Claude 3.5 Sonnet via AWS Bedrock. Human adjustors handled final decisions.
This PoC demonstrated Agentic AI's practical application in automating complex tasks,specifically in:
The PoC also validated the effectiveness of Noumena’s technological approach in addressingsome key challenges of Agentic AI in the enterprise context:
Overall, the PoC successfully showcased Agentic AI's potential to reshape business processes through efficiency gains. It also showed Noumena’s technology's ability to enable secure, auditable Agentic AI systems by providing execution guardrails.
Beyond technology choices, Agentic AI projects need to follow a well-thought-out strategy to ensure success. A phased, strategic approach that embeds risk mitigation is crucial for the deployment of Agentic AI. Follow some guiding principles when implementing agentic AI PoCs and deployments:
Moreover, make sure to follow a phased rollout approach:
1. Controlled pilot and beta deployment: Limited scope, heavy oversight, with goal to validate core functionality. With a limited production and restricted user base, test of scalability and real-world performance.
2. Broad adoption: Wider rollout, incremental expansion of responsibilities, formal integration into business processes with governance.
3. Ongoing improvement G maintenance: Continuous model updates, performance reviews, and adaptation to new rules or data.
Transparency with stakeholders and fostering cultural readiness are crucial throughout. [25]
Agentic AI marks a new frontier in automation and intelligence. To realize its promise, companies must break open the black box. This makes systems auditable, understandable, and controllable through oversight and constraints as explored in the PoC explored above.
Pragmatism is paramount. Organizations should experiment boldly yet govern responsibly. They should invest in human expertise and transparency tools. AI should be seen as a powerful assistant. However, its outputs should be questioned and verified, especially in high-risk scenarios. Policymakers and industry groups must develop standards and audit frameworks. This could be similar to those in finance or safety engineering, to ensure responsible innovation. [7]
Auditability and control will enable, rather than stifle, innovation. Trustworthy, transparent, and controllable agents will be deployed more widely and boldly. "Pragmatic agency" means building agents that are as autonomous as possible within well-understood guardrails, continually illuminating their inner workings, and maintaining a human-in-command approach. This journey demands diligence. But by transforming the AI "black box" into a "glass box," we can harness its power for a future of effective human-AI collaboration.
Authors:
1. What is Agentic AI? A Practical Guide - K2view, Abgerufen am 23. Mai 2025, https://www.k2view.com/what-is-agentic-ai/(opens in new tab)
2. Agentic AI: What you need to know about AI agents | CSAIL Alliances, Abgerufen am 23. Mai 2025, https://cap.csail.mit.edu/agentic-ai-what-you-need-know-about-ai-agents(opens in new tab)
3. What is AI? - AI Tools and Resources for Biomedical Research, Abgerufen am 23. Mai 2025, https://laneguides.stanford.edu/AI/what-is-ai(opens in new tab)
4. What on Earth are Agents? Jensen Low, Abgerufen am 23. Mai 2025,
https://www.jensenlwt.com/blog/what-on-earth-are-agents/(opens in new tab)
5. Agents Simplified: What we mean in the context of AI | Weaviate, Abgerufen am 23. Mai 2025, https://weaviate.io/blog/ai-agents(opens in new tab)
6. Reducing LLM Hallucinations: A Developer's Guide
- Zep, Abgerufen am 23. Mai 2025, https://www.getzep.com/ai-agents/reducing-llm-hallucinations(opens in new tab)
7. The AI Act Explorer | EU Artificial Intelligence Act, Abgerufen am 23. Mai 2025, https://artificialintelligenceact.eu/ai-act-explorer/(opens in new tab)
8. Art. 22 GDPR – Automated individual decision-making, including ..., Abgerufen am 23. Mai 2025, https://gdpr-info.eu/art-22-gdpr/(opens in new tab)
9. FINMA Guidance 08/2024 Governance and Risk Management when ..., Abgerufen am 23. Mai 2025, https://www.mll-news.com/finma-guidance-08-2024-governance-and- risk-management-when-using-artificial-intelligence/?lang=en(opens in new tab)
10. Control Plane as a Tool: A Scalable Design Pattern for Agentic AI Systems - arXiv, Abgerufen am 23. Mai 2025, https://arxiv.org/html/2505.06817v1(opens in new tab)
11. Agentic AI: A guide to the next wave of CX innovation - PolyAI, Abgerufen am 23. Mai 2025, https://poly.ai/agentic-ai/(opens in new tab)
12. Understanding AI Agents through the Thought-Action-Observation ..., Abgerufen am 23. Mai 2025, https://huggingface.co/learn/agents-course/unit1/agent-steps-and-structure(opens in new tab)
13. Agentic AI in Insurance: Transforming Insurance with AI, Abgerufen am 23. Mai 2025, https://hexaware.com/blogs/agentic-ai-in-insurance-transforming-the-industry-with- enterprise-ai/(opens in new tab)
14. Agentic AI for Insurance | Real-Time Insights and Outcomes - XenonStack, Abgerufen am 23. Mai 2025, https://www.xenonstack.com/blog/agentic-ai-insurance-claims(opens in new tab)
15. Agentic AI: Automated Claims Processing Multi-Agent System - IDC, Abgerufen am 23. Mai 2025 https://my.idc.com/getdoc.jsp?containerId=US52880425CpageType=PRINTFRIENDLY(opens in new tab)
16. Use Cases for AI Agents C Agentic Automation | Beam AI, Abgerufen am 23. Mai 2025, https://beam.ai/articles(opens in new tab)
17. AI-based predictive maintenance - Siemens Global, Abgerufen am 23. Mai 2025, https://www.siemens.com/global/en/products/automation/topic-areas/industrial- ai/usecases/ai-based-predictive-maintenance.html(opens in new tab)
18. Senseye Predictive Maintenance - Siemens Global,
Abgerufen am 23.
Mai 2025, https://www.siemens.com/global/en/products/services/digital-enterprise- services/analytics-artificial-intelligence-services/senseye-predictive- maintenance.html(opens in new tab)
19. KION Teams with NVIDIA and Accenture to Optimize Supply Chains..., Abgerufen am23. Mai 2025, https://erp.today/kion-teams-with-nvidia-and-accenture-to-optimize- supply-chains-with-ai-powered-robots-and-digital-twins/(opens in new tab)
20. KION presents AI Control Tower at GTC in San José, California - Our current press releases | KION GROUP AG, Abgerufen am 23. Mai 2025, https://www.kiongroup.com/en/News-Stories/Press-Releases/Press-Releases- Detail.html?id=2954666(opens in new tab)
21. UK government's AI system 'Humphrey' set to review thousands of public consultations to improve civil service efficiency, Abgerufen am 23. Mai 2025, https://www.globalgovernmentforum.com/uk-governments-ai-system-humphrey-set- to-review-thousands-of-public-consultations-to-improve-civil-service-efficiency/(opens in new tab)
22. AI experiments see “Humphrey” help townhalls cut costs and ..., Abgerufen am 23. Mai 2025, https://www.gov.uk/government/news/ai-experiments-see-humphrey-help- townhalls-cut-costs-and-improve-services(opens in new tab)
23. BC Tribunal Confirms Companies Remain Liable for Information ..., Abgerufen am 23. Mai 2025, https://www.americanbar.org/groups/business_law/resources/business- law-today/2024-february/bc-tribunal-confirms-companies-remain-liable-information- provided-ai-chatbot/(opens in new tab)
24. Air Canada Held Liable For Chatbot Misinformation - One Mile at a Time, Abgerufen am23. Mai 2025, https://onemileatatime.com/news/air-canada-liable-chatbot- misinformation/(opens in new tab)
25. How COOs maximize operational impact from gen AI and agentic AI ..., Abgerufen am23. Mai 2025, https://www.mckinsey.com/capabilities/operations/our-insights/how- coos-maximize-operational-impact-from-gen-ai-and-agentic-ai(opens in new tab)
26. Microsoft Agent Framework, Abgerufen am 5. Oktober 2025, https://learn.microsoft.com/en-us/agent-framework/overview/agent-framework- overview(opens in new tab)
Senior Consultant Data & AI
Find the job or career to suit you. A career where you can make a difference and continue your personal development.
What you do is who we are.