[edit]
CardAIc-Agents: A Multimodal Framework with Hierarchical Adaptation for Cardiac Care Support
Proceedings of The 9th International Conference on Medical Imaging with Deep Learning, PMLR 315:2143-2170, 2026.
Abstract
Cardiovascular diseases (CVDs) remain the foremost cause of mortality worldwide, a burden worsened by a severe deficit of healthcare workers. Artificial intelligence (AI) agents have shown potential to alleviate this gap through automated detection and proactive screening, yet their clinical application remains limited by: (1) rigid sequential workflows, whereas clinical care often requires adaptive reasoning that selects specific tests and, based on their results, guides personalised next steps; (2) reliance solely on intrinsic model capabilities to perform role assignment without domain-specific tool support; (3) general and static knowledge bases without continuous learning capability; and (4) fixed unimodal or bimodal inputs and lack of on-demand visual outputs when clinicians require visual clarification. In response, a multimodal framework, CardAIc-Agents, is proposed to augment models with external tools and adaptively support diverse cardiac tasks. First, a CardiacRAG agent generates task-aware plans from updatable cardiac knowledge, while the Chief agent integrates tools to autonomously execute these plans and deliver decisions. Second, to enable adaptive and case-specific customization, a stepwise update strategy is developed to dynamically refine plans based on preceding execution results, once the task is assessed as complex. Third, a multidisciplinary discussion team is proposed which is automatically invoked to interpret challenging cases, thereby supporting further adaptation. In addition, visual review panels are provided to assist validation when clinicians raise concerns. Experiments across three datasets showed the efficiency of CardAIc-Agents compared to mainstream Vision–Language Models (VLMs) and state-of-the-art agentic systems.