Paper List
-
STAR-GO: Improving Protein Function Prediction by Learning to Hierarchically Integrate Ontology-Informed Semantic Embeddings
This paper addresses the core challenge of generalizing protein function prediction to unseen or newly introduced Gene Ontology (GO) terms by overcomi...
-
Incorporating indel channels into average-case analysis of seed-chain-extend
This paper addresses the core pain point of bridging the theoretical gap for the widely used seed-chain-extend heuristic by providing the first rigoro...
-
Competition, stability, and functionality in excitatory-inhibitory neural circuits
This paper addresses the core challenge of extending interpretable energy-based frameworks to biologically realistic asymmetric neural networks, where...
-
Enhancing Clinical Note Generation with ICD-10, Clinical Ontology Knowledge Graphs, and Chain-of-Thought Prompting Using GPT-4
This paper addresses the core challenge of generating accurate and clinically relevant patient notes from sparse inputs (ICD codes and basic demograph...
-
Hypothesis-Based Particle Detection for Accurate Nanoparticle Counting and Digital Diagnostics
This paper addresses the core challenge of achieving accurate, interpretable, and training-free nanoparticle counting in digital diagnostic assays, wh...
-
MCP-AI: Protocol-Driven Intelligence Framework for Autonomous Reasoning in Healthcare
This paper addresses the critical gap in healthcare AI systems that lack contextual reasoning, long-term state management, and verifiable workflows by...
-
Model Gateway: Model Management Platform for Model-Driven Drug Discovery
This paper addresses the critical bottleneck of fragmented, ad-hoc model management in pharmaceutical research by providing a centralized, scalable ML...
-
Tree Thinking in the Genomic Era: Unifying Models Across Cells, Populations, and Species
This paper addresses the fragmentation of tree-based inference methods across biological scales by identifying shared algorithmic principles and stati...
An AI Implementation Science Study to Improve Trustworthy Data in a Large Healthcare System
Georgia Institute of Technology, Atlanta, GA, USA | Shriners Hospitals for Children, Tampa, FL, USA
The 30-Second View
IN SHORT: This paper addresses the critical gap between theoretical AI research and real-world clinical implementation by providing a practical framework for assessing and improving healthcare data quality using trustworthy AI principles.
Innovation (TL;DR)
- Methodology Developed a Python-based extension of OHDSI's Data Quality Dashboard (DQD) that integrates the METRIC framework for trustworthy AI assessment, addressing informative missingness, timeliness, and distribution consistency.
- Methodology Implemented a real-world case study modernizing a large pediatric healthcare system's Research Data Warehouse from OMOP CDM v5.1/5.2 to v5.4 within Microsoft Fabric, achieving 4% improvement in data quality test success rate (84.78% to 88.88%).
- Biology Demonstrated that data harmonization using OMOP CDM concept codes does not significantly impact AI model performance (mean AUROC: 71.3% with source codes vs. 70.0% with OMOP codes) while increasing interoperability for Craniofacial Microsomia case study.
Key conclusions
- Modernizing SC's OMOP CDM database from v5.1/5.2 to v5.4 improved overall data quality by 4% (84.78% to 88.88% success rate) and conformance by 8% (80.73% to 88.09%).
- Data harmonization using OMOP CDM concept codes maintained comparable AI model performance (mean AUROC difference: 1.3%) while enabling better interoperability across healthcare systems.
- Only 50% of ICD-9 codes shared common mappings with ICD-10 codes, revealing significant vocabulary transition challenges that could degrade AI model performance when encountering mixed coding systems.
Abstract: The rapid growth of Artificial Intelligence (AI) in healthcare has sparked interest in Trustworthy AI and AI Implementation Science, both of which are essential for accelerating clinical adoption. Yet, barriers such as strict regulations, gaps between research and clinical settings, and challenges in evaluating AI systems hinder real-world implementation. This study presents an AI implementation case study within Shriners Children’s (SC), a large multisite pediatric system, showcasing the modernization of SC’s Research Data Warehouse (RDW) to OMOP CDM v5.4 within a secure Microsoft Fabric environment. We introduce a Python-based data quality assessment tool compatible with SC’s infrastructure, an extension of OHDSI’s R/Java-based Data Quality Dashboard (DQD) that integrates Trustworthy AI principles using the METRIC framework. This extension enhances data quality evaluation by addressing informative missingness, redundancy, timeliness, and distributional consistency. We also compare systematic and case-specific AI implementation strategies for Craniofacial Microsomia (CFM) using the FHIR standard. Our contributions include a real-world evaluation of AI implementations, integration of Trustworthy AI in data quality assessment, and evidence-based insights into hybrid implementation strategies, highlighting the need to blend systematic infrastructure with use-case-driven approaches to advance AI in healthcare.