Evidence

What the literature tells us

Model Performance

ML models consistently achieve AUC >0.80 for high-acuity ED outcomes including ICU transfer and hospital admission (Rajpurkar et al., 2022; Topol, 2019).
NLP on free-text triage notes significantly improves classification accuracy over structured data alone — the core motivation for ClinAssist's dual-model architecture.
XGBoost and gradient boosting outperform simpler models across diverse ED populations in recent systematic reviews.
Clinician trust remains the primary barrier to adoption — not model performance — underscoring why explainability is central, not optional (Lundberg & Lee, 2017).

Multimodal Integration

Co-attention transformers for multimodal clinical data (Chen et al., ICCV 2022) provide theoretical grounding for ClinAssist's unified inference layer.
Attention-based NLP (Vaswani et al., 2017) forms the backbone of ClinAssist's clinical notes module, enabling both prediction and explanation from free text.

Open Datasets

MIMIC-IV — Beth Israel Deaconess Medical Centre. 300,000+ ICU and ED admissions. Gold standard for clinical AI model development and validation.
PhysioNet — Open repository of physiological and clinical data. Used for model development and cross-dataset validation alongside MIMIC-IV.

Regulatory Alignment

TGA guidance (Australia) — ClinAssist is designed from the ground up to align with the Therapeutic Goods Administration's AI-as-a-Medical-Device framework. Regulatory scoping is a Phase 1 deliverable.
Australian Privacy Act compliance — Data preprocessing pipelines are designed around Australian Privacy Act principles, including de-identification and consent frameworks.
IRB/Ethics approval — Formal ethics clearance for dataset access and any clinician interview or study components will be obtained in Phase 1 prior to data collection.

Research Questions

RQ1: What design principles enable AI models to provide real-time, accurate clinical predictions while maintaining explainability for non-technical end users?
RQ2: How can multimodal clinical data (structured records, imaging, time-series vitals) be integrated effectively within a single deployable pipeline?
RQ3: What validation frameworks and evaluation metrics are most appropriate for assessing AI clinical tools prior to real-world deployment?