Gallery

See the Platform in Action

Screenshots, architecture diagrams, and published figures from the MAPLES AI grading platform and the OSCE simulation environments it serves.

MAPLES Platform

The AI grading interface used by faculty and simulation staff.

MAPLES Dashboard

The MAPLES dashboard showing 46 active rubrics, 93 student groups, and 182 grading runs — a production system processing thousands of encounters.

Rubric Management

Rubric management view: a rubric with 64 prompts across audio, video, and note modalities, organized by OSCE station. Each item is mapped to a grading mode and scoring level.

Rubric Scoring Levels

Rubric level definitions: each item has clear scoring criteria (1-4) that both human raters and AI use — ensuring consistent interpretation across graders.

Grading Management

Grading configuration: select a student group, map rubrics to stations, enable optional features (localizer, mosaic videos), and launch AI grading with one click.

Video Review

Multimodal review: synchronized video playback with AI-generated transcript and item-level assessment. Faculty can accept the AI score or flag for manual review.

Post-Encounter Note Review

Post-encounter note grading: the student's clinical note alongside the AI assessment, rationale, and rubric scoring guide. Faculty review each item with full context.

Score Review & Edit

Batch review interface: 6,780 items across an entire OSCE cohort. Faculty can filter, accept AI scores, or drill into individual items for manual review.

Cost Estimates

Pipeline cost transparency: 622 encounters graded for $26.10 total ($0.04 per encounter) using Gemini 2.5 Pro — full token-level cost breakdown for institutional budgeting.

End-to-End Platform Overview

How MAPLES fits in the full workflow: (1) OSCE encounter captured via multi-angle cameras, (2) Platform — SimRubrics for rubric design, Elephant for multimedia cataloging, MAPLES for AI grading orchestration, (3) Human review with faculty adjudication and grades reported to students.

OSCE Setup & Simulation Environment

The clinical simulation environments where encounters are captured.

OSCE Encounters

What an OSCE looks like: medical students rotate through standardized patient scenarios (e.g., Itchy Eyes, Leg Pain, Memory Problems), each filmed from overhead cameras. Every encounter is graded — this is what MAPLES automates.

Platform Architecture & Published Figures

From the NEJM AI publication and technical report — the rubrics-to-prompts pipeline.

NEJM AI Workflow

The rubrics-to-prompts workflow published in NEJM AI — translating clinical rubrics into AI grading specifications.

Production Deployment

Production deployment model at UTSW showing the human-in-the-loop review pipeline.

Platform Pipeline

End-to-end grading pipeline: data ingestion, AI assessment, quality review, and feedback delivery.

System Architecture

MAPLES platform architecture showing the relationship between SimRubrics, Elephant, and the grading engine.

Multimodal AI Assessment

Multi-camera video analysis using Gemini — from the multimodal physical exam preprint (Kang et al., 2026).

Multimodal AI Framework

Multi-camera OSCE room setup with AI segmentation and grading pipeline — frontal, posterior, and overhead cameras feed into Gemini for physical examination assessment.

Multimodal Kappa Results

Cohen's Kappa across modalities: native 3-camera video achieves kappa = 0.837, demonstrating that "full video is king" — outperforming audio-only, transcript, and single-camera configurations.

Item-Level F1 Heatmap

Macro F1 scores by rubric item and modality configuration — showing how different physical exam maneuvers benefit from different camera angles and input types.

More screenshots of the MAPLES interface and OSCE simulation environments coming soon. Have questions about the platform? Get in touch.