AI Medical Imaging
Diagnostic Platform
with CNN & Explainable AI

This is the production AI medical imaging diagnostic platform our team delivered as a Platform-as-a-Service — AI X-ray analysis built on convolutional neural networks that identify, classify and localize abnormal patterns, with Grad-CAM explainability overlays that give clinicians a visual justification for every prediction. Specialists work through very high image volumes under constant time pressure, and subtle pathologies can still be missed after hours of back-to-back reads. We built an alternative: HIPAA-aligned, cloud-native, and continuously learning.

  • Deep learning backbone — ResNet, DenseNet, and EfficientNet fine-tuned on curated medical datasets
  • Detection & localization — Faster R-CNN and YOLO variants pinpoint regions of interest on each image
  • Explainable AI — Grad-CAM heatmaps make every inference auditable for clinicians and regulators
  • PaaS architecture — HIPAA- and GDPR-aligned pipeline with continuous learning from expert feedback

The Problem

Medical imaging sits at the center of modern diagnostics — X-rays, CT, MRI and bitewing dental radiographs drive a significant share of clinical decisions. The specialists reading those images are under constant time pressure: high case volumes, short interpretation windows, and subtle visual signals that can be easy to miss after the tenth similar study in a row. Inter-reader variability compounds the problem — two qualified radiologists can disagree on borderline findings, and that variability eventually shows up in claims, second opinions and patient outcomes.

Our client — a major dental insurance provider — needed to review enormous volumes of submitted bitewing X-rays attached to claims. Human review was expensive, slow, and inconsistent, and the back-office team was drowning. The brief was to build a system that could screen every submitted image automatically, flag suspected pathologies such as caries and periapical lesions, show the clinician exactly where the AI saw the issue, and do all of it through a HIPAA-aligned pipeline that handled PHI safely from upload to archive. A black-box classifier was never an option — explainability was a hard requirement from day one.

Solution Architecture

The platform separates concerns into three layers: a clinical-facing portal (secure upload, review, and feedback), an AI inference engine (preprocessing → CNN classification → detection → Grad-CAM), and a HIPAA-aligned data plane (encrypted storage, audit logs, continuous-learning feedback loop).

CLINICAL PORTAL AI INFERENCE ENGINE DATA PLANE C Clinician WEB PORTAL Upload X-ray Image Review Prediction View Grad-CAM Confirm / Correct SSO · TLS · RBAC AI PIPELINE 1. Image Preprocessing denoise · normalize · augment 2. CNN Feature Extraction ResNet · DenseNet · EfficientNet 3. Detection & Localization Faster R-CNN · YOLO 4. Grad-CAM Explainability class-activation heatmap 5. Result Assembly labels · bboxes · heatmap · confidence ENCRYPTED STORE PHI Images · Metadata AES-256 at rest MODEL REGISTRY Versioned Weights A/B rollout AUDIT LOG Access · Predictions continuous learning feedback
Fig. 1 — AI medical imaging diagnostic platform: portal → CNN pipeline → Grad-CAM → HIPAA-aligned data plane with continuous-learning feedback

Solution Components

Clinical Web Portal
Secure upload, case queue, side-by-side original and Grad-CAM overlay viewer, and a one-click confirm / correct workflow that feeds the training loop.
Image Preprocessing
Denoising, histogram normalization, resizing and on-the-fly augmentation (rotation, flip, contrast). Clean input stabilizes CNN confidence across devices and exposures.
CNN Feature Extraction
Backbones built on ResNet, DenseNet and EfficientNet, fine-tuned via transfer learning on curated, expert-labeled medical datasets.
Detection & Localization
Faster R-CNN and YOLO variants produce bounding boxes around regions of interest so clinicians see where the model thinks the pathology is, not just a label.
Grad-CAM Explainability
Class-activation heatmaps render on top of the original X-ray, giving a visual justification for every prediction — auditable for clinicians, compliance and regulators.
HIPAA & GDPR Data Plane
TLS in transit, AES-256 at rest, role-based access, full audit trail, and a versioned model registry with A/B rollout for safe continuous deployment.

What We Learned Building This

Preprocessing earns more accuracy than any model swap

Before we touched a network architecture, we spent serious time on image preprocessing. Medical X-rays come in wildly different exposures, resolutions and compressions depending on the sensor and the clinic. Denoising, histogram normalization and standardized resizing brought the signal-to-noise ratio up to a point where even a modest backbone began producing trustworthy confidence scores. Augmentation — rotation, flips, brightness and contrast jitter — expanded the training distribution and measurably reduced overfitting on the dental datasets we started with. The biggest lesson: every hour spent on input quality paid off more than an equivalent hour spent hyperparameter-tuning a bigger model.

Transfer learning beats training from scratch — every time

Labeled medical imaging data is scarce, expensive and bound by strict privacy rules. Training a CNN from scratch on a few thousand bitewing X-rays would have produced a brittle model. Instead, we started from backbones pretrained on ImageNet and fine-tuned the upper layers on the client’s curated dataset. ResNet gave us a strong, stable baseline; DenseNet pushed accuracy on subtle lesions thanks to its feature-reuse topology; EfficientNet delivered the best accuracy-per-inference-millisecond trade-off once we needed to run at scale. We kept all three in the model registry and routed traffic based on case type.

Localization matters more than classification for clinician trust

A classifier that says “caries: 0.87” is almost useless to a working clinician. They need to see where. We layered Faster R-CNN and YOLO-style detectors on top of the CNN backbone to draw bounding boxes around candidate regions. The combination of a classifier head plus a detection head turned out to be the point where clinicians actually started trusting the tool. Interestingly, a weaker classifier with good localization beat a stronger classifier with no localization in every blind usability test we ran.

Explainability is not optional — Grad-CAM earned its keep

The requirement for explainable AI was not a nice-to-have imposed by compliance — it was the feature clinicians asked for first. We implemented Grad-CAM (Gradient-weighted Class Activation Mapping) to produce a heatmap that shows which pixels drove the prediction. The overlay is rendered in the portal directly on top of the original X-ray so the reviewer can instantly judge whether the model is looking at a real finding or an artefact of the film edge. Grad-CAM also became our primary debugging tool: whenever the model misfired in staging, the heatmap usually told us exactly why — the model was latching onto film markers, anatomical overlap, or compression artefacts rather than the pathology itself.

HIPAA-aligned privacy shaped the architecture, not the wrapper

HIPAA and GDPR alignment was treated as an architectural constraint from day one, not a last-mile checklist. Every network hop is encrypted (TLS 1.2+). Images and metadata are encrypted at rest with AES-256. Access is role-based and every request — upload, inference, correction, export — is written to an immutable audit log. Where federated learning was feasible, we used it so that sensitive imagery could contribute to model improvement without ever leaving the client’s secure environment. The result is a platform where privacy and AI quality reinforce each other rather than compete.

Continuous learning only works with good feedback plumbing

Clinician corrections are the most valuable training signal a medical AI platform can collect. We built the “confirm / correct” action straight into the review screen, and every correction flows into a labeled queue that feeds the next retraining cycle. New model versions are registered, shadow-tested against a hold-out set, and then rolled out through an A/B mechanism backed by the model registry. Model performance on the client’s validation set improved measurably after each retraining pass — not because the architecture changed, but because the feedback loop was short and frictionless.

Model drift monitoring — the part most write-ups skip

A model that was accurate on the validation set six months ago is not automatically accurate today. Sensor firmware updates, new clinic workflows, different patient demographics — all of it can shift the input distribution in ways the model was never trained on. We instrumented the platform to track confidence distributions, class-frequency drift and Grad-CAM pattern stability over rolling windows, with alerts that trip long before accuracy degrades noticeably on the validation set. Continuous learning without continuous monitoring is a trap; this piece of plumbing is what keeps the system safe to leave in production.

Why this reads differently from most AI-in-radiology write-ups

Most content on this topic stays at the “AI can do radiology” level — what the technology might do, not what actually happens when you ship it to real clinicians on real images. We shared the preprocessing pipeline, the specific backbones, the Grad-CAM overlay choice, the HIPAA-aligned architecture, the drift monitoring and the continuous-learning loop because these are the details that matter when a platform goes into production. This architecture has been delivered for a major dental insurance client and is the pattern we reuse whenever the brief involves trustworthy, explainable AI on sensitive images. At DMexec, we don’t just consult on AI platforms — we provide the dedicated AI, data science and platform engineers who design, build and maintain them. The platform is assistive, not a standalone diagnostic device — regulatory classification (including the applicable FDA 510(k) or EU MDR pathway) is agreed with each client based on clinical intended use.

Related use case: Offline Field Data Capture with D365 F&O & Power Platform — another production-tested architecture from the DMexec team.

Who Needs This

An AI-assisted radiology workflow fits any specialty where clinicians review high volumes of images under time pressure and where an explainable, HIPAA-aligned second reader earns its keep.

Dental Insurance
Automated screening of bitewing X-rays attached to claims — caries, restorations, periapical lesions.
Radiology
Triage assist for chest X-rays and CT studies with visual localization and explainable overlays.
Orthopedics
Fracture detection, implant alignment checks and subtle cortical-break localization on long-bone studies.
Oncology & Cardiology
Lesion and nodule detection, cardiac silhouette analysis, and second-reader workflows where explainability is mandatory.

FAQ

A cloud-based platform that ingests medical images (X-ray, CT, MRI, dental radiographs), runs them through a deep-learning pipeline, and returns labeled predictions with localization and visual explanations. It sits alongside existing clinical workflows as an assistive tool for radiologists and specialists, not a replacement.

For feature extraction and classification we used CNN backbones — ResNet, DenseNet and EfficientNet — fine-tuned via transfer learning on curated medical datasets. For detection and localization we layered Faster R-CNN and YOLO variants on top to produce bounding boxes around regions of interest.

Every prediction is accompanied by a Grad-CAM heatmap rendered directly over the original X-ray. The overlay highlights the pixels that most influenced the model’s decision, so the clinician can instantly judge whether the AI is focused on a real pathology or on an irrelevant artefact.

Yes. The architecture is HIPAA- and GDPR-aligned: TLS 1.2+ in transit, AES-256 encryption at rest, role-based access control, immutable audit logs for every access and prediction, and — where applicable — federated learning so that sensitive images can contribute to model improvement without leaving the client’s secure environment.

Three techniques work together. Transfer learning from large general-purpose datasets gives the backbones a strong starting point. Data augmentation expands the effective training set. Continuous learning from clinician corrections keeps improving the model after deployment, turning every correction into a new labeled example.

The clinical portal has a built-in confirm / correct action on every prediction. Those corrections flow into a labeled queue and feed the next retraining cycle. New model versions are registered, shadow-tested against a hold-out set, and rolled out through an A/B mechanism so accuracy improves continuously without risking production quality.

Yes. The same architecture — preprocessing, CNN backbones, detection heads, Grad-CAM explainability, HIPAA-aligned data plane — maps directly to chest X-ray triage, fracture detection in orthopedics, lesion and nodule detection in oncology, and cardiac imaging. Any specialty where clinicians review images under time pressure is a fit.

No. The platform is assistive — it screens, flags and explains, but the clinician is always in the loop and always makes the final call. The goal is to help reduce the chance of missed findings, standardize quality across reviewers, and free up specialist time for the cases that actually need human judgement.

Accuracy depends on case mix, imaging modality and the specific pathology being flagged. Every deployment is validated against a client-specific hold-out set labeled by their own specialists before go-live, and monitored for drift afterwards. We publish headline metrics only in client-owned reports — not as generic marketing claims — because a number that is true on one dataset can be misleading on another.

Need a Team That
Builds This?

This platform was delivered by our AI & data science team. If you’re evaluating a similar diagnostic solution, we’re happy to walk through the architecture.

Book a Technical Call
Made on
Tilda