LymphAware is a domain-aware bias disruption framework designed to improve the reliability, robustness, and clinical relevance of AI systems for lymphoma histopathology diagnosis.
Modern medical AI models often achieve high accuracy by exploiting non-biological shortcuts β such as stain color, scanner signatures, or slide artifacts β instead of true pathological morphology. While effective in-domain, these shortcuts lead to fragile performance under cross-center variability, which is unacceptable for clinical deployment.
LymphAware explicitly addresses this challenge by separating morphology-relevant signals from shortcut-driven acquisition factors, enabling models to βthink more like pathologists.β π§ π¬
πΉ Tri-Path Morphology Purification Architecture
πΉ Artifact-Shift Counterfactual Training
πΉ Domain-Aware Robustness Without Explicit Labels
Models trained without LymphAware rely heavily on stain tone, background artifacts, and acquisition noise. With LymphAware, attention shifts toward diagnostically meaningful lymphoid morphology.
Across five independent medical centers:
β Higher AUC β Lower false positive rates β Reduced variance across backbones β Stronger causal consistency metrics
This work has been peer-reviewed and accepted for publication in IEEE Access, highlighting its contribution to reliable medical AI research.
π Official IEEE Xplore Link
https://ieeexplore.ieee.org/document/11408775
π DOI:
10.1109/ACCESS.2026.3667575
Medical AI systems must be:
LymphAware moves the field closer to trustworthy computational pathology by addressing shortcut bias at the representation level, rather than relying solely on dataset curation or domain labels.
We provide a clean, reproducible PyTorch pipeline located in the src/ directory for training LymphAware across multi-center lymphoma datasets.
The framework is backbone-agnostic and supports:
LymphAware/
β
βββ src/
β βββ train_lymphaware.py
β βββ models/
β βββ datasets/
β βββ losses/
β βββ utils/
β
βββ data/
β βββ CenterA/
β βββ CenterB/
β βββ CenterC/
β βββ CenterD/
β βββ CenterE/
β
βββ outputs/
Each center directory should contain class folders:
CenterA/
CLL/
FL/
MCL/
git clone https://github.com/kaopanboonyuen/LymphAware.git
cd LymphAware
conda create -n lymphaware python=3.10
conda activate lymphaware
pip install -r requirements.txt
Train on a specific center (e.g., Center A):
python src/train_lymphaware.py \
--train_dir data/CenterA/train \
--val_dir data/CenterA/test \
--backbone resnet50 \
--epochs 100 \
--batch_size 16 \
--lr 3e-4
python src/train_lymphaware.py \
--train_dir data/CenterA/train \
--val_dir data/CenterA/test \
--backbone vit_large_patch16_224 \
--epochs 100
Training artifacts will be saved to:
outputs/
best_model.pth
The script automatically:
β Tracks validation AUC β Computes False Positive Rate (FPR) β Saves the best checkpoint β Supports GPU acceleration
To reproduce the paper results:
Example loop:
for CENTER in CenterA CenterB CenterC CenterD CenterE
do
python src/train_lymphaware.py \
--train_dir data/${CENTER}/train \
--val_dir data/${CENTER}/test \
--backbone vit_large_patch16_224
done
For best performance reported in IEEE Access:
Unlike standard pipelines, LymphAware training:
The model learns cancer morphology β not acquisition artifacts.
If you find this work useful, please β star the repository.
This research is supported by:
π Talent Scholarship for Exceptional Ability π« College of Computing, Khon Kaen University
LymphAware learns the cancer β not the confounders.
By enforcing morphology-grounded representations and suppressing shortcut bias, we aim to build AI systems that clinicians can truly trust.
β If you find this project useful, please consider starring the repository!
@article{panboonyuen2026lymphaware,
author = {Teerapong Panboonyuen},
title = {LymphAware: Domain-Aware Bias Disruption for Reliable Lymphoma Cancer AI Diagnosis},
journal = {IEEE Access},
year = {2026},
pages = {1--1},
doi = {10.1109/ACCESS.2026.3667575},
publisher = {IEEE}
}
If you use this work in your research, please cite the official IEEE version via the DOI above.