Medical Imaging Classification via MobileNetV3
Early detection of retinal abnormalities prevents permanent blindness. Automated edge-deployable screening democratizes access.
Rural patients, remote diagnostic clinics, and overburdened ophthalmologists.
N/A (Jupyter Notebook / Python Scripts for Research environment)
End-to-end PyTorch pipeline: Local data generators -> Augmentation -> Mixed Precision (FP16) Forward Pass -> Grad-CAM interpretability overlay.
Chose MobileNetV3Large over heavy models like ResNet152 specifically to support future Edge-device deployment without massive VRAM.
Tested ResNet50 and EfficientNet. EfficientNet had marginally better AUC but 3x the inference latency, violating edge-compute constraints.
MobileNetV3Large (ImageNet Pretrained)
Two-stage transfer learning: Frozen feature extractor (10 epochs) followed by unfrozen end-to-end fine-tuning at an aggressively lower learning rate (1e-5).
84,000+ Optical Coherence Tomography (OCT) images (Kermany et al.) split into 4 classes: CNV, DME, Drusen, Normal.
Efficiently distinguishes between Normal, Choroidal Neovascularization, Diabetic Macular Edema, and Drusen.
Generates interpretable visual heatmaps highlighting the exact retinal topology causing the prediction.
Mixed precision training and native lightweight convolutions make the model directly exportable to ONNX for mobile.
Proves viability for low-cost, automated medical screening in regions lacking immediate specialized care.
Mastered PyTorch optimization loops. Learned how crucial Explainable AI (XAI) is in the medical domain; accuracy alone does not drive adoption.
MobileNet inherently trades a fraction of a percent of accuracy for massive gains in latency and compute efficiency.
Incorporate Vision Transformers (ViT) to see if global attention mechanisms outperform spatial convolutions on this specific OCT texture.
Let's talk about how I can build something similar for your team.