Alzheimer’s isn’t just a medical condition I read about in textbooks—it’s something I’ve lived through in my own family. My grandmother struggled with dementia and eventually Alzheimer’s disease, and as a child, I didn’t fully understand what was happening. What I did see was the toll it took on her, and on us as a family. The gradual loss of memory, the confusion, the helplessness—it left a deep mark on me.
As I grew older, I realised that one of the greatest challenges with Alzheimer’s is late diagnosis. By the time symptoms become obvious, opportunities for meaningful intervention are often limited. Families are left reacting instead of preparing. That experience made me determined to change the story—not just for my family, but for countless others.
That determination led me to build something: an AI‑driven early detection tool. Unlike black‑box models that spit out results without explanation, this system shows its reasoning. Clinicians don’t just get a prediction; they see why the model made that call. And because it’s lightweight enough to run on low‑resource devices, even rural clinics can use it. At its heart, my vision is simple: give families more time. Time to prepare, to connect, and to live with dignity. This project isn’t just about technology—it’s about hope, equity, and impact.
Why This Matters Globally
Alzheimer’s is one of the most pressing public health challenges of our time. Right now, about 55 million people live with Alzheimer’s or related dementias worldwide. By 2030, healthcare costs are expected to exceed $1 trillion, driven by long‑term care, hospitalisations, and informal caregiving. And by 2050, cases could rise to 139 million.
Behind those numbers are real lives—families stretched thin, caregivers exhausted, healthcare systems overwhelmed. The barriers to early detection are everywhere. There simply aren’t enough specialists to diagnose early. Symptoms are often brushed off as “normal ageing.” And access to MRI interpretation and standardised screening tools is limited, especially in rural and underserved communities. The result is that families lose precious time, and healthcare systems face enormous strain.
How the AI Tool Helps
That’s where my system comes in. It’s designed to break those barriers by being practical, transparent, and globally accessible. Through a mobile and web app, clinics can upload MRI scans directly, bringing screening closer to patients. The AI backend then classifies Alzheimer’s stage (0–3) with visual explanations clinicians can trust. A multilingual interface ensures the tool works across diverse populations, while offline mode makes sure rural communities aren’t left behind.
This isn’t about replacing doctors—it’s about empowering them with tools that extend their reach. By combining accessibility with interpretability, the system makes early detection possible in places where it was previously out of reach.
Key Benefits
The benefits of this approach are clear. By staging Alzheimer’s with AI, diagnosis can happen faster and with less reliance on specialists. A mobile‑first design means screening can happen at the community level, increasing adherence and accessibility. Most importantly, early detection enables timely intervention, care planning, and improved patient outcomes.
Trust is built through transparency. Grad‑CAM overlays show clinicians what the AI “sees,” while local dashboards help district health officers track cases and allocate resources effectively. Together, these features make the tool not just powerful, but practical.
The Bigger Vision
This tool isn’t just about technology—it’s about impact. Imagine a world where diagnostic delays are dramatically reduced, where long‑term healthcare costs shrink because interventions start earlier, and where underserved communities gain access to trustworthy technology. Most importantly, imagine patients preserving dignity, independence, and quality of life. That’s the vision driving this project.
Under the Hood
For those curious about the technology, the system uses EfficientNet to classify Alzheimer’s stages from MRI scans. It doesn’t just output a label—it provides stage predictions with high accuracy, Grad‑CAM overlays to highlight brain regions, SHAP visualisations for voxel‑level attribution, confidence scores and entropy plots to assess certainty, and confusion matrices to evaluate performance. Every output is designed to be interpretable, keeping clinicians in the loop rather than sidelining them.
Building the Pipeline
The journey of building this tool was as important as the outcome. We started with public Kaggle Alzheimer’s MRI datasets, carefully structured into labelled DataFrames. The model was EfficientNet, adapted for multi‑label classification. Training involved mixup augmentation, class weighting, early stopping, and adaptive learning rate scheduling. Evaluation included accuracy tracking, entropy‑based uncertainty, and batch‑level visualisations.
Interpretability was a priority from the start. SHAP, Grad‑CAM, confidence scores, and confusion matrix heat maps were integrated to ensure transparency. Every design choice was made with clinicians in mind.
Challenges Along the Way
Of course, it wasn’t easy. Class imbalance across Alzheimer’s stages required careful weighting and sampling. Integrating interpretability tools like SHAP and Grad‑CAM into EfficientNet was technically challenging. Deployment stability on Render with a FastAPI backend demanded resilience. And designing visual outputs that were both informative and intuitive for clinicians took iteration. Each challenge forced us to innovate, and each solution strengthened the system.
What We Achieved
The results speak for themselves. We achieved 97.67% test accuracy on multi‑stage classification. More importantly, we built a fully interpretable pipeline with SHAP, Grad‑CAM, and entropy visualisations. The framework is modular and reproducible, making it adaptable for other medical imaging tasks. And the dashboards we designed support real clinical decision‑making, bridging the gap between AI and practice.
Lessons Learned
Along the way, we learned that balancing performance with interpretability is key. Handling imbalanced data takes creativity. Transparency builds clinician trust. Integrating interpretability tools isn’t always straightforward—but it’s worth it. And above all, ethical deployment and reproducibility matter.
What’s Next
The journey doesn’t stop here. Next steps include expanding to include multimodal inputs such as PET scans, cognitive scores, and genetic markers. We plan to integrate with EMR systems for real‑time use, launch pilot studies in rural clinics, and publish validation results in peer‑reviewed journals. Beyond Alzheimer’s, the pipeline could extend to other neurodegenerative diseases such as Parkinson’s and stroke. And we aim to build an open‑source interpretability toolkit for medical AI transparency.
🔗 The code is open‑source. You can explore, fork, or adapt it:
- CNN Model: Alzheimers_Detection_ML
- API Service: Alzheimers_Detection
About the author
Reetam Biswas is passionate about blending technology with health and human care, based in Cary, North Carolina, USA, Fellow of the Soft Computing Research Society (SCRS), and Associate Member of the International Academy of Digital Arts and Sciences (IADAS).
