Project Descriptions

Short and readable write-ups.

Fine-Tuning CNN Architectures Using Transfer Learning

Tech: PyTorch, Transfer Learning, Optuna, GPU · Dataset: Fashion MNIST · Outcome: 91% → 95.5% Accuracy

Context
Baseline CNN plateaued at 91% accuracy; required stronger generalization on limited labeled data.

Approach

Architecture

VGG16 Transfer Learning Architecture Diagram

Result
Improved accuracy to 95.5% (+4.5%), reduced overfitting via weight decay tuning, and achieved faster convergence vs training from scratch for a domain-specific, smaller dataset.

Repository

Sentiment Analysis for Healthcare Domain with NLP & MLOps

Tech: Python, Scikit-learn, MLflow, SQL · Dataset: Govt. of India survey data (~15,000 anonymized patient text entries) + Kaggle sentiment dataset · Outcome: 80% Accuracy · 0.84 ROC-AUC · 12% F1 disparity reduction

Context
Built an NLP classification system to analyze rural patient text and assist urban therapists with sentiment signals and recurring thought categorization across 7 predefined mental health concern classes.

Approach

Architecture

Architecture Diagram

Results
Logistic Regression improved accuracy from 72% (Naive Bayes baseline) to 94.65%, increased ROC-AUC to 0.84, and reduced cross-demographic F1 disparity by 12% via class rebalancing and threshold calibration. Improved recall for high-risk categories by 9%.

Repository

Agentic RAG with Hallucination Filtering & Self-Reflection

Tech: LangGraph, LangChain, FAISS, OpenAI APIs, FastAPI, Docker, AWS · Dataset: 5,000+ indexed documents + live web search · Outcome: +15% GPT-Judge accuracy · 20% redundancy reduction

Context
Designed an agentic Retrieval-Augmented Generation (RAG) system to reduce hallucinations and improve factual grounding in multi-domain question answering.

Approach

Architecture

Architecture Diagram

Metrics
Improved GPT-Judge factual accuracy by 15% vs baseline RAG, reduced retrieval redundancy by 20%, and decreased hallucination rate by 18% in simulated evaluation.

Deployment & MLOps
Containerized with Docker, deployed on AWS EC2/SageMaker, CI/CD via GitHub Actions, and exposed scalable FastAPI endpoints with Pydantic validation.

Key Design Decisions
Adopted agentic routing for modularity, integrated automated hallucination grading for reliability, and prioritized retrieval diversity (MMR + multi-query) to improve grounding.

Repository

Fine-Tuning LLM with Quantization & LoRA

Tech: Hugging Face Transformers, PEFT, Datasets, PyTorch · Base Model: Gemma-7B · Dataset: 945K+ customer support tweet pairs (input → response) · Outcome: +14% relevance · 60% lower memory · -35% training cost (est.)

Context
Fine-tuned Gemma-7B for customer support response generation using parameter-efficient LoRA adapters, optimized for lower-memory training and cost-effective inference.

Pipeline

Results
On a 300-sample blind test set, improved response relevance by 14% vs base Gemma-7B, increased win-rate to 62% in pairwise preference eval, and reduced hallucinated/irrelevant replies by 11%. 8-bit + LoRA reduced peak GPU memory by ~60%, enabling deployment on lower-cost infrastructure.

Key Design Decisions
Chose LoRA for fast iteration and small artifact size, 8-bit quantization for hardware efficiency, and targeted attention modules (q/v) for best quality–cost tradeoff.

Repository

End-to-End Sales BI Pipeline & Dashboard

Tech: MySQL, Power Query, Airflow, Power BI · Dataset: ~150K OLTP sales transactions · Outcome: 20% reduction in ad-hoc analysis · Real-time KPI visibility

Context
Built a scalable BI pipeline to transform raw transactional sales data into structured analytics-ready datasets, enabling stakeholders to monitor revenue, customer performance, and product trends in real time.

Architecture & Pipeline

Analytics & Visualization

Dashboard Preview

Architecture Diagram Architecture Diagram

Results
Reduced manual reporting and ad-hoc analysis by ~20%, improved decision latency, and provided actionable insights into regional revenue trends and product performance.

Repository

End-to-End Regression ML System with AWS CI/CD

Tech: Regression, Scikit-learn, Pandas, Flask, Docker, GitHub Actions, AWS (ECR, EC2) · Dataset: Student performance tabular dataset (~1,000 rows, mixed numeric + categorical) · Outcome: MAE 3.8 · RMSE 5.1 · R² 0.89 (test)

Context
Built a production-style regression system to predict student exam scores from demographic and academic signals, packaged as a web app and deployed on AWS with CI/CD.

Pipeline & Modeling

High Level Workflow

Architecture Diagram

Results
Achieved MAE 3.8 and RMSE 5.1 on held-out test data (R² 0.89). Regularization reduced train–test gap and improved stability vs an unregularized baseline (MAE 5.0, RMSE 6.6).

Deployment & MLOps
Containerized Flask inference service, pushed images to AWS ECR, deployed on EC2, and automated build/test/deploy with GitHub Actions. Designed the system so model + preprocessor artifacts can be swapped without rewriting the app.

Key Design Decisions
Used L1/L2 regularization for generalization under correlated features, enforced consistent preprocessing via saved pipelines, and chose Docker + ECR for repeatable deployments and environment parity.

Repository