import streamlit as st st.markdown( """ """, unsafe_allow_html=True, ) st.html( """
Creator of AiGymBuddy.in | Machine Learning | Deep Learning | Flutter | Math | MLops | TensorFlow | FastAPI
(He/Him)
Passionate about Machine Learning, Computer Science and Mathematics. Has a strong grip over Machine & Deep Learning Fundamentals, Computer Networks, OS and DSA. Solved 100+ problems on leetcode. Have a keen interest in learning Mathematics, Deep Learning and Statistics. I like to understand things in a deeper way.
Private Client | Sydney, Australia
October - November 2024
Current WebAPP which you are using
Deployed
Initially, I aimed to build a recommendation system from scratch using TensorFlow Recommenders (TFRS) on the MovieLens 1M dataset. This involved creating user and movie embeddings with a candidate generation and ranking model.
However, this TFRS approach proved too resource-intensive and time-consuming for effective training and testing. Crucially, the initial results weren't satisfactory for deployment.
Therefore, for the deployed web application, I switched to pre-trained models (BGE embeddings and re-ranking). This offered:
While the TFRS code is also included in the web app, the pre-trained model approach was chosen for its superior results and efficiency in a deployment setting. A future improvement could be fine-tuning the pre-trained models for even better performance.
Technologies used: TensorFlow Recommenders, Scann, Vector DB, Distributed GPU Training, Langchain, Streamlit, BAAI BGE Models
Deployed
Objective: Predicted employee attrition with 85% AUC to improve employee retention and business performance.
Model Development: Hyperparameter optimized Multi-Layer Perceptron, XGBoost, Logistic Regression with Inference as well as Training Pipeline
Backend: Developed a FastAPI backend for real-time predictions, using Pydantic for schema validation for incoming and outgoing requests
Deployment: Containerized with Docker, deployed on AWS EC2, managed via AWS ECR.
CI/CD: Set up an automated CI/CD pipeline using GitHub Actions for seamless updates.
Web Application: Built a user-friendly interface using Flutter Web for real-time interaction.
Security: Handled HTTPS requests using Caddy as a reverse proxy server
Technologies used: TensorFlow, AWS, Docker, FastAPI, CI/CD Pipeline, Multi-Layer Perceptron, Neural Network, XGBoost, Logistic Regression, Hyperparameter Tuned Models, GitHub Actions, Pydantic, Flutter Web, Reverse-Proxy-Server: Caddy
Deployed
Personalized AI-Driven Workouts with Smart Equipment Detection and Progress Tracking
Features: Al Instrument Detection (Camera or Gallery), Exercises based on Available Equipments, Time, Preffered Muscle Groups & Custom requests, Dynamic Video Tutorial Finder for each exercise, Super personalized Al generated routine, Workout History Tracker, Easy SignUp/Login with Google Oauth
Technologies used: Dart, flutter, firebase, gemini 1.5 flash, riverpod, langchain, fastapi, google oauth
Licenses: This code of this app/website is written from scratch and I hold all the rights over distribution
Deployed
Multi-Modal Cancer Detection: Developed a novel multi-output deep learning model for breast cancer detection, predicting cancer presence, invasiveness, and difficult-negative case status. The model incorporates both mammogram images and tabular clinical data, leveraging a non-sequential architecture to process distinct data modalities.
Fine-Tuned Image Feature Extraction: Utilized a pre-trained EfficientNetV2B3 model for image feature extraction, fine-tuning layers from block 6 onwards to enhance its applicability to the specific task, thus improving the quality of learned representations and potentially making the model more robust and accurate.
Distributed Training: Accelerated model training through distributed training using TensorFlow's MirroredStrategy on 2xT4 GPUs for 9 hours on Kaggle, demonstrating proficiency in optimizing model training with limited computational resources.
Technologies used: TensorFlow, Transfer Learning, EfficientNetV2, Fused MB-CNN
Deployed
Problem Statement: E-commerce and healthcare industries struggle to efficiently extract product details (weight, volume, dimensions) from images at scale.
Action: Developed a large-scale image-to-text inference pipeline using Qwen2 VL: 2B, incorporating image preprocessing, Regex, and parallel processing. Processed 84,000 of 131,000 test images.
Result: Successfully extracted product values from a significant portion of the dataset. Our team of four ranked 172nd out of ~75,000 in the Amazon ML Challenge with Fl-Score=0.47, demonstrating the solution's potential for automated product information extraction.
Technologies used: Qwen2 VL, Python, Regex, Parallel Processing
Technologies used: Langchain, VertexAI Embedding, StreamLit, PostGresVector
Ongoing
Carried out sessions to teach juniors the fundamentals of Git and GitHub, covering version control, collaboration, and best practices.
Acharya Technical Club - Steigen
Acharya Institute of Technology, Bangalore
2021-2025
CGPA-8.12
Kalyani Public School, Barasat, Kolkata
2021
77% (Auto Pass Covid Batch)
Sacred Heart Day High School, Kolkata
2019
90%