Ajay Kumar

Data Scientist @ AllHeart Web
Technical Data Scientist specializing specializing in the development of Generative AI, Computer Vision, and high-performance Deep Learning systems. I bridge the gap between theoretical research and production-grade engineering by implementing SOTA architectures—including Transformers, YOLO, and U-Net—from first principles using PyTorch.
My professional expertise spans the end-to-end ML lifecycle: from pretraining custom Foundation Models (1.3B+ tokens) and fine-tuning via PEFT (LoRA), to architecting RAG-based domain intelligence and optimizing models for on-device inference (GGUF/Quantization). I have a proven track record of managing big data pipelines handling millions of records and deploying scalable, fault-tolerant AI services via FastAPI and MLOps frameworks. Core Focus: Transformer Architectures, Self-Supervised Learning (MAE/ViT), On-Device LLM Inference, and Scalable Backend Engineering.
About Me
Name: Ajay Kumar
Education: B.Tech. in Electronics and Communication
University: Himachal Pradesh Technical University
Current Address: Chandigarh, India
Email: kumarajaypaonta@gmail.com

Work Experience

Data Scientist — All Heart Web Apr 2025 – Present
  • Built and optimized a Whois Data chatbot for whoisdatacenter.com by integrating large-scale datasets and automating data pipelines using Python and Shell scripting.
  • Developed a FastAPI-based WHOIS data service enabling structured querying and efficient domain intelligence retrieval.
  • Worked extensively with SQL, MongoDB, and Elasticsearch for scalable data storage, querying, and analytics.
  • Designed and executed big data pipelines handling millions of records with high reliability.
  • Applied Selenium and web scraping techniques to extract structured data from diverse web sources.
  • Deployed and monitored production workflows on servers, ensuring performance, stability, and fault tolerance.
Data Analyst Intern — Indicraft Vintage Private Limited May 2024 – Sep 2024
  • Performed data cleaning, preprocessing, and exploratory data analysis to support business decision-making.
  • Analyzed large datasets to identify trends, patterns, and actionable insights.
  • Created dashboards and reports to communicate findings to stakeholders.
  • Collaborated with cross-functional teams to improve data-driven workflows.
Tools and Skills I use.
Python
Git
VS Code
Jupyter Notebook
Google Colab
TensorFlow
PyTorch
Machine Learning
Data Science
Deep Learning
NLP
Scikit-learn
Neural Networks
Streamlit
HTML5
CSS3
Hugging Face Spaces
MLOPS
dvc
mlflow
GEN AI
LLMs
RAG
Web Scraping
FAST API
GANs
Transformers
Vision Transformers

Projects

A curated set of projects exploring data science, machine learning, and AI research. From exploratory analysis and classical ML to deep learning and generative models, these projects highlight my journey of learning AI through theory, code, and experimentation.

On-Device LLM Systems: From Model Conversion to Offline Mobile Inference (Qwen2 + GGUF + Flutter).
Built an end-to-end edge AI pipeline to run Large Language Models fully offline on mobile devices, from model conversion to real-time inference in a Flutter app. Developed a custom pipeline to convert Qwen2 models (.safetensors) into GGUF format for compatibility with llama.cpp, with full control over tensor mapping and tokenizer integration. Models were then quantized to optimize performance for resource-constrained devices. Integrated llama.cpp into a Flutter-based chat application to enable real-time, on-device inference with no API calls, no internet dependency, and complete privacy.
Apk Link App Code Code LinkedIn Quantized model
OnDeviceAI Qwen2 0.5B llama.cpp gguf conversion GGUF Quantization (Q4) LLM Flutter C++ CPU-only inference Hugging Face Spaces
Akkadian to English Translation using ByT5 + LoRA (Kaggle competition)
Built a neural machine translation system for Akkadian (ancient, low-resource language) to English as part of the Deep Past Challenge (Kaggle). The project focuses on handling extreme data scarcity, noisy labels, and unknown scripts using modern parameter-efficient fine-tuning techniques.
Code Kaggle Dataset / Model Kaggle Notebook
Transformers PEFT (LoRA) PyTorch SacreBLEU google/byt5-small Kaggle competition Akkadian to English Hugging Face Spaces
Vision Transformer & MAE for Semantic Image Retrieval.
Built a semantic image retrieval system using a Vision Transformer encoder trained with Masked Autoencoder pretraining on the Food-101 dataset. The encoder generates high-dimensional image embeddings, which are used to retrieve visually similar food images through similarity search. Additionally, experimented with a Transformer-based caption generation pipeline using cross-attention, highlighting practical challenges in aligning visual features with natural language.
Live Link Code LinkedIn Thread
PyTorch Vision Transformer (ViT) Masked Autoencoder (MAE) FAISS (for similarity search) FastAPI Computer Vision Self-Supervised Learning Feature Extraction Hugging Face Spaces
हिंदी GPT Model from Scratch — From Pretraining on 1.35B tokens to SFT & Deployment using PyTorch.
Designed and trained a decoder-only GPT-style language model from scratch in PyTorch, implementing masked multi-head self-attention with Rotary Positional Embeddings (RoPE), feed-forward networks, residual connections, normalization, and a stacked decoder architecture from first principles. The model was pretrained on ~1.35B Hindi tokens using a custom SentencePiece BPE tokenizer (Hindi-only vocabulary), enabling experimentation with large-scale language modeling, convergence stability, and efficient training on limited hardware.
Live Link Pretraining Code SFT Code Tokenizer Code Model files LinkedIn Thread
Hindi GPT GPT Pretraining SFT Custome Hindi Tokenizer Generative AI Foundation Models Decoder Only Transformers NLP AttentionIsAllYouNeed PyTorch Gradio
Implementing the Transformer Encoder from scratch using PyTorch.
Implementing the Transformer Encoder from scratch using PyTorch — exactly as described in “Attention is All You Need” and also implements Multi-Head Attention, Positional Encoding, Feed-Forward networks, Residual+LayerNorm, and an encoder stack from first principles. The encoder was trained on KP20k (BIO labels) to validate the implementation and to experiment with attention behavior, head effects and extraction heuristics.
Live Link Code LinkedIn
Generative AI Foundation Models Encoder Transformers NLP AttentionIsAllYouNeed PyTorch
YOLOv1-Research Paper-Implementation-using-PyTorch-from-scratch.
This project is a from-scratch implementation of the YOLOv1 (You Only Look Once) object detection paper using PyTorch. I implemented the entire pipeline — architecture, loss function, dataset parsing, and model training — to deeply understand how YOLO works at its core.
Live Link Code LinkedIn
YOLOv1 Foundation Models ComputerVision ObjectDetection AIResearch PyTorch
Implementation of GAN Research Papers: From Concept to Face Generation.
Implemented and trained a DCGAN model to generate realistic human faces using the CelebA dataset. Built an interactive web interface with Flask, HTML, CSS, and JavaScript that allows users to generate faces, morph between two faces with a slider, and create batches of 100 synthetic faces. Deployed the project on Hugging Face Spaces, showcasing end-to-end GAN implementation from research paper to real-time user interaction.
Live Link Code LinkedIn
GAN Foundation Models DCGAN ComputerVision FaceGeneration ImageGeneration PyTorch
AI-Powered Stock Market Insights: Building a Chrome Extension with Selenium, FastAPI, and LangChain.
I built a Chrome extension powered by Selenium, FastAPI, and LangChain that works as an AI-driven stock market chatbot assistant. With a simple natural language query, it can fetch real-time stock performance, show detailed stats for any stock, provide sector-wise insights, pull the latest market news, and display interactive charts — all directly inside the browser.
Medium Blog Code LinkedIn
ChromeExtension AI Agent StockMarket_Bot LangChain Selenium Scraping FastAPI MarketInsights Automation
Implementation of U-Net Research Paper using PyTorch for Car Image Segmentation and Mask Generation.
I have implemented the original U-Net research paper using PyTorch and then trained this model on a car dataset. created the frontend to use this model and generate the mask of the car. For the frontend, I used Flask, HTML, and CSS. In the app, I can upload an image, and by clicking a button, it generates the mask and overlay of the original image.
Live Link Code Original Research Paper
UNet ImageSegmentation MaskGeneration Flask
Show more projects
An AI-powered medical assistant leveraging WHO data via RAG.
An AI-powered medical assistant leveraging Retrieval-Augmented Generation (RAG) on WHO data to deliver accurate medical responses. Built with a modular Flask (frontend) and FastAPI (backend) architecture, integrated with Pinecone for semantic search, MongoDB for patient data, and containerized with Docker for seamless deployment on Hugging Face Spaces.
Live Link Code LinkedIn
RAG MedicalAI FastAPI Flask Docker Pinecone MongoDB LLM HuggingFaceSpaces
Dog Cat Classifier using pretrained model Resnet50
This project is deep learning-based image classification model that distinguishes between dogs and cats using the ResNet50 architecture. The model leverages transfer learning from a pretrained ResNet50 network (trained on ImageNet) to achieve high accuracy (95.36%) with minimal training time.
Live Link Code
resnet50 transferlearning Fine Tuning dogvscatclassifier Streamlit ImageNet Data HuggingFaceSpaces
RainFall Prediction using ML and Experiment tracking with MLflow
This project focuses on predicting rainfall using machine learning techniques while implementing experiment tracking with MLflow and DagsHub. The goal is to build an accurate model that forecasts rainfall based on meteorological features.
Live Link Code
RainFallprediction mlflow MachineLearning Experiment Tracking Streamlit HuggingFaceSpaces
Email Spam Classifier
This project focuses on developing a robust Email/SMS Spam Classifier that leverages machine learning techniques to accurately distinguish between spam (unwanted) and ham (legitimate) messages.
Live Link Code Medium Blog
SpamDetection NLP MachineLearning DataScience Streamlit HuggingFaceSpaces
Cx Churn Prediction With PyTorch
This project predicts customer churn in credit card services using an Artificial Neural Network (ANN) built with PyTorch. The model was trained on the Credit Card Customer Churn Prediction dataset from Kaggle, achieving an accuracy of 86%.
Live Link Code
CustomerChurn ArtificialNeuralNetwork PyTorch DeepLearning Streamlit HuggingFaceSpaces
Book Recommender System
In this project, I developed a Book Recommender System using a combination of popularity-based and collaborative filtering techniques. The system helps users discover books based on their reading history and interactions with books, and it also highlights the top-rated books. The model is deployed using a web application built with Streamlit and hosted on Hugging Face.
Live Link Code Medium Blog
RecommenderSystem CollaborativeFiltering MachineLearning BookRecommendations Streamlit HuggingFaceSpaces
Exploratory Data Analysis of IPL Dataset
I’ve developed a dynamic web application using Streamlit that offers insightful analysis of IPL data. This user-friendly app allows enthusiasts to explore the vast IPL dataset interactively.
Live Link Code Medium Blog
IPLEDA DataAnalysis Cricket Streamlit HuggingFaceSpaces
Olympics Unveiled: An In-Depth Exploratory Data Analysis
I’ve developed a Streamlit web application to visualize and analyze Olympic data. This app leverages interactive charts and plots to explore various metrics, such as athlete performance and medal distribution, providing insights into trends and patterns across different Olympic events. It enhances data understanding through user-friendly, real-time data exploration tools.
Live Link Code Medium Blog
olympicsdataanalysis DataAnalysis EDA Olympics Streamlit HuggingFaceSpaces

Blogs

I write about Machine Learning, Deep Learning, NLP, and practical AI concepts with real-world intuition.

Behind the Code: Building a Stock Market Chatbot with FastAPI, Selenium and LangChain
Introduction, The stock market moves fast, and so does the information around it. Traders often need to check multiple things at once — best and worst performers of the day, single stock details, the latest news, sector trends, or even a stock chart for a specific timeframe. Doing this manually can feel repetitive and time-consuming. That’s where my project comes in.
Linear Regression 101: Everything You Need to Know for Data Science Interviews
Linear regression is one of the most fundamental algorithms in machine earning, widely used for predictive modeling. In this guide, we’ll break down the key concepts, interview questions, and advanced techniques to help you master linear regression from the ground up.
Fighting Spam with Machine Learning: Building an Effective Email and SMS Classifier
Spam emails and SMS messages have become a significant concern in today’s digital world, often cluttering inboxes with unwanted content or, worse, phishing attacks. These messages are an everyday nuisance, clogging our inboxes with unsolicited or malicious content. Machine learning provides a powerful solution to automatically classify and filter these messages, improving the user experience.
From Data to Recommendations: Creating a Machine Learning Book Recommender System
I implemented a popularity-based recommendation system for the Top 50 Books section. This system ranks books based on overall popularity, considering factors like highest ratings, most reviews, and frequent user interactions. It effectively showcases widely-loved and highly-rated books, making it ideal for users seeking to discover popular titles that have already gained a strong reputation.

“AI is going to bring a new renaissance for humanity, a new form of enlightenment, if you want, because AI is going to amplify everybody's intelligence.”

— Yann LeCun