From Sketch to Scale.
Early-career full stack AI developer passionate about Backend engineering and Infra for Next-Gen Agentic LLM Systems.
EVERYTHING ABOUT
VAIBHAV
Hi, I'm Vaibhav Tanwar — an Applied AI Engineer who builds end-to-end ML systems that ship to production, from model training to containerized deployment at scale.
My core stack spans PyTorch, FastAPI, LangChain, Docker, and PostgreSQL. I've built computer vision pipelines with YOLO and CLIP, multi-agent LLM systems, and MLOps platforms handling millions of records — all with rigorous testing and CI/CD.
Whether it's a zero-to-one AI product or scaling existing infrastructure, I focus on reliability, low-latency inference, and clean architecture that teams can build on.
MY TECH TOOLBOX
AI & Machine Learning
PyTorch, Transformers, Scikit-Learn, YOLO, CLIP, OpenCV, SpaCy
Training, fine-tuning, and deploying CV & NLP models at scale.
LLM Ops & Agents
LangChain, LangGraph, DSPy, Gemini-ADK, Mem0, Vector DBs
Building multi-agent systems, RAG pipelines, and LLM toolchains.
Backend & APIs
FastAPI, Node.js, NestJS, Kafka, RabbitMQ
High-throughput APIs and async task queues.
Databases
PostgreSQL, MongoDB, Redis, MySQL, pgvector
Relational, document, and vector stores at scale.
Infrastructure
Docker, AWS, GitHub Actions, CI/CD, Prometheus
Containerized deployments with full observability.
Previous Endeavors

Infosys Centre for Artificial Intelligence
ML + Backend Engineer
Developed advanced wildlife monitoring capabilities by fine-tuning YOLO and custom Transformer based architectures along with (CUDA, TensorRT) inference optimization. Built a robust backend infrastructure (FastAPI, PostgreSQL, Docker) featuring optimized queries and an end-to-end MLOps pipeline for continual learning from camera trap data. Mitigated annotation bottlenecks using Active Learning algorithms and ensured system health via custom API monitoring tools.

Scale AI
LLM Post Training Contributor
Enhanced zero shot inference capability of LLMs through supervised fine-tuning (SFT) and Reinforcement Learning from Human Feedback (RLHF). Curated and refined domain-specific datasets for complex reasoning tasks. Optimized reward models using human preference data to better align outputs with user expectations (truthfulness, harmlessness, instruction-following). Collaborated with ML engineers on refining annotation guidelines and feedback mechanisms.

Networked Systems and Security Research Lab
Undergraduate Researcher
Improved latency and throughput for live media transfer from semi-autonomous vehicles to edge servers under Dr. Arani Bhattacharya, utilizing state-of-the-art multipath QUIC protocols. Critically analyzed and tested Alibaba's XQUIC and Tencent's TQUIC frameworks to identify solutions for latency bottlenecks. Reported detailed findings on performance and build library inconsistencies.

MIDAS Research Group
Working on improving Foundational models for Self Supervised Speech Representation Learning like HuBERT and MS-HuBERT.
Projects
Let's Build
Something Amazing
Have a project, idea, or collaboration in mind? I'd love to hear from you. Let's create something impactful together.