Computer Engineering undergrad at K.C. College of Engineering (Mumbai University, 2024-2028) with deep, hands-on experience across the ML stack. I don't just use models, I build them from scratch, study their architectures, and deploy them on real hardware.
My work spans transformer language models, state space models (Mamba), agentic AI systems, and RAG pipelines, with a focus on taking cutting-edge research and turning it into working systems.
Built a complete decoder-only transformer language model from first principles. Custom BPE tokenizer, multi-head self-attention, mixed-precision (FP16) training, cosine LR scheduling. Key Results:
|
In-depth study and reproduction of the Mamba architecture, a linear-time sequence model using selective state spaces as an efficient alternative to attention. Key Results:
|
Multi-agent AI productivity platform with RAG pipelines and vector databases for context-aware, personalized assistance. Key Results:
|
A persistent, full-stack multi-agent system acting as an always-on personal intelligence layer across calendars, tasks, health, finances, and goals. Key Results:
|
| Domain | Technologies |
|---|---|
| ML / Deep Learning | PyTorch · NumPy · Pandas · Scikit-learn · Hugging Face Transformers |
| LLM & Agents | LangChain · LlamaIndex · RAG Pipelines · Agentic AI · Tokenization & Embeddings |
| Research Areas | Transformer Architecture · State Space Models (Mamba) · Scaling Laws · MLOps |
| Languages | Python · Java · JavaScript/TypeScript · SQL · Bash |
| Infrastructure | Docker · REST APIs · Spring Boot · Git · Linux · Raspberry Pi |
- Founded Focuzdrvn platform:for Tech Entreprenurs
- Directed Singularity Hackathon (300+ participants, end-to-end delivery)
- E-Cell IIT Bombay Campus Ambassador, organized 5+ workshops for 200+ students
- Built a decoder-only transformer LLM entirely from scratch
- Researched Mamba SSM architecture and reproduced benchmark experiments
- Won Gold medal in Kaggle Tunix competition 2025


