I'm Sourab Mangrulkar; an Applied Scientist and Machine Learning Engineer from India ๐ฎ๐ณ.
- ๐ญ Iโm currently working as an Applied Scientist at Amazon.
- ๐ฑ Exploring Natural Language Processing, Computer Vision and Distributed Training at Scale. Always up for meaningful collaboration.
- ๐ Pronouns: He/His/Him.
- โก Painting ๐จ, sketching โ๏ธ and poetry ๐ are my favourite hobbies. Recently, I've started reading up on stocks and economic markets.
- ๐ซ How to reach me: ย
- Arxiv 2023 BigCode Project: SantaCoder: don't reach for the stars!
- SIGIR 2022 (eCommerce Workshop): HISS: A Novel Hybrid Inference Architecture in Embedding Based Product Sourcing using Knowledge Distillation
- KDD 2022 (ADS Track): BE3R: BERT-based early-exit using expert routing
- WWW 2022 (Industry Track): Multilingual Semantic Sourcing using Product Images for Cross-lingual Alignment
- SIGDial 2018: A Context-aware Convolutional Natural Language Generation model for Dialogue Systems
- March 2024: You can now train a 70b language model at home - In collaboration with Answer.AI
- February 2024: ๐ค PEFT welcomes new merging methods
- January 2024: Finetune LLMs on your own consumer hardware using tools from PyTorch and Hugging Face ecosystem
- December 2023: Mixture of Experts Explained
- October 2023: Personal Copilot: Train Your Own Coding Assistant
- October 2023: Falcon 180B Finetuning using ๐ค PEFT and DeepSpeed
- September 2023: Fine-tuning Llama 2 70B using PyTorch FSDP
- August 2023: Large Scale Training of Hugging Face Transformers on TPUs With PyTorch/XLA FSDP
- June 2023: The Falcon has landed in the Hugging Face ecosystem
- May 2023: Making LLMs even more accessible with bitsandbytes, 4-bit quantization and QLoRA
- February 2023: ๐ค PEFT: Parameter-Efficient Fine-Tuning of Billion-Scale Models on Low-Resource Hardware
- June 2022: Accelerate Large Model Training using DeepSpeed
- May 2022: Accelerate Large Model Training using PyTorch Fully Sharded Data Parallel
- December 2023: Hands on session for the ACM Winter School focused on Generative AI @ ACM Winter School India Chapter 2023
- December 2023: Generative AI for All. ๐ค PEFT: Finetuning made simple, efficient and extendable @ NeurIPS Conference 2023 - Large Language Model Efficiency Challenge: 1 LLM + 1GPU + 1Day
- October 2023: Training a LLaMA in your Backyard: Fine-tuning Very Large Models on Consumer Hardware @ PyTorch Conference 2023
- August 2023: Unleashing LLMs: Training, Finetuning, and Evaluating @ DataHack Summit 2023 (Analytics Vidhya)
- August 2023: Parameter-Efficient Fine-Tuning: Doing more with less @ DataHack Summit 2023 (Analytics Vidhya)