My background, education, professional journey, and technical expertise
I'm a passionate AI/ML Engineer and Researcher recently graduated from Northeastern University with a Master's in Information Systems. My journey in technology began with a deep fascination for artificial intelligence and its potential to solve complex real-world problems.
With a strong foundation in computer science and extensive experience in machine learning, I specialize in developing innovative AI solutions, optimizing ML pipelines, and building scalable data systems. My work spans across computer vision, natural language processing, and distributed computing, with a particular focus on large language models and diffusion models.
I believe in the power of open-source collaboration and actively contribute to the AI/ML community through research, technical writing, and knowledge sharing. I recently contributed to the LLVM project by refactoring the `ilogbf128` math function to a header-only implementation in libc ( LLVM libc open-source contribution ). When I'm not coding or researching, you'll find me exploring the latest AI trends, mentoring fellow students, or contributing to open-source projects.
Boston, MA
Master of Science in Information Systems
2023 - 2025
Specializing in AI/ML, Data Engineering, and Distributed Systems. Focus on cutting-edge technologies including large language models, computer vision, and scalable data processing platforms.
Master's Thesis:
ModelOpt: Research Framework for Zero-Shot Computer Vision Model Optimization With Tree Search and Federated Knowledge Sharing, Northeastern University, 2025.
Chennai, India
Bachelor of Technology in Computer Science
2019 - 2023
Comprehensive foundation in computer science fundamentals, algorithms, data structures, and software engineering principles. Graduated with distinction and active involvement in technical projects.
Developed and optimized high-performance GPU kernels for inference/training workloads, demonstrating deep knowledge of memory hierarchy and compute/memory-bound optimization strategies.
Worked with graph compilers (CUDA, HIP) to optimize deep learning frameworks for various hardware architectures including AMD GPUs, ensuring streams integration.
Identified 70% redundant model calls through profiling; architected Redis caching system reducing monthly GPU costs by $12K and inference latency by 42%.