|
Mukul Ranjan
I am a Master's student in Machine Learning at MBZUAI, where I work with Prof. Zhiqiang Shen. My research focuses on efficient machine learning systems, particularly developing alternate architectures and hardware-aware algorithms through sparsity, quantization, and hardware-software co-design. Most recently I also worked with Prof. Deming Chen at UIUC from Jul. 2025 to Oct. 2025. I also work on creating evaluation methodologies and benchmarks for these systems.
Previously, I was a Data Scientist at Meesho, where I built personalized ranking systems, and an AI Research Scientist at Qure.ai, developing automated stroke severity assessment systems deployed in hospitals worldwide.
I hold a B.Tech. in Electronics and Communication Engineering from IIT Guwahati.
Email /
CV /
Scholar /
Github /
LinkedIn /
Medium /
Blogs
|
|
| Oct 2025 |
Elastic-Cache for Diffusion LLMs released |
| May 2025 |
SpookyBench paper on temporal reasoning in VLMs released |
| Feb 2025 |
KITAB-Bench accepted at ACL 2025! |
| Oct 2024 |
Won 1st place at GITEX DGE Elite Hackathon for Cybersecurity |
| Aug 2024 |
Started MS in Machine Learning at MBZUAI |
Publications
Representative papers are highlighted. * indicates equal contribution (random order).
|
|
Attention Is All You Need for KV Cache in Diffusion LLMs
Mukul Ranjan*, Quan Nguyen-Tri*, and Zhiqiang Shen
Under Review, 2025
arXiv
/
code
/
project page
Elastic-Cache is a training-free framework that accelerates Diffusion LLMs up to 45.1x with higher accuracy by adaptively refreshing the KV cache.
|
|
Time Blindness: Why Video-Language Models Can't See What Humans Can?
Mukul Ranjan*, Ujjwal Upadhyay*, Zhiqiang Shen, and Mohamed Elhoseiny
Under Review, 2025
arXiv
/
code
/
project page
SpookyBench reveals that patterns in temporal noise that humans recognize with 98% accuracy, state-of-the-art VLMs fails completely achieving 0%.
|
|
Mobile-MMLU: A Mobile Intelligence Language Understanding Benchmark
Mukul Ranjan*, Sondos Mahmoud Bsharat*, Aidar Myrzakhan*, et al.
Under Review, 2025
arXiv
/
dataset
/
project page
Mobile-MMLU is a benchmark with with 16,000+ questions across 80 mobile-related fields to evaluate LLM performance under real-world constraints.
|
|
Beyond Size: How Gradients Shape Pruning Decisions in Large Language Models
Rocktim Jyoti Das*, Mukul Ranjan*, Mingjie Sun*, Liqun Ma, and Zhiqiang Shen
Under Review
arXiv (coming soon)
/
code
GBLM-Pruner is a gradient-based pruning method that is extremeley faster than weight-update methods like SparseGPT.
|
|
One-for-All: Generalized LoRA for Parameter-Efficient Fine-tuning
Arnav Chavan*, Mukul Ranjan*, Zhuang Liu, Deepak Gupta, Eric Xing, and Zhiqiang Shen
Pending Submission for IEEE TPAMI, 2025
arXiv (coming soon)
/
code
GLoRA is a unified PEFT framework achieving state-of-the-art accuracy with zero inference overhead through structural re-parameterization.
|
|
KITAB-Bench: A Comprehensive Multi-Domain Arabic OCR Benchmark
Mukul Ranjan*, Ahmed Heakl*, Muhammad Abdullah Sohail*, et al.
ACL 2025 (Findings), 2025
arXiv
/
dataset
KITAB-Bench has 8,809 samples across 9 domains. It reveals that vision-language models outperform traditional OCR by 60%.
|
|
Deep-ASPECTS: A Segmentation-Assisted Model for Stroke Severity Measurement
Ujjwal Upadhyay, Mukul Ranjan, et al.
ECCV, 2022
arXiv
Deep-ASPECTS is an automated ASPECT scoring system achieving radiologist-level performance, now deployed in hospitals worldwide.
|
|
An Evaluation of Google Translate for Sanskrit to English Translation
Akshat Shukla, Chaarvi Bansal, Sushrut Badhe, Mukul Ranjan, and Rohitash Chandra
Natural Language Processing Journal, 2023
paper
|
|
Artificial Intelligence for Topic Modelling in Hindu Philosophy
Mukul Ranjan* and Rohitash Chandra*
PLOS ONE, 2022
paper
|
|
Last updated: January 2026
© 2026 Mukul Ranjan. Design inspired by Jon Barron.
|
|