AI & ML interests
Efficient and adaptive foundation models across language and multimodal intelligence.
Recent Activity
Papers
Demystifying When Pruning Works via Representation Hierarchies
Making Large Language Models Efficient Dense Retrievers
LLM-Drop
š¤ LLM-Drop hosts research artifacts for efficient foundation models, with a focus on large language models and unified multimodal models.
Our work studies how modern foundation models can be made more efficient while preserving their core capabilities. This page collects model weights, code links, project pages, and related resources from our research projects.
š Projects
š§© LLM-Drop
Uncovering the Redundancy in Transformers via a Unified Study of Layer Dropping
TMLR 2026
- š Paper: https://openreview.net/forum?id=1I7PCbOPfe
- š» Code: https://github.com/CASE-Lab-UMD/LLM-Drop
- š¤ Models: https://huggingface.co/collections/LLM-Drop/llm-drop
š Pruning on Representations
Demystifying When Pruning Works via Representation Hierarchies
- š Project Page: https://case-lab-umd.github.io/Pruning-on-Representations/
- š Paper: https://arxiv.org/abs/2603.24652
- š» Code: https://github.com/CASE-Lab-UMD/Pruning-on-Representations
š Sparse Unified Models
Understanding and Harnessing Sparsity in Unified Multimodal Models
- š Project Page: https://shwai-he.github.io/SparseUnifiedModel/
- š Paper: https://huggingface.co/papers/2512.02351
- š» Code: https://github.com/Shwai-He/SparseUnifiedModel
- š¤ Models:
š¬ Contact
For questions or collaborations, please contact:
- Shwai He: shwaihe@umd.edu
- Guoheng Sun: ghsun@umd.edu