Pinned Loading
-
CASE-Lab-UMD/LLM-Drop
CASE-Lab-UMD/LLM-Drop PublicThe official implementation of the paper "What Matters in Transformers? Not All Attention is Needed".
-
CASE-Lab-UMD/Unified-MoE-Compression
CASE-Lab-UMD/Unified-MoE-Compression PublicThe official implementation of the paper "Demystifying the Compression of Mixture-of-Experts Through a Unified Framework".
-
CASE-Lab-UMD/Router-Tuning-Mixture-of-Depths
CASE-Lab-UMD/Router-Tuning-Mixture-of-Depths PublicThe open-source Mixture of Depths code and the official implementation of the paper "Router-Tuning: A Simple and Effective Approach for Enabling Dynamic Depth in Transformers."
-
SparseAdapter
SparseAdapter PublicSource code of EMNLP 2022 Findings paper "SparseAdapter: An Easy Approach for Improving the Parameter-Efficiency of Adapters"
Python 18
Something went wrong, please refresh the page to try again.
If the problem persists, check the GitHub status page or contact support.
If the problem persists, check the GitHub status page or contact support.