🎯
Focusing
Pinned Loading
-
mit-han-lab/llm-awq
mit-han-lab/llm-awq Public[MLSys 2024 Best Paper Award] AWQ: Activation-aware Weight Quantization for LLM Compression and Acceleration
-
mit-han-lab/smoothquant
mit-han-lab/smoothquant Public[ICML 2023] SmoothQuant: Accurate and Efficient Post-Training Quantization for Large Language Models
-
mit-han-lab/tinyengine
mit-han-lab/tinyengine Public[NeurIPS 2020] MCUNet: Tiny Deep Learning on IoT Devices; [NeurIPS 2021] MCUNetV2: Memory-Efficient Patch-based Inference for Tiny Deep Learning; [NeurIPS 2022] MCUNetV3: On-Device Training Under 2…
-
-
mit-han-lab/gan-compression
mit-han-lab/gan-compression Public[CVPR 2020] GAN Compression: Efficient Architectures for Interactive Conditional GANs
-
mit-han-lab/temporal-shift-module
mit-han-lab/temporal-shift-module Public[ICCV 2019] TSM: Temporal Shift Module for Efficient Video Understanding
Something went wrong, please refresh the page to try again.
If the problem persists, check the GitHub status page or contact support.
If the problem persists, check the GitHub status page or contact support.