Skip to content

Latest commit

Β 

History

History

Applications

Folders and files

NameName
Last commit message
Last commit date

parent directory

..
Β 
Β 

Distillation Applications


Large Language Models

  • TinyBERT: Distilling BERT for Natural Language Understanding, arXiv 2019, πŸ”— :octocat:
  • Distilbert, a distilled version of bert: smaller, faster, cheaper and lighter, arXiv 2019, πŸ”—
  • Patient Knowledge Distillation for BERT Model Compression, arXiv 2019, πŸ”— :octocat:
  • Well-read students learn better: On the importance of pre-training compact models, arXiv 2019, πŸ”— :octocat:
  • Xtremedistil: Multi-stage distillation for massive multilingual models, arXiv 2020, πŸ”— :octocat:
  • MobileBERT: a Compact Task-Agnostic BERT for Resource-Limited Devices, ACL 2020, πŸ”— :octocat:
  • MINILM: deep self-attention distillation for task-agnostic compression of pre-trained transformers, NeurIPS 2020, πŸ”—
  • xtremedistiltransformers: Task transfer for task-agnostic distillation, arXiv 2021, πŸ”— :octocat:
  • Explanations from Large Language Models Make Small Reasoners Better, arXiv 2022, πŸ”—
  • BERT Learns to Teach: Knowledge Distillation with Meta Learning, ACL 2022, πŸ”— :octocat:
  • Distilling Step-by-Step! Outperforming Larger Language Models with Less Training Data and Smaller Model Sizes, arXiv 2023, πŸ”— :octocat:
  • Large language models are reasoning teachers, ACL 2023, πŸ”— :octocat:
  • Less is More: Task-aware Layer-wise Distillation for Language Model Compression, ICML 2023, πŸ”— :octocat:
  • MCC-KD: Multi-CoT Consistent Knowledge Distillation, EMNLP 2023, πŸ”— :octocat:
  • Teaching Small Language Models to Reason, ACL 2023, πŸ”—
  • Symbolic Chain-of-Thought Distillation: Small Models Can Also β€œThink” Step-by-Step, ACL 2023, πŸ”—
  • Gkd: Generalized knowledge distillation for auto-regressive sequence models, arXiv 2023, πŸ”—
  • SCOTT: Self-consistent chain-of-thought distillation, NeurIPS 2023, πŸ”— :octocat:
  • Dialogue Chain-of-Thought Distillation for Commonsense-aware Conversational Agents, EMNLP 2023, πŸ”— :octocat:
  • MiniLLM: Knowledge Distillation of Large Language Models, ICLR 2024, πŸ”— :octocat:
  • On-Policy Distillation of Language Models: Learning from Self-Generated Mistakes, ICLR 2024, πŸ”—
  • PaD: Program-aided Distillation Specializes Large Models in Reasoning, ACL 2024, πŸ”—
  • Turning dust into gold: Distilling complex reasoning capabilities from llms by leveraging negative data, AAAI 2024, πŸ”— :octocat:
  • PC-LoRA: Low-Rank Adaptation for Progressive Model Compression with Knowledge Distillation, arXiv 2024, πŸ”—
  • Reverse Thinking Makes LLMs Stronger Reasoners, arXiv 2024, πŸ”—
  • UniCoTT: A Unified Framework for Structural Chain-of-Thought Distillation, ICLR 2025, πŸ”—
  • MiniPLM: Knowledge Distillation for Pre-Training Language Models, ICLR 2025, πŸ”— :octocat:

Foundation Models

Prominent Foundation Models

  • CLIP : Learning Transferable Visual Models From Natural Language Supervision, PMLR 2021, πŸ”— :octocat:
  • GPT
    • GPT-1: Improving Language Understanding by Generative Pre-Training, OpenAI 2018, πŸ”—
    • Gpt-3: Language Models are Few-Shot Learners, NeurIPS 2020, πŸ”—
    • Gpt-4: A Review on Advancements and Opportunities in Natural Language Processing, arXiv 2023, πŸ”—
  • DALL-E: Zero-Shot Text-to-Image Generation, ICML 2021, πŸ”— :octocat:
  • SAM: Segment Anything, ICCV 2023, πŸ”— :octocat:
  • ViT: An Image is Worth 16x16 Words: Transformers for Image Recognition at Scale, arXiv 2020, πŸ”— :octocat:
  • DeiT: Training data-efficient image transformers & distillation through attention, ICML 2021, πŸ”— :octocat:
  • DINO: Emerging Properties in Self-Supervised Vision Transformers, ICCV 2021, πŸ”— :octocat:
  • DETR: End-to-End Object Detection with Transformers, ECCV 2020, πŸ”— :octocat:
  • Swin Transformer: Hierarchical Vision Transformer using Shifted Windows, ICCV 2021, πŸ”— :octocat:

Vision-Language Models (VLMs)

  • Simple Open-Vocabulary Object Detection with Vision Transformers, ECCV 2022, πŸ”— :octocat:

  • Learning to Detect and Segment for Open Vocabulary Object Detection, CVPR 2023, πŸ”—

  • PracticalDG: Perturbation Distillation on Vision-Language Models for Hybrid Domain Generalization, CVPR 2024, πŸ”— :octocat:

  • CLIP

    • Open-Vocabulary Object Detection

      • Open-vocabulary Object Detection via Vision and Language Knowledge Distillation, ICLR 2021, πŸ”— :octocat:
      • Detecting Twenty-thousand Classes using Image-level Supervision, ECCV 2022, πŸ”— :octocat:
      • Learning to Prompt for Open-Vocabulary Object Detection with Vision-Language Model, CVPR 2022, πŸ”— :octocat:
      • Open-Vocabulary DETR with Conditional Matching, ECCV 2022, πŸ”— :octocat:
      • F-VLM: Open-Vocabulary Object Detection upon Frozen Vision and Language Models, arXiv 2022, πŸ”—
      • Bridging the Gap between Object and Image-level Representations for Open-Vocabulary Detection, NeurIPS 2022, πŸ”— :octocat:
      • PromptDet: Towards Open-vocabulary Detection using Uncurated Images, ECCV 2022, πŸ”— :octocat:
      • DetCLIP: Dictionary-Enriched Visual-Concept Paralleled Pre-training for Open-world Detection, NeurIPS 2022, πŸ”—
      • Exploiting Unlabeled Data with Vision and Language Models for Object Detection, ECCV 2022, πŸ”— :octocat:
      • Learning Object-Language Alignments for Open-Vocabulary Object Detection, arXiv 2022, πŸ”— :octocat:
      • Open-Vocabulary One-Stage Detection with Hierarchical Visual-Language Knowledge Distillation, CVPR 2022, πŸ”— :octocat:
      • Open-Vocabulary Instance Segmentation via Robust Cross-Modal Pseudo-Labeling, CVPR 2022, πŸ”— :octocat:
      • Open Vocabulary Object Detection with Pseudo Bounding-Box Labels, ECCV 2022, πŸ”— :octocat:
      • Fine-grained Visual-Text Prompt-Driven Self-Training for Open-Vocabulary Object Detection, TNNLS 2022, πŸ”—
      • Zero-shot Object Detection Through Vision-Language Embedding Alignment, ICDM 2022, πŸ”— :octocat:
      • Aligning Bag of Regions for Open-Vocabulary Object Detection, CVPR 2023, πŸ”— :octocat:
      • Region-Aware Pretraining for Open-Vocabulary Object Detection with Vision Transformers, CVPR 2023, πŸ”—
      • Object-Aware Distillation Pyramid for Open-Vocabulary Object Detection, CVPR 2023, πŸ”— :octocat:
    • Open-Vocabulary Semantic Segmentation

      • Language-driven Semantic Segmentation, ICLR 2021, πŸ”— :octocat:
      • Semantic Segmentation In-the-Wild Without Seeing Any Segmentation Examples, arXiv 2021, πŸ”—
      • Extract Free Dense Labels from CLIP, ECCV 2022, πŸ”— :octocat:
      • Image Segmentation Using Text and Image Prompts, CVPR 2022, πŸ”— :octocat:
      • Scaling Open-Vocabulary Image Segmentation with Image-Level Labels, ECCV 2022, πŸ”— :octocat:
      • Decoupling Zero-Shot Semantic Segmentation, CVPR 2022, πŸ”— :octocat:
      • A Simple Baseline for Open-Vocabulary Semantic Segmentation with Pre-trained Vision-language Model, ECCV 2022, πŸ”— :octocat:
      • ReCo: Retrieve and Co-segment for Zero-shot Transfer, NeurIPS 2022, πŸ”— :octocat:
      • Open-vocabulary Semantic Segmentation with Frozen Vision-Language Models, arXiv 2022, πŸ”— :octocat:
      • Open-Vocabulary Semantic Segmentation with Mask-adapted CLIP, CVPR 2023, πŸ”— :octocat:
      • ZegCLIP: Towards Adapting CLIP for Zero-shot Semantic Segmentation, CVPR 2023, πŸ”— :octocat:
      • FreeSeg: Unified, Universal and Open-Vocabulary Image Segmentation, CVPR 2023, πŸ”— :octocat:
      • Exploring Open-Vocabulary Semantic Segmentation from CLIP Vision Encoder Distillation Only, ICCV 2023, πŸ”— :octocat:
      • CLIP-DIY: CLIP Dense Inference Yields Open-Vocabulary Semantic Segmentation For-Free, WACV 2024, πŸ”—
    • Open-Vocabulary Part Segmentation

      • Understanding Multi-Granularity for Open-Vocabulary Part Segmentation, NeurIPS 2025, πŸ”— :octocat:
    • Open-Vocabulary Customization

      • Open-Vocabulary Customization from CLIP via Data-Free Knowledge Distillation, ICLR 2025, πŸ”—
    • Weakly Supervised Semantic Segmentation

      • Cross Language Image Matching for Weakly Supervised Semantic Segmentation, CVPR 2022, πŸ”— :octocat:
      • CLIP is Also an Efficient Segmenter: A Text-Driven Approach for Weakly Supervised Semantic Segmentation, CVPR 2023, πŸ”— :octocat:
      • Learning Multi-Modal Class-Specific Tokens for Weakly Supervised Dense Object Localization, CVPR 2023, πŸ”— :octocat:
      • Question-Answer Cross Language Image Matching for Weakly Supervised Semantic Segmentation, ACM MM 2023, πŸ”— :octocat:
      • Prompting classes: Exploring the Power of Prompt Class Learning in Weakly Supervised Semantic Segmentation, WACV 2024, πŸ”— :octocat:
      • Frozen CLIP: A Strong Backbone for Weakly Supervised Semantic Segmentation, CVPR 2024, πŸ”— :octocat:
      • DIAL: Dense Image-text ALignment for Weakly Supervised Semantic Segmentation, ECCV 2024, πŸ”—
      • WeakCLIP: Adapting CLIP for Weakly-Supervised Semantic Segmentation, IJCV 2025, πŸ”— :octocat:
    • Prompt Learning

      • Learning to Prompt for Vision-Language Models, IJCV 2022, πŸ”— :octocat:
      • Conditional Prompt Learning for Vision-Language Models, CVPR 2022, πŸ”— :octocat:
      • DenseCLIP: Language-Guided Dense Prediction with Context-Aware Prompting, CVPR 2022, πŸ”— :octocat:
      • Domain Adaptation via Prompt Learning, T-NNLS 2023, πŸ”—
      • PromptKD: Unsupervised Prompt Distillation for Vision-Language Models, CVPR 2024, πŸ”— :octocat:
    • Generation and Editing

      • StyleCLIP: Text-Driven Manipulation of StyleGAN Imagery, ICCV 2021, πŸ”— :octocat:
      • StyleGAN-NADA: CLIP-Guided Domain Adaptation of Image Generators, ACM SIGGRAPH 2022, πŸ”— :octocat:
      • Zero-Shot Text-Guided Object Generation with Dream Fields, CVPR 2022, πŸ”— :octocat:
      • CLIP-NeRF: Text-and-Image Driven Manipulation of Neural Radiance Fields, CVPR 2022, πŸ”— :octocat:
      • Text2Mesh: Text-Driven Neural Stylization for Meshes, CVPR 2022, πŸ”— :octocat:
      • Decomposing NeRF for Editing via Feature Field Distillation, NeurIPS 2022, πŸ”— :octocat:
      • AvatarCLIP: Zero-Shot Text-Driven Generation and Animation of 3D Avatars, ACM Trans. Graph 2022, πŸ”—
      • CLIP-Forge: Towards Zero-Shot Text-to-Shape Generation, CVPR 2022, πŸ”— :octocat:
      • CLIP2StyleGAN: Unsupervised Extraction of StyleGAN Edit Directions, ACM SIGGRAPH 2022, πŸ”—
      • Text and Image Guided 3D Avatar Generation and Manipulation, WACV 2023, πŸ”—
      • Local 3D Editing via 3D Distillation of CLIP Knowledge, CVPR 2023, πŸ”—
    • Person Re-Identification

      • Distilling CLIP with Dual Guidance for Learning Discriminative Human Body Shape Representation, CVPR 2024, πŸ”—
    • Zero-Shot Human-Object Interaction (HOI) Detection

      • Exploiting CLIP for Zero-shot HOI Detection Requires Knowledge Distillation at Multiple Levels, WACV 2024, πŸ”— :octocat:
    • Open-Vocabulary Out-of-Distribution Classification

      • Distilling Large Vision-Language Model with Out-of-Distribution Generalizability, ICCV 2023, πŸ”— :octocat:
    • Domain Generalization

      • A Sentence Speaks a Thousand Images: Domain Generalization through Distilling CLIP with Language Guidance, ICCV 2023, πŸ”—
    • Video-Language Retrieval

      • CLIPPING: Distilling CLIP-Based Models with a Student Base for Video-Language Retrieval, CVPR 2023, πŸ”—
    • Affordance Detection

      • What does CLIP know about peeling a banana?, CVPR 2024, πŸ”—
    • Video Highlight Detection

      • Unleash the Potential of CLIP for Video Highlight Detection, CVPR 2024, πŸ”—
    • Monocular Depth Estimation

      • CaBins: CLIP-based Adaptive Bins for Monocular Depth Estimation, CVPR 2024, πŸ”—
    • Multiple Tasks (Image Classification, Object Detection, Semantic Segmentation, Instance Segmentation, Image-Text Retrieval)

      • MaskCLIP: Masked Self-Distillation Advances Contrastive Language-Image Pretraining, CVPR 2023, πŸ”— :octocat:
      • TinyCLIP: CLIP Distillation via Affinity Mimicking and Weight Inheritance, ICCV 2023, πŸ”— :octocat:
      • CLIP-KD: An Empirical Study of CLIP Model Distillation, CVPR 2024, πŸ”— :octocat:
      • CLIP-Embed-KD: Computationally Efficient Knowledge Distillation Using Embeddings as Teachers, HPEC 2024, πŸ”— :octocat:

Segment Anything (SAM)

  • Road Network Graph Extraction

    • Segment Anything Model for Road Network Graph Extraction, CVPR 2024, πŸ”—
  • Polyp Segmentation

    • PP-SAM: Perturbed Prompts for Robust Adaptation of Segment Anything Model for Polyp Segmentation, CVPR 2024, πŸ”— :octocat:
  • Image Restoration

    • Distilling Semantic Priors from SAM to Efficient Image Restoration Models, CVPR 2024, πŸ”—
  • Weakly Supervised Semantic Segmentation

    • Segment Anything Model (SAM) Enhanced Pseudo Labels for Weakly Supervised Semantic Segmentation, arXiv 2023, πŸ”— :octocat:
    • An Alternative to WSSS? An Empirical Study of the Segment Anything Model (SAM) on Weakly-Supervised Semantic Segmentation Problems, arXiv 2023, πŸ”— :octocat:
    • Segment Anything is A Good Pseudo-label Generator for Weakly Supervised Semantic Segmentation, arXiv 2023, πŸ”—
    • Weakly-Supervised Semantic Segmentation with Image-Level Labels: from Traditional Models to Foundation Models, arXiv 2023, πŸ”—
    • From SAM to CAMs: Exploring Segment Anything Model for Weakly Supervised Semantic Segmentation, CVPR 2024, πŸ”— :octocat:

Multi-model Applications

  • SAM + CLIP

    • Foundation Model Assisted Weakly Supervised Semantic Segmentation, WACV 2023, πŸ”— :octocat:
    • SAM-CLIP: Merging Vision Foundation Models towards Semantic and Spatial Understanding, CVPR 2024, πŸ”—
    • Test-Time Adaptation with SaLIP: A Cascade of SAM and CLIP for Zero shot Medical Image Segmentation, CVPR 2024, πŸ”— :octocat:
  • DINO + CLIP

    • LERF: Language Embedded Radiance Fields, ICCV 2023, πŸ”— :octocat:
    • CLIP-DINOiser: Teaching CLIP a few DINO tricks for open-vocabulary semantic segmentation, ECCV 2025, πŸ”— :octocat:
  • DETR + CLIP

    • T-Rex2: Towards Generic Object Detection via Text-Visual Prompt Synergy, ECCV 2024, πŸ”— :octocat:
  • SAM + GDINO

    • Grounded SAM: Assembling Open-World Models for Diverse Visual Tasks, arXiv 2024, πŸ”— :octocat:
    • Prompting Foundational Models for Omni-supervised Instance Segmentation, CVPR 2024, πŸ”—
  • Others

    • FeatureNeRF: Learning Generalizable NeRFs by Distilling Foundation Models, ICCV 2023, πŸ”— :octocat:
    • DIME-FM: DIstilling Multimodal and Efficient Foundation Models, ICCV 2023, πŸ”— :octocat:
    • AM-RADIO: Agglomerative Vision Foundation Model -- Reduce All Domains Into One, CVPR 2024, πŸ”— :octocat:
    • Theia: Distilling Diverse Vision Foundation Models for Robot Learning, arXiv 2024, πŸ”— :octocat:
    • Strategies to Leverage Foundational Model Knowledge in Object Affordance Grounding, CVPR 2024, πŸ”—
    • Enrich Distill and Fuse: Generalized Few-Shot Semantic Segmentation in Remote Sensing Leveraging Foundation Model's Assistance, CVPR 2024, πŸ”—
    • Exploring the Benefits of Vision Foundation Models for Unsupervised Domain Adaptation, CVPR 2024, πŸ”— :octocat:
    • Distilling Knowledge from Multiple Foundation Models for Zero-shot Image classification, PlosOne 2024, πŸ”— :octocat:

Vision Transformers

  • Vision Transformer (ViT)

    • Supervised Masked Knowledge Distillation for Few-Shot Transformers, CVPR 2023, πŸ”— :octocat:
    • Masked Autoencoders Enable Efficient Knowledge Distillers, CVPR 2023, πŸ”— :octocat:
    • Distilling Self-Supervised Vision Transformers for Weakly-Supervised Few-Shot Classification & Segmentation, CVPR 2023, πŸ”— :octocat:
    • TinyMIM: An Empirical Study of Distilling MIM Pre-trained Models, CVPR 2023, πŸ”— :octocat:
    • Incrementer: Transformer for Class-Incremental Semantic Segmentation With Knowledge Distillation Focusing on Old Class, CVPR 2023, πŸ”—
    • Cumulative Spatial Knowledge Distillation for Vision Transformers, ICCV 2023, πŸ”— :octocat:
    • Feature-based Knowledge Distillation for Vision Transformers, CVPR 2024, πŸ”— :octocat:
  • Distillation with No Labels (DINO)

    • Unsupervised Object Discovery

      • Localizing Objects with Self-Supervised Transformers and no Labels, BMVC 2021, πŸ”— :octocat:
      • Self-Supervised Transformers for Unsupervised Object Discovery using Normalized Cut, CVPR 2022, πŸ”— :octocat:
      • Unsupervised Object Localization: Observing the Background to Discover Objects, CVPR 2023, πŸ”— :octocat:
      • Cut and Learn for Unsupervised Object Detection and Instance Segmentation, CVPR 2023, πŸ”— :octocat:
      • Grounding DINO: Marrying DINO with Grounded Pre-Training for Open-Set Object Detection, ECCV 2024, πŸ”— :octocat:
      • DIOD: Self-Distillation Meets Object Discovery, CVPR 2024, πŸ”— :octocat:
    • Open-world Counting

      • CountGD: Multi-Modal Open-World Counting, NeurIPS 2025, πŸ”— :octocat:
    • Zero-shot Semantic Segmentation

      • Diffusion Models for Open-Vocabulary Segmentation, arXiv 2023, πŸ”—
  • DEtection TRansformer (DETR)

    • Knowledge Distillation for Detection Transformer with Consistent Distillation Points Sampling, CVPR 2024, πŸ”—
    • DETRDistill: A Universal Knowledge Distillation Framework for DETR-families, ICCV 2023, πŸ”— :octocat:
  • Swin Transformer

    • Per-Pixel Classification is Not All You Need for Semantic Segmentation, NeurIPS 2021, πŸ”— :octocat:
    • Dynamic Feature Regularized Loss for Weakly Supervised Semantic Segmentation, arXiv 2021, πŸ”—
    • SwinNet: Swin Transformer Drives Edge-Aware RGB-D and RGB-T Salient Object Detection, TCSVT 2021, πŸ”— :octocat:
    • SwinTrack: A Simple and Strong Baseline for Transformer Tracking, NeurIPS 2022, πŸ”— :octocat:
    • SwinF: Swin Transformer with feature fusion in target detection, JPCS 2022, πŸ”—
    • Self-Supervised Pre-Training of Swin Transformers for 3D Medical Image Analysis, CVPR 2022, πŸ”— :octocat:
    • SSformer: A Lightweight Transformer for Semantic Segmentation, MMSP 2022, πŸ”— :octocat:
    • Weakly Supervised Intracranial Hemorrhage Segmentation using Head-Wise Gradient-Infused Self-Attention Maps from a Swin Transformer in Categorical Learning, arXiv 2023, πŸ”— :octocat:
    • HiFormer: Hierarchical Multi-scale Representations Using Transformers for Medical Image Segmentation, WACV 2023, πŸ”— :octocat:
    • Swin-Fusion: Swin-Transformer with Feature Fusion for Human Action Recognition, NPL 2023, πŸ”—
    • Pyramid Swin Transformer: Different-Size Windows Swin Transformer for Image Classification and Object Detection, VISIGRAPP 2023, πŸ”—
    • SCUNet++: Swin-UNet and CNN Bottleneck Hybrid Architecture with Multi-Fusion Dense Skip Connection for Pulmonary Embolism CT Image Segmentation, WACV 2024, πŸ”— :octocat:
    • Leveraging Swin Transformer for Local-to-Global Weakly Supervised Semantic Segmentation, MVIP 2024, πŸ”— :octocat:

Self-Supervised Learning

  • A simple framework for contrastive learning of visual representations, NeurIPS 2020, πŸ”— :octocat:
  • Momentum contrast for unsupervised visual representation learning, CVPR 2020, πŸ”— :octocat:
  • Improved Baselines with Momentum Contrastive Learning, arXiv 2020, πŸ”— :octocat:
  • Bootstrap your own latent-a new approach to self-supervised learning, NeurIPS 2020, πŸ”— :octocat:
  • Unsupervised learning of visual features by contrasting cluster assignments, NeurIPS 2020, πŸ”— :octocat:
  • Self-Supervised Learning of Pretext-Invariant Representations, CVPR 2020, πŸ”— :octocat:
  • CompRess: Self-Supervised Learning by Compressing Representations, NeurIPS 2020, πŸ”— :octocat:
  • SimReg: Regression as a Simple Yet Effective Tool for Self-supervised Knowledge Distillation, BMVC 2021, πŸ”— :octocat:
  • ISD: Self-Supervised Learning by Iterative Similarity Distillation, ICCV 2021, πŸ”— :octocat:
  • SEED: Self-supervised Distillation For Visual Representation, ICLR 2021, πŸ”— :octocat:
  • Simple Distillation Baselines for Improving Small Self-supervised Models, ICCV 2021, πŸ”— :octocat:
  • Emerging Properties in Self-Supervised Vision Transformers, ICCV 2021, πŸ”— :octocat:
  • Bag of Instances Aggregation Boosts Self-supervised Distillation, ICLR 2022, πŸ”—
  • DisCo: Remedying Self-supervised Learning on Lightweight Models with Distilled Contrastive Learning, ECCV 2022, πŸ”— :octocat:
  • Auxiliary Learning for Self-Supervised Video Representation via Similarity-Based Knowledge Distillation, CVPR 2022, πŸ”— :octocat:
  • Masked Video Distillation: Rethinking Masked Feature Modeling for Self-Supervised Video Representation Learning, CVPR 2023, πŸ”— :octocat:
  • Multi-Mode Online Knowledge Distillation for Self-Supervised Visual Representation Learning, CVPR 2023, πŸ”— :octocat:
  • DINOv2: Learning Robust Visual Features without Supervision, TMLR 2024, πŸ”— :octocat:

Diffusion Models

  • Knowledge Distillation in Iterative Generative Models for Improved Sampling Speed, arXiv 2021, πŸ”— :octocat:
  • Progressive Distillation for Fast Sampling of Diffusion Models, ICLR 2022, πŸ”— :octocat:
  • Accelerating diffusion sampling with classifier-based feature distillation, ICME 2023, πŸ”— :octocat:
  • Consistency Models, PMLR 2023, πŸ”— :octocat:
  • Fast Sampling of Diffusion Models via Operator Learning, PMLR 2023, πŸ”—
  • Diffusion-GAN: Training GANs with Diffusion, ICLR 2023, πŸ”— :octocat:
  • ProlificDreamer: High-Fidelity and Diverse Text-to-3D Generation with Variational Score Distillation, NeurIPS 2023, πŸ”— :octocat:
  • Diff-Instruct: A Universal Approach for Transferring Knowledge From Pre-trained Diffusion Models, NeurIPS 2023, πŸ”— :octocat:
  • Latent Consistency Models: Synthesizing High-Resolution Images with Few-Step Inference, arXiv 2023, πŸ”— :octocat:
  • Adversarial Diffusion Distillation, ECCV 2024, πŸ”— :octocat:
  • Improved Techniques for Training Consistency Models, ICLR 2024, πŸ”—
  • Relational diffusion distillation for efficient image generation, ACM MM 2024, πŸ”— :octocat:
  • Multistep Distillation of Diffusion Models via Moment Matching, NeurIPS 2024, πŸ”—
  • EM Distillation for One-step Diffusion Models, NeurIPS 2024, πŸ”—
  • Simple and Fast Distillation of Diffusion Models, NeurIPS 2024, πŸ”— :octocat:
  • One-step Diffusion with Distribution Matching Distillation, CVPR 2024, πŸ”— :octocat:
  • Simplifying, Stabilizing and Scaling Continuous-Time Consistency Models, arXiv 2024, πŸ”— :octocat:
  • Consistency Models Made Easy, arXiv 2024, πŸ”— :octocat:
  • TLCM: Training-efficient Latent Consistency Model for Image Generation with 2-8 Steps, arXiv 2024, πŸ”— :octocat:

Knowledge Distillation in Visual Recognition

Object Detection

  • Learning Efficient Object Detection Models with Knowledge Distillation, NIPS 2017, πŸ”—
  • Distilling Object Detectors with Fine-grained Feature Imitation, CVPR 2019, πŸ”— :octocat:
  • An end-to-end architecture for class-incremental object detection with knowledge distillation, ICME 2019, πŸ”— :octocat:
  • LabelEnc: A New Intermediate Supervision Method for Object Detection, ECCV 2020, πŸ”— :octocat:
  • MimicDet: Bridging the Gap Between One-Stage and Two-Stage Object Detection, ECCV 2020, πŸ”—
  • Improve object detection with feature-based knowledge distillation: Towards accurate and efficient detectors, ICLR 2021, πŸ”— :octocat:
  • Distilling Object Detectors via Decoupled Features, CVPR 2021, πŸ”— :octocat:
  • General Instance Distillation for Object Detection, CVPR 2021, πŸ”—
  • Instance-Conditional Knowledge Distillation for Object Detection, NIPS 2021, πŸ”— :octocat:
  • Improving Object Detection by Label Assignment Distillation, WACV 2022, πŸ”— :octocat:
  • Distilling image classifiers in object detectors, NIPS 2021, πŸ”— :octocat:
  • G-DetKD: Towards General Distillation Framework for Object Detectors via Contrastive and Semantic-guided Feature Imitation, ICCV 2021, πŸ”—
  • Distilling Object Detectors with Feature Richness, NIPS 2021, πŸ”— :octocat:
  • Prediction-Guided Distillation for Dense Object Detection, ECCV 2022, πŸ”— :octocat:
  • Focal and Global Knowledge Distillation for Detectors, CVPR 2022, πŸ”— :octocat:
  • GLAMD: Global and Local Attention Mask Distillation for Object Detectors, ECCV 2022, πŸ”—
  • Masked Generative Distillation, ECCV 2022, πŸ”— :octocat:
  • LGD: Label-guided Self-distillation for Object Detection, AAAI 2022, πŸ”— :octocat:
  • Knowledge Distillation for Object Detection via Rank Mimicking and Prediction-Guided Feature Imitation, AAAI 2022, πŸ”—
  • Task-balanced distillation for object detection, Pattern Recognition 2023, πŸ”—
  • Structured Knowledge Distillation for Accurate and Efficient Object Detection, TPAMI 2023, πŸ”—
  • Spatial Self-Distillation for Object Detection with Inaccurate Bounding Boxes, ICCV 2023, πŸ”— :octocat:
  • Dual Relation Knowledge Distillation for Object Detection, IJCAI 2023, πŸ”—
  • ScaleKD: Distilling Scale-Aware Knowledge in Small Object Detector, CVPR 2023, πŸ”—
  • Multi-level Logit Distillation, CVPR 2023, πŸ”— :octocat:
  • Bridging Cross-task Protocol Inconsistency for Distillation in Dense Object Detection, ICCV 2023, πŸ”— :octocat:
  • Distilling DETR with Visual-Linguistic Knowledge for Open-Vocabulary Object Detection, ICCV 2023, πŸ”—
  • UniKD: Universal Knowledge Distillation for Mimicking Homogeneous or Heterogeneous Object Detectors, ICCV 2023, πŸ”—
  • Texture-Guided Saliency Distilling for Unsupervised Salient Object Detection, CVPR 2023, πŸ”— :octocat:
  • Object-Aware Distillation Pyramid for Open-Vocabulary Object Detection, CVPR 2023, πŸ”— :octocat:
  • Localization distillation for object detection, IEEE TPAMI 2023, πŸ”— :octocat:
  • CrossKD: Cross-Head Knowledge Distillation for Dense Object Detection, CVPR 2024, πŸ”— :octocat:
  • MSSD: multi-scale self-distillation for object detection, Visual Intelligence 2024, πŸ”—
  • CLIPSelf: Vision Transformer Distills Itself for Open-Vocabulary Dense Prediction, ICLR 2024, πŸ”— :octocat:
  • Gradient-Guided Knowledge Distillation for Object Detectors, WACV 2024, πŸ”—
  • Efficient Feature Distillation for Zero-Shot Annotation Object Detection, WACV 2024, πŸ”— :octocat:

Super Resolution

  • Image Super-Resolution Using Knowledge Distillation, ACCV 2018, πŸ”—
  • Feature-Affinity Based Knowledge Distillation for Efficient Image Super-Resolution, ICIP 2020, πŸ”— :octocat:
  • Learning with Privileged Information for Efficient Image Super-Resolution, ECCV 2020, πŸ”— :octocat:
  • Data-Free Knowledge Distillation for Image Super-Resolution, CVPR 2021, πŸ”—
  • Towards Compact Single Image Super-Resolution via Contrastive Self-distillation, IJCAI 2021, πŸ”— :octocat:
  • DIPNet: Efficiency Distillation and Iterative Pruning for Image Super-Resolution, CVPRW 2023, πŸ”— :octocat:
  • Generative Adversarial Super-Resolution at the edge with knowledge distillation, Engineering Applications of Artificial Intelligence 2023, πŸ”— :octocat:
  • Hybrid knowledge distillation from intermediate layers for efficient Single Image Super-Resolution, Neurocomputing 2023, πŸ”—
  • Dual cross knowledge distillation for image super-resolution, Journal of Visual Communication and Image Representation 2023, πŸ”—
  • Pairwise Distance Distillation for Unsupervised Real-World Image Super-Resolution, ECCV 2024, πŸ”— :octocat:
  • Knowledge Distillation for Single Image Super-Resolution via Contrastive Learning, ICMR 2024, πŸ”—
  • MTKD: Multi-Teacher Knowledge Distillation for Image Super-Resolution, ECCV 2024, πŸ”— :octocat:
  • You Only Need One Step: Fast Super-Resolution with Stable Diffusion via Scale Distillation, ECCV 2024, πŸ”—
  • Attention Guidance Distillation Network for Efficient Image Super-Resolution, CVPRW 2024, πŸ”— :octocat:
  • SinSR: Diffusion-Based Image Super-Resolution in a Single Step, CVPR 2024, πŸ”— :octocat:
  • OSFFNet: Omni-Stage Feature Fusion Network for Lightweight Image Super-Resolution, AAAI 2024, πŸ”—
  • Semantic Super-Resolution via Self-Distillation and Adversarial Learning, IEEE Access 2024, πŸ”—

Image Segmentation

  • Structured knowledge distillation for semantic segmentation, CVPR 2019, πŸ”— :octocat:
  • Knowledge Adaptation for Efficient Semantic Segmentation, CVPR 2019, πŸ”—
  • Intra-class Feature Variation Distillation for Semantic Segmentation, ECCV 2020, πŸ”— :octocat:
  • Domain Adaptive Knowledge Distillation for Driving Scene Semantic Segmentation, WACVW 2020, πŸ”— :octocat:
  • Channel-wise Knowledge Distillation for Dense Prediction, ICCV 2021, πŸ”— :octocat:
  • Double Similarity Distillation for Semantic Image Segmentation, IEEE Transactions on Image Processing (TIP) 2021, πŸ”—
  • Robust Semantic Segmentation With Multi-Teacher Knowledge Distillation, IEEE Access 2021, πŸ”—
  • Cross-Image Relational Knowledge Distillation for Semantic Segmentation, CVPR 2022, πŸ”— :octocat:
  • Structural and Statistical Texture Knowledge Distillation for Semantic Segmentation, CVPR 2022, πŸ”—
  • Adaptive Perspective Distillation for Semantic Segmentation, TPAMI 2022, πŸ”— :octocat:
  • Masked Generative Distillation, ECCV 2022, πŸ”— :octocat:
  • Class Similarity Weighted Knowledge Distillation for Continual Semantic Segmentation, CVPR 2022, πŸ”—
  • Cross-Domain Correlation Distillation for Unsupervised Domain Adaptation in Nighttime Semantic Segmentation, CVPR 2022, πŸ”—
  • DiGA: Distil to Generalize and then Adapt for Domain Adaptive Semantic Segmentation, CVPR 2023, πŸ”— :octocat:
  • A Good Student is Cooperative and Reliable: CNN-Transformer Collaborative Learning for Semantic Segmentation, ICCV 2023, πŸ”—
  • Endpoints Weight Fusion for Class Incremental Semantic Segmentation, CVPR 2023, πŸ”— :octocat:
  • Multi-Task Learning with Knowledge Distillation for Dense Prediction, ICCV 2023, πŸ”—
  • Hierarchical Dense Correlation Distillation for Few-Shot Segmentation, CVPR 2023, πŸ”—
  • Distilling Self-Supervised Vision Transformers for Weakly-Supervised Few-Shot Classification & Segmentation, CVPR 2023, πŸ”— :octocat:
  • Incrementer: Transformer for Class-Incremental Semantic Segmentation With Knowledge Distillation Focusing on Old Class, CVPR 2023, πŸ”—
  • FAKD: Feature Augmented Knowledge Distillation for Semantic Segmentation, WACV 2024, πŸ”—
  • Knowledge Distillation for Efficient Instance Semantic Segmentation with Transformers, CVPRW 2024, πŸ”—
  • CLIPSelf: Vision Transformer Distills Itself for Open-Vocabulary Dense Prediction, ICLR 2024, πŸ”— :octocat:
  • P2Seg: Pointly-supervised Segmentation via Mutual Distillation, ICLR 2024, πŸ”— :octocat:
  • Rethinking Knowledge Distillation With Raw Features for Semantic Segmentation, WACV 2024, πŸ”—
  • BPKD: Boundary Privileged Knowledge Distillation for Semantic Segmentation, WACV 2024, πŸ”— :octocat:
  • Guided Distillation for Semi-Supervised Instance Segmentation, WACV 2024, πŸ”— :octocat:
  • Distilling efficient Vision Transformers from CNNs for semantic segmentation, Pattern Recognition 2025, πŸ”—
  • AICSD: Adaptive Inter-Class Similarity Distillation for Semantic Segmentation, Multimedia Tools and Applications 2025, πŸ”— :octocat:

Depth Estimation

  • Knowledge Distillation for Fast and Accurate Monocular Depth Estimation on Mobile Devices, CVPRW 2021, πŸ”—
  • Realtime single image depth perception in the wild with handheld devices, Sensors 2021, πŸ”—
  • Excavating the Potential Capacity of Self-Supervised Monocular Depth Estimation, ICCV 2021, πŸ”— :octocat:
  • Self-distilled Feature Aggregation for Self-supervised Monocular Depth Estimation, ECCV 2022, πŸ”—
  • Attention-based depth distillation with 3d-aware positional encoding for monocular 3d object detection, AAAI 2022, πŸ”—
  • Boosting Light-Weight Depth Estimation Via Knowledge Distillation, KSEM 2023, πŸ”— :octocat:
  • Urcdc-depth: Uncertainty rectified cross-distillation with cutflip for monocular depth estimation, IEEE Transactions on Multimedia 2023, πŸ”— :octocat:
  • Multi-task learning with knowledge distillation for dense prediction, CVPR 2023, πŸ”—
  • Attention-Based Knowledge Distillation in Scene Recognition: The Impact of a DCT-Driven Loss, IEEE Transactions on Circuits and Systems for Video Technology 2023, πŸ”—
  • Monocular Depth Estimation from a Fisheye Camera Based on Knowledge Distillation, Sensors 2023, πŸ”—
  • 3D Distillation: Improving Self-Supervised Monocular Depth Estimation on Reflective Surfaces, ICCV 2023, πŸ”—
  • Multi-Task Learning with Knowledge Distillation for Dense Prediction, ICCV 2023, πŸ”—
  • GasMono: Geometry-Aided Self-Supervised Monocular Depth Estimation for Indoor Scenes, ICCV 2023, πŸ”— :octocat:
  • Depth Anywhere: Enhancing 360 Monocular Depth Estimation via Perspective Distillation and Unlabeled Data Augmentation, NIPS 2024, πŸ”— :octocat:
  • Structure-Centric Robust Monocular Depth Estimation via Knowledge Distillation, ACCV 2024, πŸ”—
  • MonoProb: Self-Supervised Monocular Depth Estimation With Interpretable Uncertainty, WACV 2024, πŸ”— :octocat:
  • Monocular Depth Estimation via Self-Supervised Self-Distillation, Sensors 2024, πŸ”—
  • MAL: Motion-Aware Loss with Temporal and Distillation Hints for Self-Supervised Depth Estimation, ICRA 2024, πŸ”— :octocat:
  • Stereo-Matching Knowledge Distilled Monocular Depth Estimation Filtered by Multiple Disparity Consistency, ICASSP 2024, πŸ”—

Medical Image Analysis

  • Categorical Relation-Preserving Contrastive Knowledge Distillation for Medical Image Classification, MICCAI 2021, πŸ”— :octocat:
  • Efficient Medical Image Segmentation Based on Knowledge Distillation, IEEE Transactions on Medical Imaging (TMI) 2021, πŸ”— :octocat:
  • DeSD: Self-Supervised Learning with Deep Self-Distillation for 3D Medical Image Segmentation, MICCAI 2022, πŸ”— :octocat:
  • Efficient Biomedical Instance Segmentation via Knowledge Distillation, MICCAI 2022, πŸ”—
  • Flat-aware Cross-stage Distilled Framework for Imbalanced Medical Image Classification, MICCAI 2022, πŸ”—
  • Free Lunch for Surgical Video Understanding by Distilling Self-Supervisions, MICCAI 2022, πŸ”— :octocat:
  • Simcvd: Simple contrastive voxel-wise representation distillation for semi-supervised medical image segmentation, IEEE Trans. Med. Imaging 2022, πŸ”—
  • Deep Mutual Distillation for Semi-Supervised Medical Image Segmentation, MICCAI 2023, πŸ”— :octocat:
  • Distilling BlackBox to Interpretable models for Efficient Transfer Learning, MICCAI 2023, πŸ”— :octocat:
  • Learnable Cross-modal Knowledge Distillation for Multi-modal Learning with Missing Modality, MICCAI 2023, πŸ”—
  • Self-distillation for surgical action recognition, MICCAI 2023, πŸ”— :octocat:
  • Semi-supervised Pathological Image Segmentation via Cross Distillation of Multiple Attentions, MICCAI 2023, πŸ”— :octocat:
  • Bootstrapping Chest CT Image Understanding by Distilling Knowledge from X-ray Expert Models, CVPR 2024, πŸ”—
  • Reverse Knowledge Distillation: Training a Large Model Using a Small One for Retinal Image Matching on Limited Data, WACV 2024, πŸ”—
  • Learning Robust Shape Regularization for Generalizable Medical Image Segmentation, IEEE Trans. Med. Imaging 2024, πŸ”— :octocat:
  • Enhancing Medical Imaging with GANs Synthesizing Realistic Images from Limited Data, ICETCI 2024, πŸ”—
  • Exploring Generalizable Distillation for Efficient Medical Image Segmentation, IEEE Journal of Biomedical and Health Informatics 2024, πŸ”— :octocat:
  • Confidence Matters: Enhancing Medical Image Classification Through Uncertainty-Driven Contrastive Self-Distillation, MICCAI 2024, πŸ”— :octocat:
  • DES-SAM: Distillation-Enhanced Semantic SAM for Cervical Nuclear Segmentation with Box Annotation, MICCAI 2024, πŸ”— :octocat:
  • Hallucinated Style Distillation for Single Domain Generalization in Medical Image Segmentation, MICCAI 2024, πŸ”—
  • Progressively Correcting Soft Labels via Teacher Team for Knowledge Distillation in Medical Image Segmentation, MICCAI 2024, πŸ”—
  • Reprogramming Distillation for Medical Foundation Models, MICCAI 2024, πŸ”—
  • HDKD: Hybrid Data-Efficient Knowledge Distillation Network for Medical Image Classification, Engineering Applications of Artificial Intelligence 2024, πŸ”— :octocat:
  • Shape-Intensity Knowledge Distillation For Robust Medical Image Segmentation, Frontiers of Computer Science 2024, πŸ”— :octocat:

Object Tracking

  • There is More than Meets the Eye: Self-Supervised Multi-Object Detection and Tracking with Sound by Distilling Multimodal Knowledge, CVPR 2021, πŸ”— :octocat:
  • Ensemble learning with siamese networks for visual tracking, Neurocomputing 2021, πŸ”—
  • Distilled Siamese Networks for Visual Tracking, TPAMI 2022, πŸ”—
  • SDSTrack: Self-Distillation Symmetric Adapter Learning for Multi-Modal Visual Object Tracking, CVPR 2024, πŸ”— :octocat:
  • Event Stream-based Visual Object Tracking: A High-Resolution Benchmark Dataset and A Novel Baseline, CVPR 2024, πŸ”— :octocat:
  • Object Knowledge Distillation for Joint Detection and Tracking in Satellite Videos, IEEE Transactions on Geoscience and Remote Sensing 2024, πŸ”—

Face Recognition

  • Teacher Supervises Students How to Learn From Partially Labeled Images for Facial Landmark Detection, ICCV 2019, πŸ”— :octocat:
  • Fair Feature Distillation for Visual Recognition, CVPR 2021, πŸ”—
  • Rectifying the Data Bias in Knowledge Distillation, ICCV 2021, πŸ”—
  • Teaching Where to Look: Attention Similarity Knowledge Distillation for Low Resolution Face Recognition, ECCV 2022, πŸ”— :octocat:
  • Evaluation-oriented knowledge distillation for deep face recognition, CVPR 2022, πŸ”—
  • Rethinking Feature-Based Knowledge Distillation for Face Recognition, CVPR 2023, πŸ”—
  • Probabilistic Knowledge Distillation of Face Ensembles, CVPR 2023, πŸ”—
  • SynthDistill: Face Recognition with Knowledge Distillation from Synthetic Data, IJCB 2023, πŸ”— :octocat:
  • AdaDistill: Adaptive Knowledge Distillation for Deep Face Recognition, ECCV 2024, πŸ”— :octocat:
  • How Knowledge Distillation Mitigates the Synthetic Gap in Fair Face Recognition, ECCV Workshop 2024, πŸ”— :octocat:
  • MST-KD: Multiple Specialized Teachers Knowledge Distillation for Fair Face Recognition, ECCV Workshop 2024, πŸ”— :octocat:
  • Enhanced Face Recognition using Intra-class Incoherence Constraint, ICLR 2024, πŸ”—
  • ProS: Facial Omni-Representation Learning via Prototype-Based Self-Distillation, WACV 2024, πŸ”—
  • AI-KD: Towards Alignment Invariant Face Image Quality Assessment Using Knowledge Distillation, IWBF 2024, πŸ”— :octocat:

Action Recognition

  • Structural Knowledge Distillation for Efficient Skeleton-Based Action Recognition, IEEE Trans. Image Processing 2021, πŸ”— :octocat:
  • Video Pose Distillation for Few-Shot, Fine-Grained Sports Action Recognition, ICCV 2021, πŸ”— :octocat:
  • Learning an Augmented RGB Representation with Cross-modal Knowledge Distillation for Action Detection, ICCV 2021, πŸ”—
  • Privileged Knowledge Distillation for Online Action Detection, Pattern recognition 2022, πŸ”—
  • Multimodal Distillation for Egocentric Action Recognition, ICCV 2023, πŸ”— :octocat:
  • FSAR: Federated Skeleton-based Action Recognition with Adaptive Topology Structure and Knowledge Distillation, ICCV 2023, πŸ”—
  • Decomposed Cross-Modal Distillation for RGB-Based Temporal Action Detection, CVPR 2023, πŸ”—
  • Generative Model-Based Feature Knowledge Distillation for Action Recognition, AAAI 2024, πŸ”— :octocat:
  • FROSTER: Frozen CLIP is A Strong Teacher for Open-Vocabulary Action Recognition, ICLR 2024, πŸ”— :octocat:

Pose Estimation

  • Dynamic Kernel Distillation for Efficient Pose Estimation in Videos, ICCV 2019, πŸ”—
  • Fast Human Pose Estimation, CVPR 2019, πŸ”— :octocat:
  • Online Knowledge Distillation for Efficient Pose Estimation, ICCV 2021, πŸ”— :octocat:
  • When Human Pose Estimation Meets Robustness: Adversarial Algorithms and Benchmarks, CVPR 2021, πŸ”— :octocat:
  • From Synthetic to Real: Unsupervised Domain Adaptation for Animal Pose Estimation, CVPR 2021, πŸ”— :octocat:
  • Combining Weight Pruning and Knowledge Distillation for CNN Compression, CVPR Workshop 2021, πŸ”—
  • AlphaPose: Whole-Body Regional Multi-Person Pose Estimation and Tracking in Real-Time, IEEE TPAMI 2022, πŸ”— :octocat:
  • ViTPose: Simple Vision Transformer Baselines for Human Pose Estimation, NeurIPS 2022, πŸ”— :octocat:
  • DistilPose: Tokenized Pose Regression With Heatmap Distillation, CVPR 2023, πŸ”— :octocat:
  • Knowledge Distillation for 6D Pose Estimation by Aligning Distributions of Local Predictions, CVPR 2023, πŸ”— :octocat:
  • Effective Whole-Body Pose Estimation with Two-Stages Distillation, ICCV 2023, πŸ”— :octocat:
  • HRPose: Real-Time High-Resolution 6D Pose Estimation Network Using Knowledge Distillation, Chinese Journal of Electronics 2023, πŸ”—
  • SDPose: Tokenized Pose Estimation via Circulation-Guide Self-Distillation, CVPR 2024, πŸ”— :octocat:

Image Retrieval

  • Uncertainty-aware multi-shot knowledge distillation for image-based object re-identification, AAAI 2020, πŸ”—
  • Robust re-identification by multiple views knowledge distillation, ECCV 2020, πŸ”— :octocat:
  • Relationship-Preserving Knowledge Distillation for Zero-Shot Sketch Based Image Retrieval, ACMM 2021, πŸ”—
  • Asymmetric metric learning for knowledge transfer, CVPR 2021, πŸ”—
  • Contextual Similarity Distillation for Asymmetric Image Retrieval, CVPR 2022, πŸ”—
  • Large-to-small image resolution asymmetry in deep metric learning, WACV 2023, πŸ”—
  • Towards a Smaller Student: Capacity Dynamic Distillation for Efficient Image Retrieval, CVPR 2023, πŸ”—
  • D3still: Decoupled Differential Distillation for Asymmetric Image Retrieval, CVPR 2024, πŸ”—
  • LSTKC: Long Short-Term Knowledge Consolidation for Lifelong Person Re-identification, AAAI 2024, πŸ”— :octocat:
  • Pairwise difference relational distillation for object re-identification, Pattern Recognition 2024, πŸ”—