- Large Language Models
- Foundation Models
- Self-Supervised Learning
- Diffusion Models
- Visual Recognition
- TinyBERT: Distilling BERT for Natural Language Understanding, arXiv 2019, π
- Distilbert, a distilled version of bert: smaller, faster, cheaper and lighter, arXiv 2019, π
- Patient Knowledge Distillation for BERT Model Compression, arXiv 2019, π
- Well-read students learn better: On the importance of pre-training compact models, arXiv 2019, π
- Xtremedistil: Multi-stage distillation for massive multilingual models, arXiv 2020, π
- MobileBERT: a Compact Task-Agnostic BERT for Resource-Limited Devices, ACL 2020, π
- MINILM: deep self-attention distillation for task-agnostic compression of pre-trained transformers, NeurIPS 2020, π
- xtremedistiltransformers: Task transfer for task-agnostic distillation, arXiv 2021, π
- Explanations from Large Language Models Make Small Reasoners Better, arXiv 2022, π
- BERT Learns to Teach: Knowledge Distillation with Meta Learning, ACL 2022, π
- Distilling Step-by-Step! Outperforming Larger Language Models with Less Training Data and Smaller Model Sizes, arXiv 2023, π
- Large language models are reasoning teachers, ACL 2023, π
- Less is More: Task-aware Layer-wise Distillation for Language Model Compression, ICML 2023, π
- MCC-KD: Multi-CoT Consistent Knowledge Distillation, EMNLP 2023, π
- Teaching Small Language Models to Reason, ACL 2023, π
- Symbolic Chain-of-Thought Distillation: Small Models Can Also βThinkβ Step-by-Step, ACL 2023, π
- Gkd: Generalized knowledge distillation for auto-regressive sequence models, arXiv 2023, π
- SCOTT: Self-consistent chain-of-thought distillation, NeurIPS 2023, π
- Dialogue Chain-of-Thought Distillation for Commonsense-aware Conversational Agents, EMNLP 2023, π
- MiniLLM: Knowledge Distillation of Large Language Models, ICLR 2024, π
- On-Policy Distillation of Language Models: Learning from Self-Generated Mistakes, ICLR 2024, π
- PaD: Program-aided Distillation Specializes Large Models in Reasoning, ACL 2024, π
- Turning dust into gold: Distilling complex reasoning capabilities from llms by leveraging negative data, AAAI 2024, π
- PC-LoRA: Low-Rank Adaptation for Progressive Model Compression with Knowledge Distillation, arXiv 2024, π
- Reverse Thinking Makes LLMs Stronger Reasoners, arXiv 2024, π
- UniCoTT: A Unified Framework for Structural Chain-of-Thought Distillation, ICLR 2025, π
- MiniPLM: Knowledge Distillation for Pre-Training Language Models, ICLR 2025, π
- CLIP : Learning Transferable Visual Models From Natural Language Supervision, PMLR 2021, π
- GPT
- DALL-E: Zero-Shot Text-to-Image Generation, ICML 2021, π
- SAM: Segment Anything, ICCV 2023, π
- ViT: An Image is Worth 16x16 Words: Transformers for Image Recognition at Scale, arXiv 2020, π
- DeiT: Training data-efficient image transformers & distillation through attention, ICML 2021, π
- DINO: Emerging Properties in Self-Supervised Vision Transformers, ICCV 2021, π
- DETR: End-to-End Object Detection with Transformers, ECCV 2020, π
- Swin Transformer: Hierarchical Vision Transformer using Shifted Windows, ICCV 2021, π
-
Simple Open-Vocabulary Object Detection with Vision Transformers, ECCV 2022, π
-
Learning to Detect and Segment for Open Vocabulary Object Detection, CVPR 2023, π
-
PracticalDG: Perturbation Distillation on Vision-Language Models for Hybrid Domain Generalization, CVPR 2024, π
-
CLIP
-
Open-Vocabulary Object Detection
- Open-vocabulary Object Detection via Vision and Language Knowledge Distillation, ICLR 2021, π
- Detecting Twenty-thousand Classes using Image-level Supervision, ECCV 2022, π
- Learning to Prompt for Open-Vocabulary Object Detection with Vision-Language Model, CVPR 2022, π
- Open-Vocabulary DETR with Conditional Matching, ECCV 2022, π
- F-VLM: Open-Vocabulary Object Detection upon Frozen Vision and Language Models, arXiv 2022, π
- Bridging the Gap between Object and Image-level Representations for Open-Vocabulary Detection, NeurIPS 2022, π
- PromptDet: Towards Open-vocabulary Detection using Uncurated Images, ECCV 2022, π
- DetCLIP: Dictionary-Enriched Visual-Concept Paralleled Pre-training for Open-world Detection, NeurIPS 2022, π
- Exploiting Unlabeled Data with Vision and Language Models for Object Detection, ECCV 2022, π
- Learning Object-Language Alignments for Open-Vocabulary Object Detection, arXiv 2022, π
- Open-Vocabulary One-Stage Detection with Hierarchical Visual-Language Knowledge Distillation, CVPR 2022, π
- Open-Vocabulary Instance Segmentation via Robust Cross-Modal Pseudo-Labeling, CVPR 2022, π
- Open Vocabulary Object Detection with Pseudo Bounding-Box Labels, ECCV 2022, π
- Fine-grained Visual-Text Prompt-Driven Self-Training for Open-Vocabulary Object Detection, TNNLS 2022, π
- Zero-shot Object Detection Through Vision-Language Embedding Alignment, ICDM 2022, π
- Aligning Bag of Regions for Open-Vocabulary Object Detection, CVPR 2023, π
- Region-Aware Pretraining for Open-Vocabulary Object Detection with Vision Transformers, CVPR 2023, π
- Object-Aware Distillation Pyramid for Open-Vocabulary Object Detection, CVPR 2023, π
- Open-vocabulary Object Detection via Vision and Language Knowledge Distillation, ICLR 2021, π
-
Open-Vocabulary Semantic Segmentation
- Language-driven Semantic Segmentation, ICLR 2021, π
- Semantic Segmentation In-the-Wild Without Seeing Any Segmentation Examples, arXiv 2021, π
- Extract Free Dense Labels from CLIP, ECCV 2022, π
- Image Segmentation Using Text and Image Prompts, CVPR 2022, π
- Scaling Open-Vocabulary Image Segmentation with Image-Level Labels, ECCV 2022, π
- Decoupling Zero-Shot Semantic Segmentation, CVPR 2022, π
- A Simple Baseline for Open-Vocabulary Semantic Segmentation with Pre-trained Vision-language Model, ECCV 2022, π
- ReCo: Retrieve and Co-segment for Zero-shot Transfer, NeurIPS 2022, π
- Open-vocabulary Semantic Segmentation with Frozen Vision-Language Models, arXiv 2022, π
- Open-Vocabulary Semantic Segmentation with Mask-adapted CLIP, CVPR 2023, π
- ZegCLIP: Towards Adapting CLIP for Zero-shot Semantic Segmentation, CVPR 2023, π
- FreeSeg: Unified, Universal and Open-Vocabulary Image Segmentation, CVPR 2023, π
- Exploring Open-Vocabulary Semantic Segmentation from CLIP Vision Encoder Distillation Only, ICCV 2023, π
- CLIP-DIY: CLIP Dense Inference Yields Open-Vocabulary Semantic Segmentation For-Free, WACV 2024, π
- Language-driven Semantic Segmentation, ICLR 2021, π
-
Open-Vocabulary Part Segmentation
- Understanding Multi-Granularity for Open-Vocabulary Part Segmentation, NeurIPS 2025, π
- Understanding Multi-Granularity for Open-Vocabulary Part Segmentation, NeurIPS 2025, π
-
Open-Vocabulary Customization
- Open-Vocabulary Customization from CLIP via Data-Free Knowledge Distillation, ICLR 2025, π
-
Weakly Supervised Semantic Segmentation
- Cross Language Image Matching for Weakly Supervised Semantic Segmentation, CVPR 2022, π
- CLIP is Also an Efficient Segmenter: A Text-Driven Approach for Weakly Supervised Semantic Segmentation, CVPR 2023, π
- Learning Multi-Modal Class-Specific Tokens for Weakly Supervised Dense Object Localization, CVPR 2023, π
- Question-Answer Cross Language Image Matching for Weakly Supervised Semantic Segmentation, ACM MM 2023, π
- Prompting classes: Exploring the Power of Prompt Class Learning in Weakly Supervised Semantic Segmentation, WACV 2024, π
- Frozen CLIP: A Strong Backbone for Weakly Supervised Semantic Segmentation, CVPR 2024, π
- DIAL: Dense Image-text ALignment for Weakly Supervised Semantic Segmentation, ECCV 2024, π
- WeakCLIP: Adapting CLIP for Weakly-Supervised Semantic Segmentation, IJCV 2025, π
- Cross Language Image Matching for Weakly Supervised Semantic Segmentation, CVPR 2022, π
-
Prompt Learning
- Learning to Prompt for Vision-Language Models, IJCV 2022, π
- Conditional Prompt Learning for Vision-Language Models, CVPR 2022, π
- DenseCLIP: Language-Guided Dense Prediction with Context-Aware Prompting, CVPR 2022, π
- Domain Adaptation via Prompt Learning, T-NNLS 2023, π
- PromptKD: Unsupervised Prompt Distillation for Vision-Language Models, CVPR 2024, π
- Learning to Prompt for Vision-Language Models, IJCV 2022, π
-
Generation and Editing
- StyleCLIP: Text-Driven Manipulation of StyleGAN Imagery, ICCV 2021, π
- StyleGAN-NADA: CLIP-Guided Domain Adaptation of Image Generators, ACM SIGGRAPH 2022, π
- Zero-Shot Text-Guided Object Generation with Dream Fields, CVPR 2022, π
- CLIP-NeRF: Text-and-Image Driven Manipulation of Neural Radiance Fields, CVPR 2022, π
- Text2Mesh: Text-Driven Neural Stylization for Meshes, CVPR 2022, π
- Decomposing NeRF for Editing via Feature Field Distillation, NeurIPS 2022, π
- AvatarCLIP: Zero-Shot Text-Driven Generation and Animation of 3D Avatars, ACM Trans. Graph 2022, π
- CLIP-Forge: Towards Zero-Shot Text-to-Shape Generation, CVPR 2022, π
- CLIP2StyleGAN: Unsupervised Extraction of StyleGAN Edit Directions, ACM SIGGRAPH 2022, π
- Text and Image Guided 3D Avatar Generation and Manipulation, WACV 2023, π
- Local 3D Editing via 3D Distillation of CLIP Knowledge, CVPR 2023, π
- StyleCLIP: Text-Driven Manipulation of StyleGAN Imagery, ICCV 2021, π
-
Person Re-Identification
- Distilling CLIP with Dual Guidance for Learning Discriminative Human Body Shape Representation, CVPR 2024, π
-
Zero-Shot Human-Object Interaction (HOI) Detection
- Exploiting CLIP for Zero-shot HOI Detection Requires Knowledge Distillation at Multiple Levels, WACV 2024, π
- Exploiting CLIP for Zero-shot HOI Detection Requires Knowledge Distillation at Multiple Levels, WACV 2024, π
-
Open-Vocabulary Out-of-Distribution Classification
- Distilling Large Vision-Language Model with Out-of-Distribution Generalizability, ICCV 2023, π
- Distilling Large Vision-Language Model with Out-of-Distribution Generalizability, ICCV 2023, π
-
Domain Generalization
- A Sentence Speaks a Thousand Images: Domain Generalization through Distilling CLIP with Language Guidance, ICCV 2023, π
-
Video-Language Retrieval
- CLIPPING: Distilling CLIP-Based Models with a Student Base for Video-Language Retrieval, CVPR 2023, π
-
Affordance Detection
- What does CLIP know about peeling a banana?, CVPR 2024, π
-
Video Highlight Detection
- Unleash the Potential of CLIP for Video Highlight Detection, CVPR 2024, π
-
Monocular Depth Estimation
- CaBins: CLIP-based Adaptive Bins for Monocular Depth Estimation, CVPR 2024, π
-
Multiple Tasks (Image Classification, Object Detection, Semantic Segmentation, Instance Segmentation, Image-Text Retrieval)
- MaskCLIP: Masked Self-Distillation Advances Contrastive Language-Image Pretraining, CVPR 2023, π
- TinyCLIP: CLIP Distillation via Affinity Mimicking and Weight Inheritance, ICCV 2023, π
- CLIP-KD: An Empirical Study of CLIP Model Distillation, CVPR 2024, π
- CLIP-Embed-KD: Computationally Efficient Knowledge Distillation Using Embeddings as Teachers, HPEC 2024, π
- MaskCLIP: Masked Self-Distillation Advances Contrastive Language-Image Pretraining, CVPR 2023, π
-
-
Road Network Graph Extraction
- Segment Anything Model for Road Network Graph Extraction, CVPR 2024, π
-
Polyp Segmentation
- PP-SAM: Perturbed Prompts for Robust Adaptation of Segment Anything Model for Polyp Segmentation, CVPR 2024, π
- PP-SAM: Perturbed Prompts for Robust Adaptation of Segment Anything Model for Polyp Segmentation, CVPR 2024, π
-
Image Restoration
- Distilling Semantic Priors from SAM to Efficient Image Restoration Models, CVPR 2024, π
-
Weakly Supervised Semantic Segmentation
- Segment Anything Model (SAM) Enhanced Pseudo Labels for Weakly Supervised Semantic Segmentation, arXiv 2023, π
- An Alternative to WSSS? An Empirical Study of the Segment Anything Model (SAM) on Weakly-Supervised Semantic Segmentation Problems, arXiv 2023, π
- Segment Anything is A Good Pseudo-label Generator for Weakly Supervised Semantic Segmentation, arXiv 2023, π
- Weakly-Supervised Semantic Segmentation with Image-Level Labels: from Traditional Models to Foundation Models, arXiv 2023, π
- From SAM to CAMs: Exploring Segment Anything Model for Weakly Supervised Semantic Segmentation, CVPR 2024, π
- Segment Anything Model (SAM) Enhanced Pseudo Labels for Weakly Supervised Semantic Segmentation, arXiv 2023, π
-
SAM + CLIP
- Foundation Model Assisted Weakly Supervised Semantic Segmentation, WACV 2023, π
- SAM-CLIP: Merging Vision Foundation Models towards Semantic and Spatial Understanding, CVPR 2024, π
- Test-Time Adaptation with SaLIP: A Cascade of SAM and CLIP for Zero shot Medical Image Segmentation, CVPR 2024, π
- Foundation Model Assisted Weakly Supervised Semantic Segmentation, WACV 2023, π
-
DINO + CLIP
-
DETR + CLIP
- T-Rex2: Towards Generic Object Detection via Text-Visual Prompt Synergy, ECCV 2024, π
- T-Rex2: Towards Generic Object Detection via Text-Visual Prompt Synergy, ECCV 2024, π
-
SAM + GDINO
-
Others
- FeatureNeRF: Learning Generalizable NeRFs by Distilling Foundation Models, ICCV 2023, π
- DIME-FM: DIstilling Multimodal and Efficient Foundation Models, ICCV 2023, π
- AM-RADIO: Agglomerative Vision Foundation Model -- Reduce All Domains Into One, CVPR 2024, π
- Theia: Distilling Diverse Vision Foundation Models for Robot Learning, arXiv 2024, π
- Strategies to Leverage Foundational Model Knowledge in Object Affordance Grounding, CVPR 2024, π
- Enrich Distill and Fuse: Generalized Few-Shot Semantic Segmentation in Remote Sensing Leveraging Foundation Model's Assistance, CVPR 2024, π
- Exploring the Benefits of Vision Foundation Models for Unsupervised Domain Adaptation, CVPR 2024, π
- Distilling Knowledge from Multiple Foundation Models for Zero-shot Image classification, PlosOne 2024, π
- FeatureNeRF: Learning Generalizable NeRFs by Distilling Foundation Models, ICCV 2023, π
-
Vision Transformer (ViT)
- Supervised Masked Knowledge Distillation for Few-Shot Transformers, CVPR 2023, π
- Masked Autoencoders Enable Efficient Knowledge Distillers, CVPR 2023, π
- Distilling Self-Supervised Vision Transformers for Weakly-Supervised Few-Shot Classification & Segmentation, CVPR 2023, π
- TinyMIM: An Empirical Study of Distilling MIM Pre-trained Models, CVPR 2023, π
- Incrementer: Transformer for Class-Incremental Semantic Segmentation With Knowledge Distillation Focusing on Old Class, CVPR 2023, π
- Cumulative Spatial Knowledge Distillation for Vision Transformers, ICCV 2023, π
- Feature-based Knowledge Distillation for Vision Transformers, CVPR 2024, π
- Supervised Masked Knowledge Distillation for Few-Shot Transformers, CVPR 2023, π
-
Distillation with No Labels (DINO)
-
Unsupervised Object Discovery
- Localizing Objects with Self-Supervised Transformers and no Labels, BMVC 2021, π
- Self-Supervised Transformers for Unsupervised Object Discovery using Normalized Cut, CVPR 2022, π
- Unsupervised Object Localization: Observing the Background to Discover Objects, CVPR 2023, π
- Cut and Learn for Unsupervised Object Detection and Instance Segmentation, CVPR 2023, π
- Grounding DINO: Marrying DINO with Grounded Pre-Training for Open-Set Object Detection, ECCV 2024, π
- DIOD: Self-Distillation Meets Object Discovery, CVPR 2024, π
- Localizing Objects with Self-Supervised Transformers and no Labels, BMVC 2021, π
-
Open-world Counting
- CountGD: Multi-Modal Open-World Counting, NeurIPS 2025, π
- CountGD: Multi-Modal Open-World Counting, NeurIPS 2025, π
-
Zero-shot Semantic Segmentation
- Diffusion Models for Open-Vocabulary Segmentation, arXiv 2023, π
-
-
DEtection TRansformer (DETR)
-
Swin Transformer
- Per-Pixel Classification is Not All You Need for Semantic Segmentation, NeurIPS 2021, π
- Dynamic Feature Regularized Loss for Weakly Supervised Semantic Segmentation, arXiv 2021, π
- SwinNet: Swin Transformer Drives Edge-Aware RGB-D and RGB-T Salient Object Detection, TCSVT 2021, π
- SwinTrack: A Simple and Strong Baseline for Transformer Tracking, NeurIPS 2022, π
- SwinF: Swin Transformer with feature fusion in target detection, JPCS 2022, π
- Self-Supervised Pre-Training of Swin Transformers for 3D Medical Image Analysis, CVPR 2022, π
- SSformer: A Lightweight Transformer for Semantic Segmentation, MMSP 2022, π
- Weakly Supervised Intracranial Hemorrhage Segmentation using Head-Wise Gradient-Infused Self-Attention Maps from a Swin Transformer in Categorical Learning, arXiv 2023, π
- HiFormer: Hierarchical Multi-scale Representations Using Transformers for Medical Image Segmentation, WACV 2023, π
- Swin-Fusion: Swin-Transformer with Feature Fusion for Human Action Recognition, NPL 2023, π
- Pyramid Swin Transformer: Different-Size Windows Swin Transformer for Image Classification and Object Detection, VISIGRAPP 2023, π
- SCUNet++: Swin-UNet and CNN Bottleneck Hybrid Architecture with Multi-Fusion Dense Skip Connection for Pulmonary Embolism CT Image Segmentation, WACV 2024, π
- Leveraging Swin Transformer for Local-to-Global Weakly Supervised Semantic Segmentation, MVIP 2024, π
- Per-Pixel Classification is Not All You Need for Semantic Segmentation, NeurIPS 2021, π
- A simple framework for contrastive learning of visual representations, NeurIPS 2020, π
- Momentum contrast for unsupervised visual representation learning, CVPR 2020, π
- Improved Baselines with Momentum Contrastive Learning, arXiv 2020, π
- Bootstrap your own latent-a new approach to self-supervised learning, NeurIPS 2020, π
- Unsupervised learning of visual features by contrasting cluster assignments, NeurIPS 2020, π
- Self-Supervised Learning of Pretext-Invariant Representations, CVPR 2020, π
- CompRess: Self-Supervised Learning by Compressing Representations, NeurIPS 2020, π
- SimReg: Regression as a Simple Yet Effective Tool for Self-supervised Knowledge Distillation, BMVC 2021, π
- ISD: Self-Supervised Learning by Iterative Similarity Distillation, ICCV 2021, π
- SEED: Self-supervised Distillation For Visual Representation, ICLR 2021, π
- Simple Distillation Baselines for Improving Small Self-supervised Models, ICCV 2021, π
- Emerging Properties in Self-Supervised Vision Transformers, ICCV 2021, π
- Bag of Instances Aggregation Boosts Self-supervised Distillation, ICLR 2022, π
- DisCo: Remedying Self-supervised Learning on Lightweight Models with Distilled Contrastive Learning, ECCV 2022, π
- Auxiliary Learning for Self-Supervised Video Representation via Similarity-Based Knowledge Distillation, CVPR 2022, π
- Masked Video Distillation: Rethinking Masked Feature Modeling for Self-Supervised Video Representation Learning, CVPR 2023, π
- Multi-Mode Online Knowledge Distillation for Self-Supervised Visual Representation Learning, CVPR 2023, π
- DINOv2: Learning Robust Visual Features without Supervision, TMLR 2024, π
- Knowledge Distillation in Iterative Generative Models for Improved Sampling Speed, arXiv 2021, π
- Progressive Distillation for Fast Sampling of Diffusion Models, ICLR 2022, π
- Accelerating diffusion sampling with classifier-based feature distillation, ICME 2023, π
- Consistency Models, PMLR 2023, π
- Fast Sampling of Diffusion Models via Operator Learning, PMLR 2023, π
- Diffusion-GAN: Training GANs with Diffusion, ICLR 2023, π
- ProlificDreamer: High-Fidelity and Diverse Text-to-3D Generation with Variational Score Distillation, NeurIPS 2023, π
- Diff-Instruct: A Universal Approach for Transferring Knowledge From Pre-trained Diffusion Models, NeurIPS 2023, π
- Latent Consistency Models: Synthesizing High-Resolution Images with Few-Step Inference, arXiv 2023, π
- Adversarial Diffusion Distillation, ECCV 2024, π
- Improved Techniques for Training Consistency Models, ICLR 2024, π
- Relational diffusion distillation for efficient image generation, ACM MM 2024, π
- Multistep Distillation of Diffusion Models via Moment Matching, NeurIPS 2024, π
- EM Distillation for One-step Diffusion Models, NeurIPS 2024, π
- Simple and Fast Distillation of Diffusion Models, NeurIPS 2024, π
- One-step Diffusion with Distribution Matching Distillation, CVPR 2024, π
- Simplifying, Stabilizing and Scaling Continuous-Time Consistency Models, arXiv 2024, π
- Consistency Models Made Easy, arXiv 2024, π
- TLCM: Training-efficient Latent Consistency Model for Image Generation with 2-8 Steps, arXiv 2024, π
- Learning Efficient Object Detection Models with Knowledge Distillation, NIPS 2017, π
- Distilling Object Detectors with Fine-grained Feature Imitation, CVPR 2019, π
- An end-to-end architecture for class-incremental object detection with knowledge distillation, ICME 2019, π
- LabelEnc: A New Intermediate Supervision Method for Object Detection, ECCV 2020, π
- MimicDet: Bridging the Gap Between One-Stage and Two-Stage Object Detection, ECCV 2020, π
- Improve object detection with feature-based knowledge distillation: Towards accurate and efficient detectors, ICLR 2021, π
- Distilling Object Detectors via Decoupled Features, CVPR 2021, π
- General Instance Distillation for Object Detection, CVPR 2021, π
- Instance-Conditional Knowledge Distillation for Object Detection, NIPS 2021, π
- Improving Object Detection by Label Assignment Distillation, WACV 2022, π
- Distilling image classifiers in object detectors, NIPS 2021, π
- G-DetKD: Towards General Distillation Framework for Object Detectors via Contrastive and Semantic-guided Feature Imitation, ICCV 2021, π
- Distilling Object Detectors with Feature Richness, NIPS 2021, π
- Prediction-Guided Distillation for Dense Object Detection, ECCV 2022, π
- Focal and Global Knowledge Distillation for Detectors, CVPR 2022, π
- GLAMD: Global and Local Attention Mask Distillation for Object Detectors, ECCV 2022, π
- Masked Generative Distillation, ECCV 2022, π
- LGD: Label-guided Self-distillation for Object Detection, AAAI 2022, π
- Knowledge Distillation for Object Detection via Rank Mimicking and Prediction-Guided Feature Imitation, AAAI 2022, π
- Task-balanced distillation for object detection, Pattern Recognition 2023, π
- Structured Knowledge Distillation for Accurate and Efficient Object Detection, TPAMI 2023, π
- Spatial Self-Distillation for Object Detection with Inaccurate Bounding Boxes, ICCV 2023, π
- Dual Relation Knowledge Distillation for Object Detection, IJCAI 2023, π
- ScaleKD: Distilling Scale-Aware Knowledge in Small Object Detector, CVPR 2023, π
- Multi-level Logit Distillation, CVPR 2023, π
- Bridging Cross-task Protocol Inconsistency for Distillation in Dense Object Detection, ICCV 2023, π
- Distilling DETR with Visual-Linguistic Knowledge for Open-Vocabulary Object Detection, ICCV 2023, π
- UniKD: Universal Knowledge Distillation for Mimicking Homogeneous or Heterogeneous Object Detectors, ICCV 2023, π
- Texture-Guided Saliency Distilling for Unsupervised Salient Object Detection, CVPR 2023, π
- Object-Aware Distillation Pyramid for Open-Vocabulary Object Detection, CVPR 2023, π
- Localization distillation for object detection, IEEE TPAMI 2023, π
- CrossKD: Cross-Head Knowledge Distillation for Dense Object Detection, CVPR 2024, π
- MSSD: multi-scale self-distillation for object detection, Visual Intelligence 2024, π
- CLIPSelf: Vision Transformer Distills Itself for Open-Vocabulary Dense Prediction, ICLR 2024, π
- Gradient-Guided Knowledge Distillation for Object Detectors, WACV 2024, π
- Efficient Feature Distillation for Zero-Shot Annotation Object Detection, WACV 2024, π
- Image Super-Resolution Using Knowledge Distillation, ACCV 2018, π
- Feature-Affinity Based Knowledge Distillation for Efficient Image Super-Resolution, ICIP 2020, π
- Learning with Privileged Information for Efficient Image Super-Resolution, ECCV 2020, π
- Data-Free Knowledge Distillation for Image Super-Resolution, CVPR 2021, π
- Towards Compact Single Image Super-Resolution via Contrastive Self-distillation, IJCAI 2021, π
- DIPNet: Efficiency Distillation and Iterative Pruning for Image Super-Resolution, CVPRW 2023, π
- Generative Adversarial Super-Resolution at the edge with knowledge distillation, Engineering Applications of Artificial Intelligence 2023, π
- Hybrid knowledge distillation from intermediate layers for efficient Single Image Super-Resolution, Neurocomputing 2023, π
- Dual cross knowledge distillation for image super-resolution, Journal of Visual Communication and Image Representation 2023, π
- Pairwise Distance Distillation for Unsupervised Real-World Image Super-Resolution, ECCV 2024, π
- Knowledge Distillation for Single Image Super-Resolution via Contrastive Learning, ICMR 2024, π
- MTKD: Multi-Teacher Knowledge Distillation for Image Super-Resolution, ECCV 2024, π
- You Only Need One Step: Fast Super-Resolution with Stable Diffusion via Scale Distillation, ECCV 2024, π
- Attention Guidance Distillation Network for Efficient Image Super-Resolution, CVPRW 2024, π
- SinSR: Diffusion-Based Image Super-Resolution in a Single Step, CVPR 2024, π
- OSFFNet: Omni-Stage Feature Fusion Network for Lightweight Image Super-Resolution, AAAI 2024, π
- Semantic Super-Resolution via Self-Distillation and Adversarial Learning, IEEE Access 2024, π
- Structured knowledge distillation for semantic segmentation, CVPR 2019, π
- Knowledge Adaptation for Efficient Semantic Segmentation, CVPR 2019, π
- Intra-class Feature Variation Distillation for Semantic Segmentation, ECCV 2020, π
- Domain Adaptive Knowledge Distillation for Driving Scene Semantic Segmentation, WACVW 2020, π
- Channel-wise Knowledge Distillation for Dense Prediction, ICCV 2021, π
- Double Similarity Distillation for Semantic Image Segmentation, IEEE Transactions on Image Processing (TIP) 2021, π
- Robust Semantic Segmentation With Multi-Teacher Knowledge Distillation, IEEE Access 2021, π
- Cross-Image Relational Knowledge Distillation for Semantic Segmentation, CVPR 2022, π
- Structural and Statistical Texture Knowledge Distillation for Semantic Segmentation, CVPR 2022, π
- Adaptive Perspective Distillation for Semantic Segmentation, TPAMI 2022, π
- Masked Generative Distillation, ECCV 2022, π
- Class Similarity Weighted Knowledge Distillation for Continual Semantic Segmentation, CVPR 2022, π
- Cross-Domain Correlation Distillation for Unsupervised Domain Adaptation in Nighttime Semantic Segmentation, CVPR 2022, π
- DiGA: Distil to Generalize and then Adapt for Domain Adaptive Semantic Segmentation, CVPR 2023, π
- A Good Student is Cooperative and Reliable: CNN-Transformer Collaborative Learning for Semantic Segmentation, ICCV 2023, π
- Endpoints Weight Fusion for Class Incremental Semantic Segmentation, CVPR 2023, π
- Multi-Task Learning with Knowledge Distillation for Dense Prediction, ICCV 2023, π
- Hierarchical Dense Correlation Distillation for Few-Shot Segmentation, CVPR 2023, π
- Distilling Self-Supervised Vision Transformers for Weakly-Supervised Few-Shot Classification & Segmentation, CVPR 2023, π
- Incrementer: Transformer for Class-Incremental Semantic Segmentation With Knowledge Distillation Focusing on Old Class, CVPR 2023, π
- FAKD: Feature Augmented Knowledge Distillation for Semantic Segmentation, WACV 2024, π
- Knowledge Distillation for Efficient Instance Semantic Segmentation with Transformers, CVPRW 2024, π
- CLIPSelf: Vision Transformer Distills Itself for Open-Vocabulary Dense Prediction, ICLR 2024, π
- P2Seg: Pointly-supervised Segmentation via Mutual Distillation, ICLR 2024, π
- Rethinking Knowledge Distillation With Raw Features for Semantic Segmentation, WACV 2024, π
- BPKD: Boundary Privileged Knowledge Distillation for Semantic Segmentation, WACV 2024, π
- Guided Distillation for Semi-Supervised Instance Segmentation, WACV 2024, π
- Distilling efficient Vision Transformers from CNNs for semantic segmentation, Pattern Recognition 2025, π
- AICSD: Adaptive Inter-Class Similarity Distillation for Semantic Segmentation, Multimedia Tools and Applications 2025, π
- Knowledge Distillation for Fast and Accurate Monocular Depth Estimation on Mobile Devices, CVPRW 2021, π
- Realtime single image depth perception in the wild with handheld devices, Sensors 2021, π
- Excavating the Potential Capacity of Self-Supervised Monocular Depth Estimation, ICCV 2021, π
- Self-distilled Feature Aggregation for Self-supervised Monocular Depth Estimation, ECCV 2022, π
- Attention-based depth distillation with 3d-aware positional encoding for monocular 3d object detection, AAAI 2022, π
- Boosting Light-Weight Depth Estimation Via Knowledge Distillation, KSEM 2023, π
- Urcdc-depth: Uncertainty rectified cross-distillation with cutflip for monocular depth estimation, IEEE Transactions on Multimedia 2023, π
- Multi-task learning with knowledge distillation for dense prediction, CVPR 2023, π
- Attention-Based Knowledge Distillation in Scene Recognition: The Impact of a DCT-Driven Loss, IEEE Transactions on Circuits and Systems for Video Technology 2023, π
- Monocular Depth Estimation from a Fisheye Camera Based on Knowledge Distillation, Sensors 2023, π
- 3D Distillation: Improving Self-Supervised Monocular Depth Estimation on Reflective Surfaces, ICCV 2023, π
- Multi-Task Learning with Knowledge Distillation for Dense Prediction, ICCV 2023, π
- GasMono: Geometry-Aided Self-Supervised Monocular Depth Estimation for Indoor Scenes, ICCV 2023, π
- Depth Anywhere: Enhancing 360 Monocular Depth Estimation via Perspective Distillation and Unlabeled Data Augmentation, NIPS 2024, π
- Structure-Centric Robust Monocular Depth Estimation via Knowledge Distillation, ACCV 2024, π
- MonoProb: Self-Supervised Monocular Depth Estimation With Interpretable Uncertainty, WACV 2024, π
- Monocular Depth Estimation via Self-Supervised Self-Distillation, Sensors 2024, π
- MAL: Motion-Aware Loss with Temporal and Distillation Hints for Self-Supervised Depth Estimation, ICRA 2024, π
- Stereo-Matching Knowledge Distilled Monocular Depth Estimation Filtered by Multiple Disparity Consistency, ICASSP 2024, π
- Categorical Relation-Preserving Contrastive Knowledge Distillation for Medical Image Classification, MICCAI 2021, π
- Efficient Medical Image Segmentation Based on Knowledge Distillation, IEEE Transactions on Medical Imaging (TMI) 2021, π
- DeSD: Self-Supervised Learning with Deep Self-Distillation for 3D Medical Image Segmentation, MICCAI 2022, π
- Efficient Biomedical Instance Segmentation via Knowledge Distillation, MICCAI 2022, π
- Flat-aware Cross-stage Distilled Framework for Imbalanced Medical Image Classification, MICCAI 2022, π
- Free Lunch for Surgical Video Understanding by Distilling Self-Supervisions, MICCAI 2022, π
- Simcvd: Simple contrastive voxel-wise representation distillation for semi-supervised medical image segmentation, IEEE Trans. Med. Imaging 2022, π
- Deep Mutual Distillation for Semi-Supervised Medical Image Segmentation, MICCAI 2023, π
- Distilling BlackBox to Interpretable models for Efficient Transfer Learning, MICCAI 2023, π
- Learnable Cross-modal Knowledge Distillation for Multi-modal Learning with Missing Modality, MICCAI 2023, π
- Self-distillation for surgical action recognition, MICCAI 2023, π
- Semi-supervised Pathological Image Segmentation via Cross Distillation of Multiple Attentions, MICCAI 2023, π
- Bootstrapping Chest CT Image Understanding by Distilling Knowledge from X-ray Expert Models, CVPR 2024, π
- Reverse Knowledge Distillation: Training a Large Model Using a Small One for Retinal Image Matching on Limited Data, WACV 2024, π
- Learning Robust Shape Regularization for Generalizable Medical Image Segmentation, IEEE Trans. Med. Imaging 2024, π
- Enhancing Medical Imaging with GANs Synthesizing Realistic Images from Limited Data, ICETCI 2024, π
- Exploring Generalizable Distillation for Efficient Medical Image Segmentation, IEEE Journal of Biomedical and Health Informatics 2024, π
- Confidence Matters: Enhancing Medical Image Classification Through Uncertainty-Driven Contrastive Self-Distillation, MICCAI 2024, π
- DES-SAM: Distillation-Enhanced Semantic SAM for Cervical Nuclear Segmentation with Box Annotation, MICCAI 2024, π
- Hallucinated Style Distillation for Single Domain Generalization in Medical Image Segmentation, MICCAI 2024, π
- Progressively Correcting Soft Labels via Teacher Team for Knowledge Distillation in Medical Image Segmentation, MICCAI 2024, π
- Reprogramming Distillation for Medical Foundation Models, MICCAI 2024, π
- HDKD: Hybrid Data-Efficient Knowledge Distillation Network for Medical Image Classification, Engineering Applications of Artificial Intelligence 2024, π
- Shape-Intensity Knowledge Distillation For Robust Medical Image Segmentation, Frontiers of Computer Science 2024, π
- There is More than Meets the Eye: Self-Supervised Multi-Object Detection and Tracking with Sound by Distilling Multimodal Knowledge, CVPR 2021, π
- Ensemble learning with siamese networks for visual tracking, Neurocomputing 2021, π
- Distilled Siamese Networks for Visual Tracking, TPAMI 2022, π
- SDSTrack: Self-Distillation Symmetric Adapter Learning for Multi-Modal Visual Object Tracking, CVPR 2024, π
- Event Stream-based Visual Object Tracking: A High-Resolution Benchmark Dataset and A Novel Baseline, CVPR 2024, π
- Object Knowledge Distillation for Joint Detection and Tracking in Satellite Videos, IEEE Transactions on Geoscience and Remote Sensing 2024, π
- Teacher Supervises Students How to Learn From Partially Labeled Images for Facial Landmark Detection, ICCV 2019, π
- Fair Feature Distillation for Visual Recognition, CVPR 2021, π
- Rectifying the Data Bias in Knowledge Distillation, ICCV 2021, π
- Teaching Where to Look: Attention Similarity Knowledge Distillation for Low Resolution Face Recognition, ECCV 2022, π
- Evaluation-oriented knowledge distillation for deep face recognition, CVPR 2022, π
- Rethinking Feature-Based Knowledge Distillation for Face Recognition, CVPR 2023, π
- Probabilistic Knowledge Distillation of Face Ensembles, CVPR 2023, π
- SynthDistill: Face Recognition with Knowledge Distillation from Synthetic Data, IJCB 2023, π
- AdaDistill: Adaptive Knowledge Distillation for Deep Face Recognition, ECCV 2024, π
- How Knowledge Distillation Mitigates the Synthetic Gap in Fair Face Recognition, ECCV Workshop 2024, π
- MST-KD: Multiple Specialized Teachers Knowledge Distillation for Fair Face Recognition, ECCV Workshop 2024, π
- Enhanced Face Recognition using Intra-class Incoherence Constraint, ICLR 2024, π
- ProS: Facial Omni-Representation Learning via Prototype-Based Self-Distillation, WACV 2024, π
- AI-KD: Towards Alignment Invariant Face Image Quality Assessment Using Knowledge Distillation, IWBF 2024, π
- Structural Knowledge Distillation for Efficient Skeleton-Based Action Recognition, IEEE Trans. Image Processing 2021, π
- Video Pose Distillation for Few-Shot, Fine-Grained Sports Action Recognition, ICCV 2021, π
- Learning an Augmented RGB Representation with Cross-modal Knowledge Distillation for Action Detection, ICCV 2021, π
- Privileged Knowledge Distillation for Online Action Detection, Pattern recognition 2022, π
- Multimodal Distillation for Egocentric Action Recognition, ICCV 2023, π
- FSAR: Federated Skeleton-based Action Recognition with Adaptive Topology Structure and Knowledge Distillation, ICCV 2023, π
- Decomposed Cross-Modal Distillation for RGB-Based Temporal Action Detection, CVPR 2023, π
- Generative Model-Based Feature Knowledge Distillation for Action Recognition, AAAI 2024, π
- FROSTER: Frozen CLIP is A Strong Teacher for Open-Vocabulary Action Recognition, ICLR 2024, π
- Dynamic Kernel Distillation for Efficient Pose Estimation in Videos, ICCV 2019, π
- Fast Human Pose Estimation, CVPR 2019, π
- Online Knowledge Distillation for Efficient Pose Estimation, ICCV 2021, π
- When Human Pose Estimation Meets Robustness: Adversarial Algorithms and Benchmarks, CVPR 2021, π
- From Synthetic to Real: Unsupervised Domain Adaptation for Animal Pose Estimation, CVPR 2021, π
- Combining Weight Pruning and Knowledge Distillation for CNN Compression, CVPR Workshop 2021, π
- AlphaPose: Whole-Body Regional Multi-Person Pose Estimation and Tracking in Real-Time, IEEE TPAMI 2022, π
- ViTPose: Simple Vision Transformer Baselines for Human Pose Estimation, NeurIPS 2022, π
- DistilPose: Tokenized Pose Regression With Heatmap Distillation, CVPR 2023, π
- Knowledge Distillation for 6D Pose Estimation by Aligning Distributions of Local Predictions, CVPR 2023, π
- Effective Whole-Body Pose Estimation with Two-Stages Distillation, ICCV 2023, π
- HRPose: Real-Time High-Resolution 6D Pose Estimation Network Using Knowledge Distillation, Chinese Journal of Electronics 2023, π
- SDPose: Tokenized Pose Estimation via Circulation-Guide Self-Distillation, CVPR 2024, π
- Uncertainty-aware multi-shot knowledge distillation for image-based object re-identification, AAAI 2020, π
- Robust re-identification by multiple views knowledge distillation, ECCV 2020, π
- Relationship-Preserving Knowledge Distillation for Zero-Shot Sketch Based Image Retrieval, ACMM 2021, π
- Asymmetric metric learning for knowledge transfer, CVPR 2021, π
- Contextual Similarity Distillation for Asymmetric Image Retrieval, CVPR 2022, π
- Large-to-small image resolution asymmetry in deep metric learning, WACV 2023, π
- Towards a Smaller Student: Capacity Dynamic Distillation for Efficient Image Retrieval, CVPR 2023, π
- D3still: Decoupled Differential Distillation for Asymmetric Image Retrieval, CVPR 2024, π
- LSTKC: Long Short-Term Knowledge Consolidation for Lifelong Person Re-identification, AAAI 2024, π
- Pairwise difference relational distillation for object re-identification, Pattern Recognition 2024, π