-
Naver AI Lab
- Seoul, Korea
- https://sangdooyun.github.io
Stars
Official Implementation of MambaMia (AAAI-26 Oral)
PyTorch implementation of JiT https://arxiv.org/abs/2511.13720
coallaoh / MASEval
Forked from parameterlab/MASEvalMulti-Agent LLM Evaluation
Source code of "Dr.LLM: Dynamic Layer Routing in LLMs"
reproduction of semantic segmentation using masked autoencoder (mae)
utilities for decoding deep representations (like sentence embeddings) back to text
gpt-oss-120b and gpt-oss-20b are two open-weight language models by OpenAI
codes for R-Zero: Self-Evolving Reasoning LLM from Zero Data (https://www.arxiv.org/pdf/2508.05004)
https://huggingface.co/datasets/multimodal-reasoning-lab/Zebra-CoT
[NeurIPS 2025] Official PyTorch implementation of "Token Bottleneck: One Token to Remember Dynamics"
🤗 LeRobot: Making AI for Robotics more accessible with end-to-end learning
A curated list for Efficient Large Language Models
Fully open data curation for reasoning models
Unified Efficient Fine-Tuning of 100+ LLMs & VLMs (ACL 2024)
Source code of "C-SEO Bench: Does Conversational SEO Work?" NeurIPS D&B 2025
Source code of "Leaky Thoughts: Large Reasoning Models Are Not Private Thinkers" EMNLP 2025
Just another reasonably minimal repo for class-conditional training of pixel-space diffusion transformers.
Open-source framework for the research and development of foundation models.
Mixture-of-Transformers: A Sparse and Scalable Architecture for Multi-Modal Foundation Models. TMLR 2025.

