Skip to content
View zxdmike's full-sized avatar

Block or report zxdmike

Block user

Prevent this user from interacting with your repositories and sending you notifications. Learn more about blocking users.

You must be logged in to block users.

Please don't include any personal information such as legal names or email addresses. Maximum 100 characters, markdown supported. This note will be visible to only you.
Report abuse

Contact GitHub support about this user’s behavior. Learn more about reporting abuse.

Report abuse
Showing results

Fine-tuning & Reinforcement Learning for LLMs. 🦥 Train OpenAI gpt-oss, Qwen3, Llama 4, DeepSeek-R1, Gemma 3, TTS 2x faster with 70% less VRAM.

Python 45,485 3,700 Updated Sep 15, 2025

This repository contains the training code of ParetoQ introduced in our work "ParetoQ Scaling Laws in Extremely Low-bit LLM Quantization"

Python 100 6 Updated Jun 2, 2025

⚖️ The First Coding Agent-as-a-Judge

Python 628 91 Updated May 14, 2025

Code repo for the paper "SpinQuant LLM quantization with learned rotations"

Python 326 53 Updated Feb 14, 2025

MobileLLM Optimizing Sub-billion Parameter Language Models for On-Device Use Cases. In ICML 2024.

Python 1,342 79 Updated Apr 21, 2025

Code repo for the paper "LLM-QAT Data-Free Quantization Aware Training for Large Language Models"

Python 312 24 Updated Mar 4, 2025