Skip to content
#

memory-optimization

Here are 20 public repositories matching this topic...

Run larger LLMs with longer contexts on Apple Silicon by using differentiated precision for KV cache quantization. KVSplit enables 8-bit keys & 4-bit values, reducing memory by 59% with <1% quality loss. Includes benchmarking, visualization, and one-command setup. Optimized for M1/M2/M3 Macs with Metal support.

  • Updated May 21, 2025
  • Python

Automated memory optimization for Pandas DataFrames. Reduces memory by 50-80% and loads CSV 5-10x faster. Drop-in replacement that returns standard DataFrames - works with sklearn, matplotlib, and all your favorite libraries. Safe, fast, and zero refactoring required.

  • Updated Dec 24, 2025
  • Python

A chatbot with RBAC and Human-in-the-Loop used for booking purposes covering the research, plan formation, booking, payment. This is built for multiple users like - Booking customers, internal teams (finance, customer support etc..). Built with multi-agent two-layer supervised architecture in LangGraph powered by LLMs including openAI,Gemini, Ollam

  • Updated Jan 26, 2026
  • Python

GPU Memory Calculator for LLM Training - Calculate GPU memory requirements for training Large Language Models with support for multiple training engines including PyTorch DDP, DeepSpeed ZeRO, Megatron-LM, and FSDP.

  • Updated Jan 26, 2026
  • Python

Improve this page

Add a description, image, and links to the memory-optimization topic page so that developers can more easily learn about it.

Curate this topic

Add this topic to your repo

To associate your repository with the memory-optimization topic, visit your repo's landing page and select "manage topics."

Learn more