Skip to content

jeremyeder/mtop

Repository files navigation

mtop

A demonstration tool showcasing llm-d's enterprise-ready LLM deployment and management capabilities.

llm-d-demo

See llm-d in action: GPU-aware scheduling, SLO-driven autoscaling, and cost optimization for LLM workloads.

What is llm-d?

llm-d is Red Hat's open source platform for deploying and managing Large Language Models at scale. It provides:

  • GPU-Aware Orchestration: Intelligent scheduling across heterogeneous GPU clusters (H100, A100, DRA)
  • SLO-Driven Operations: Automatic scaling based on Time-to-First-Token (TTFT) and latency targets
  • Cost Intelligence: 40% reduction in GPU costs through dynamic right-sizing and bin-packing
  • Enterprise Integration: Native OpenShift/Kubernetes integration with RBAC and multi-tenancy

About This Demo

mtop simulates a production llm-d environment without requiring a live cluster, allowing you to:

  • Demonstrate llm-d's real-time monitoring capabilities
  • Showcase SLO convergence and autoscaling behavior
  • Illustrate GPU utilization optimization
  • Present cost savings through intelligent scheduling

Quick Demo

# Show llm-d's real-time monitoring
./mtop-main

# Demonstrate SLO-driven autoscaling
./mtop-main slo-dashboard

# Showcase deployment strategies
./mtop-main simulate canary

Key llm-d Capabilities Demonstrated

🎯 SLO-Driven Autoscaling

slo-compliance

llm-d automatically scales deployments to meet TTFT and latency targets while minimizing cost.

💰 Intelligent Cost Optimization

cost-optimization

GPU-aware bin-packing and right-sizing reduces infrastructure costs by up to 40%.

🔥 Heterogeneous GPU Management

gpu-efficiency

Seamlessly manage H100, A100, and DRA clusters with unified scheduling and monitoring.

📊 Enterprise-Ready Operations

multi-model

Production-grade monitoring, RBAC, and multi-tenant isolation for enterprise deployments.

Demo Scenarios

Pre-configured scenarios showcasing llm-d capabilities:

./scripts/demo.py startup           # Small team adopting llm-d
./scripts/demo.py enterprise        # Large-scale llm-d deployment
./scripts/demo.py cost-optimization # Cost reduction showcase

Installation

# Requirements: Python 3.12+
pip install -r requirements.txt
chmod +x mtop-main

# Run the demo
./mtop-main

Learn More About llm-d

Demo Documentation

Contributing

This demo tool is part of the llm-d project. For contributions:


Part of the llm-d project by Red Hat AI Engineering.

About

Mock CLI tool for debugging and simulating LLMInferenceService CRDs (ChatGPT generated)

Topics

Resources

License

Security policy

Stars

Watchers

Forks

Packages

No packages published

Contributors 4

  •  
  •  
  •  
  •