This repository was archived by the owner on Oct 15, 2025. It is now read-only.
  
  
  
  
llm-d-1.0.14
llm-d is a Kubernetes-native high-performance distributed LLM inference framework
What's Changed
- Fix post minikube merge prometheus install for minikube by @nerdalert in #228
 - Add 
--minikubeflag to--uninstallby @nerdalert in #224 - Add "slim" value files for each example that will run the stack on 1xL4 g6.2xlarge by @nerdalert in #230
 - Add command to minikube troubleshooting section by @nerdalert in #237
 - Add example of llama4 fp8 model by @cfchase in #217
 - README updates for values files and minor fixes by @cfchase in #244
 - Set instructions to minikube since that's tested, and vanilla kube by @cfchase in #245
 - Add doc for creating a cluster for llm-d by @cfchase in #247
 - Add more machine type clarity. by @nerdalert in #246
 - Add blurb saying only single installation is supported by @cfchase in #249
 - chore(ci): change to using public images and no extra creds in GHA by @tumido in #254
 - chore(ci): fix the badge img url for release decorator by @tumido in #255
 - chore: claim artifacthub ownership by @tumido in #258
 
Full Changelog: llm-d-1.0.13...llm-d-1.0.14