Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Loading base model/lora adapters from s3 fails and throws HF error #766

Open
robert-moyai opened this issue Feb 28, 2025 · 0 comments
Open
Labels
area/lora kind/bug Something isn't working

Comments

@robert-moyai
Copy link

🐛 Describe the bug

Im trying to initialize AIBrix using model files hosted on s3. But I am not able to get vllm-openai container to run as its incorrectly identifying the s3 link as being a HF repo.

huggingface_hub.errors.HFValidationError: Repo id must be in the form 'repo_name' or 'namespace/repo_name': 's3://s3-bucket/llama-3.1-nemoguard-8b-topic-control'. Use repo_type argument if needed.

Steps to Reproduce

  1. Move a model llama-3.1-nemoguard-8b-topic-control to a s3 bucket
  2. Deploy AIBrix following the docs: https://aibrix.readthedocs.io/latest/features/lora-dynamic-loading.html#create-base-model. Be sure to reference the model using its s3 path: s3://s3-bucket/llama-3.1-nemoguard-8b-topic-control
  3. Inspect the pod logs and find the error. Pod will crash when applying.
  4. Same issue if you just load LoRA from s3

Expected behavior

Pod should just start running like when downloading model from HF

Environment

  • AIBrix version: 0.20.0
  • k8s deployment
  • AWS s3 bucket with model files
@Jeffwan Jeffwan added kind/bug Something isn't working area/lora labels Mar 1, 2025
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
area/lora kind/bug Something isn't working
Projects
None yet
Development

No branches or pull requests

2 participants