-
Couldn't load subscription status.
- Fork 552
feat(benchmark): Create mock LLM server for use in benchmarks #1403
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
base: develop
Are you sure you want to change the base?
Changes from all commits
1bb4443
d9b73be
9021b81
687e33b
c0afd8d
e62f394
6ddcaca
3b3f49a
f142c0f
a18b514
6beb888
c056b3b
4104a1f
1cca2ff
e87715c
File filter
Filter by extension
Conversations
Jump to
Diff view
Diff view
There are no files selected for viewing
| Original file line number | Diff line number | Diff line change |
|---|---|---|
| @@ -0,0 +1,14 @@ | ||
| # SPDX-FileCopyrightText: Copyright (c) 2023 NVIDIA CORPORATION & AFFILIATES. All rights reserved. | ||
| # SPDX-License-Identifier: Apache-2.0 | ||
| # | ||
| # Licensed under the Apache License, Version 2.0 (the "License"); | ||
| # you may not use this file except in compliance with the License. | ||
| # You may obtain a copy of the License at | ||
| # | ||
| # http://www.apache.org/licenses/LICENSE-2.0 | ||
| # | ||
| # Unless required by applicable law or agreed to in writing, software | ||
| # distributed under the License is distributed on an "AS IS" BASIS, | ||
| # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | ||
| # See the License for the specific language governing permissions and | ||
| # limitations under the License. | ||
| Original file line number | Diff line number | Diff line change |
|---|---|---|
| @@ -0,0 +1,14 @@ | ||
| # SPDX-FileCopyrightText: Copyright (c) 2023 NVIDIA CORPORATION & AFFILIATES. All rights reserved. | ||
| # SPDX-License-Identifier: Apache-2.0 | ||
| # | ||
| # Licensed under the Apache License, Version 2.0 (the "License"); | ||
| # you may not use this file except in compliance with the License. | ||
| # You may obtain a copy of the License at | ||
| # | ||
| # http://www.apache.org/licenses/LICENSE-2.0 | ||
| # | ||
| # Unless required by applicable law or agreed to in writing, software | ||
| # distributed under the License is distributed on an "AS IS" BASIS, | ||
| # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | ||
| # See the License for the specific language governing permissions and | ||
| # limitations under the License. |
| Original file line number | Diff line number | Diff line change |
|---|---|---|
| @@ -0,0 +1,240 @@ | ||
| # SPDX-FileCopyrightText: Copyright (c) 2023 NVIDIA CORPORATION & AFFILIATES. All rights reserved. | ||
| # SPDX-License-Identifier: Apache-2.0 | ||
| # | ||
| # Licensed under the Apache License, Version 2.0 (the "License"); | ||
| # you may not use this file except in compliance with the License. | ||
| # You may obtain a copy of the License at | ||
| # | ||
| # http://www.apache.org/licenses/LICENSE-2.0 | ||
| # | ||
| # Unless required by applicable law or agreed to in writing, software | ||
| # distributed under the License is distributed on an "AS IS" BASIS, | ||
| # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | ||
| # See the License for the specific language governing permissions and | ||
| # limitations under the License. | ||
|
|
||
|
|
||
| import asyncio | ||
| import logging | ||
| import time | ||
| from typing import Annotated, Optional, Union | ||
|
|
||
| from fastapi import Depends, FastAPI, HTTPException, Request, Response | ||
|
|
||
| from nemoguardrails.benchmark.mock_llm_server.config import ( # get_config, | ||
| ModelSettings, | ||
| get_settings, | ||
| ) | ||
| from nemoguardrails.benchmark.mock_llm_server.models import ( | ||
| ChatCompletionChoice, | ||
| ChatCompletionRequest, | ||
| ChatCompletionResponse, | ||
| CompletionChoice, | ||
| CompletionRequest, | ||
| CompletionResponse, | ||
| Message, | ||
| Model, | ||
| ModelsResponse, | ||
| Usage, | ||
| ) | ||
| from nemoguardrails.benchmark.mock_llm_server.response_data import ( | ||
| calculate_tokens, | ||
| generate_id, | ||
| get_latency_seconds, | ||
| get_response, | ||
| ) | ||
|
|
||
| # Create a console logging handler | ||
| log = logging.getLogger(__name__) | ||
| log.setLevel(logging.INFO) # TODO Control this from the CLi args | ||
|
|
||
| # Create a formatter to define the log message format | ||
| formatter = logging.Formatter( | ||
| "%(asctime)s %(levelname)s: %(message)s", datefmt="%Y-%m-%d %H:%M:%S" | ||
| ) | ||
|
|
||
| # Create a console handler to print logs to the console | ||
| console_handler = logging.StreamHandler() | ||
| console_handler.setLevel(logging.INFO) # DEBUG and higher will go to the console | ||
| console_handler.setFormatter(formatter) | ||
|
|
||
| # Add console handler to logs | ||
| log.addHandler(console_handler) | ||
|
|
||
|
|
||
| ModelSettingsDep = Annotated[ModelSettings, Depends(get_settings)] | ||
|
|
||
|
|
||
| def _validate_request_model( | ||
| config: ModelSettingsDep, | ||
| request: Union[CompletionRequest, ChatCompletionRequest], | ||
| ) -> None: | ||
| """Check the Completion or Chat Completion `model` field is in our supported model list""" | ||
| if request.model != config.model: | ||
| raise HTTPException( | ||
| status_code=400, | ||
| detail=f"Model '{request.model}' not found. Available models: {config.model}", | ||
| ) | ||
|
|
||
|
|
||
| app = FastAPI( | ||
| title="Mock LLM Server", | ||
| description="OpenAI-compatible mock LLM server for testing and benchmarking", | ||
| version="0.0.1", | ||
| ) | ||
|
|
||
|
|
||
| @app.middleware("http") | ||
| async def log_http_duration(request: Request, call_next): | ||
| """ | ||
| Middleware to log incoming requests and their responses. | ||
| """ | ||
| request_time = time.time() | ||
| response = await call_next(request) | ||
| response_time = time.time() | ||
|
|
||
| duration_seconds = response_time - request_time | ||
| log.info( | ||
| "Request finished: %s, took %.3f seconds", | ||
| response.status_code, | ||
| duration_seconds, | ||
| ) | ||
| return response | ||
|
|
||
|
|
||
| @app.get("/") | ||
| async def root(config: ModelSettingsDep): | ||
| """Root endpoint with basic server information.""" | ||
| return { | ||
| "message": "Mock LLM Server", | ||
| "version": "0.0.1", | ||
| "description": f"OpenAI-compatible mock LLM server for model: {config.model}", | ||
| "endpoints": ["/v1/models", "/v1/chat/completions", "/v1/completions"], | ||
| "model_configuration": config, | ||
| } | ||
|
|
||
|
|
||
| @app.get("/v1/models", response_model=ModelsResponse) | ||
| async def list_models(config: ModelSettingsDep): | ||
| """List available models.""" | ||
| log.debug("/v1/models request") | ||
|
|
||
| model = Model( | ||
| id=config.model, object="model", created=int(time.time()), owned_by="system" | ||
| ) | ||
| response = ModelsResponse(object="list", data=[model]) | ||
| log.debug("/v1/models response: %s", response) | ||
| return response | ||
|
|
||
|
|
||
| @app.post("/v1/chat/completions", response_model=ChatCompletionResponse) | ||
| async def chat_completions( | ||
| request: ChatCompletionRequest, config: ModelSettingsDep | ||
| ) -> ChatCompletionResponse: | ||
| """Create a chat completion.""" | ||
|
|
||
| log.debug("/v1/chat/completions request: %s", request) | ||
|
|
||
| # Validate model exists | ||
| _validate_request_model(config, request) | ||
|
|
||
| # Generate dummy response | ||
| response_content = get_response(config) | ||
| response_latency_seconds = get_latency_seconds(config, seed=12345) | ||
|
|
||
| # Calculate token usage | ||
| prompt_text = " ".join([msg.content for msg in request.messages]) | ||
| prompt_tokens = calculate_tokens(prompt_text) | ||
| completion_tokens = calculate_tokens(response_content) | ||
|
|
||
| # Create response | ||
| completion_id = generate_id("chatcmpl") | ||
| created_timestamp = int(time.time()) | ||
|
|
||
| choices = [] | ||
| for i in range(request.n or 1): | ||
| choice = ChatCompletionChoice( | ||
| index=i, | ||
| message=Message(role="assistant", content=response_content), | ||
| finish_reason="stop", | ||
| ) | ||
| choices.append(choice) | ||
|
|
||
| response = ChatCompletionResponse( | ||
| id=completion_id, | ||
| object="chat.completion", | ||
| created=created_timestamp, | ||
| model=request.model, | ||
| choices=choices, | ||
| usage=Usage( | ||
| prompt_tokens=prompt_tokens, | ||
| completion_tokens=completion_tokens, | ||
| total_tokens=prompt_tokens + completion_tokens, | ||
| ), | ||
| ) | ||
| await asyncio.sleep(response_latency_seconds) | ||
| log.debug("/v1/chat/completions response: %s", response) | ||
| return response | ||
|
|
||
|
|
||
| @app.post("/v1/completions", response_model=CompletionResponse) | ||
|
There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. are you using completions in your benchmarking? If not, I think it is better not to support this legacy interface (https://platform.openai.com/docs/api-reference/completions/create) |
||
| async def completions( | ||
| request: CompletionRequest, config: ModelSettingsDep | ||
| ) -> CompletionResponse: | ||
| """Create a text completion.""" | ||
|
|
||
| log.debug("/v1/completions request: %s", request) | ||
|
|
||
| # Validate model exists | ||
| _validate_request_model(config, request) | ||
|
|
||
| # Handle prompt (can be string or list) | ||
| if isinstance(request.prompt, list): | ||
| prompt_text = " ".join(request.prompt) | ||
| else: | ||
| prompt_text = request.prompt | ||
|
|
||
| # Generate dummy response | ||
| response_text = get_response(config) | ||
| response_latency_seconds = get_latency_seconds(config, seed=12345) | ||
|
|
||
| # Calculate token usage | ||
| prompt_tokens = calculate_tokens(prompt_text) | ||
| completion_tokens = calculate_tokens(response_text) | ||
|
|
||
| # Create response | ||
| completion_id = generate_id("cmpl") | ||
| created_timestamp = int(time.time()) | ||
|
|
||
| choices = [] | ||
| for i in range(request.n or 1): | ||
| choice = CompletionChoice( | ||
| text=response_text, index=i, logprobs=None, finish_reason="stop" | ||
| ) | ||
| choices.append(choice) | ||
|
|
||
| response = CompletionResponse( | ||
| id=completion_id, | ||
| object="text_completion", | ||
| created=created_timestamp, | ||
| model=request.model, | ||
| choices=choices, | ||
| usage=Usage( | ||
| prompt_tokens=prompt_tokens, | ||
| completion_tokens=completion_tokens, | ||
| total_tokens=prompt_tokens + completion_tokens, | ||
| ), | ||
| ) | ||
|
|
||
| await asyncio.sleep(response_latency_seconds) | ||
| log.debug("/v1/completions response: %s", response) | ||
| return response | ||
|
|
||
|
|
||
| @app.get("/health") | ||
| async def health_check(): | ||
| """Health check endpoint.""" | ||
| log.debug("/health request") | ||
| response = {"status": "healthy", "timestamp": int(time.time())} | ||
| log.debug("/health response: %s", response) | ||
| return response | ||
| Original file line number | Diff line number | Diff line change |
|---|---|---|
| @@ -0,0 +1,68 @@ | ||
| # SPDX-FileCopyrightText: Copyright (c) 2023 NVIDIA CORPORATION & AFFILIATES. All rights reserved. | ||
| # SPDX-License-Identifier: Apache-2.0 | ||
| # | ||
| # Licensed under the Apache License, Version 2.0 (the "License"); | ||
| # you may not use this file except in compliance with the License. | ||
| # You may obtain a copy of the License at | ||
| # | ||
| # http://www.apache.org/licenses/LICENSE-2.0 | ||
| # | ||
| # Unless required by applicable law or agreed to in writing, software | ||
| # distributed under the License is distributed on an "AS IS" BASIS, | ||
| # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | ||
| # See the License for the specific language governing permissions and | ||
| # limitations under the License. | ||
|
|
||
| import os | ||
| from functools import lru_cache | ||
| from pathlib import Path | ||
| from typing import Any, Optional, Union | ||
|
|
||
| import yaml | ||
| from pydantic import BaseModel, Field | ||
| from pydantic_settings import ( | ||
| BaseSettings, | ||
| PydanticBaseSettingsSource, | ||
| SettingsConfigDict, | ||
| ) | ||
|
|
||
| CONFIG_FILE_ENV_VAR = "MOCK_LLM_CONFIG_FILE" | ||
| config_file_path = os.getenv(CONFIG_FILE_ENV_VAR, "model_settings.yml") | ||
| CONFIG_FILE = Path(config_file_path) | ||
|
|
||
|
|
||
| class ModelSettings(BaseSettings): | ||
| """Pydantic model to configure the Mock LLM Server.""" | ||
|
|
||
| # Mandatory fields | ||
| model: str = Field(..., description="Model name served by mock server") | ||
| unsafe_probability: float = Field( | ||
| default=0.1, description="Probability of unsafe response (between 0 and 1)" | ||
| ) | ||
| unsafe_text: str = Field(..., description="Refusal response to unsafe prompt") | ||
| safe_text: str = Field(..., description="Safe response") | ||
|
|
||
| # Config with default values | ||
| # Latency sampled from a truncated-normal distribution. | ||
| # Plain Normal distributions have infinite support, and can be negative | ||
| latency_min_seconds: float = Field( | ||
| default=0.1, description="Minimum latency in seconds" | ||
| ) | ||
| latency_max_seconds: float = Field( | ||
| default=5, description="Maximum latency in seconds" | ||
| ) | ||
| latency_mean_seconds: float = Field( | ||
| default=0.5, description="The average response time in seconds" | ||
| ) | ||
| latency_std_seconds: float = Field( | ||
| default=0.1, description="Standard deviation of response time" | ||
| ) | ||
|
|
||
| model_config = SettingsConfigDict(env_file=CONFIG_FILE) | ||
|
|
||
|
|
||
| def get_settings() -> ModelSettings: | ||
| """Singleton-pattern to get settings once via lru_cache""" | ||
| settings = ModelSettings() # type: ignore (These are filled in by loading from CONFIG_FILE) | ||
| print("Returning ModelSettings: %s", settings) | ||
| return settings |
| Original file line number | Diff line number | Diff line change |
|---|---|---|
| @@ -0,0 +1,21 @@ | ||
| models: | ||
| - type: main | ||
| engine: nim | ||
| model: meta/llama-3.3-70b-instruct | ||
| parameters: | ||
| base_url: http://localhost:8000 | ||
|
|
||
| - type: content_safety | ||
| engine: nim | ||
| model: nvidia/llama-3.1-nemoguard-8b-content-safety | ||
| parameters: | ||
| base_url: http://localhost:8001 | ||
|
|
||
|
|
||
| rails: | ||
| input: | ||
| flows: | ||
| - content safety check input $model=content_safety | ||
| output: | ||
| flows: | ||
| - content safety check output $model=content_safety |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Should we be updating the copyright date on new files?
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
glad you pointed this out. We should update our LICENSE.md. I'll open a PR