Skip to content

[#3334][feat] Support of CPU Inference for Scaffolding via PyTorch #4639

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Draft
wants to merge 1 commit into
base: main
Choose a base branch
from

Conversation

amemov
Copy link
Contributor

@amemov amemov commented May 25, 2025

Description

Part of #3706. Addresses #3334 and #3333

Provides a Worker pytorch_worker.py to run inference directly on CPU

Examples

Described in examples/scaffolding/contrib/PytorchCPU/pytorch_worker_run.py - similar to TRTLLMWorker in how the worker is initialized

@amemov amemov force-pushed the cpu-inference-w-scaffolding branch from 09a46af to ad2bcbf Compare May 25, 2025 15:03
@juney-nvidia juney-nvidia added Community want to contribute PRs initiated from Community Community Engagement help/insights needed from community labels May 26, 2025
@amemov amemov force-pushed the cpu-inference-w-scaffolding branch from ad2bcbf to 5c1f5e7 Compare May 27, 2025 00:06
@poweiw poweiw added the Generic Runtime General operational aspects of TRTLLM execution not in other categories. label Jun 5, 2025
@poweiw poweiw requested a review from dcampora June 5, 2025 20:27
@poweiw poweiw added the triaged Issue has been triaged by maintainers label Jun 5, 2025
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
Community Engagement help/insights needed from community Community want to contribute PRs initiated from Community Generic Runtime General operational aspects of TRTLLM execution not in other categories. triaged Issue has been triaged by maintainers
Projects
None yet
Development

Successfully merging this pull request may close these issues.

3 participants