-
Notifications
You must be signed in to change notification settings - Fork 364
TRT-LLM loading mechanism tool #3398
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
base: main
Are you sure you want to change the base?
Conversation
57dbb3f
to
3e38e87
Compare
f"Ensure the path is correct and the library is compatible", | ||
exc_info=e_os_error, | ||
else: | ||
py_version = f"cp{sys.version_info.major}{sys.version_info.minor}" |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Why do we restrict to cp310 and cp312, It shouldnt matter if we are pulling the whl and unzipping ourselves
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
https://pypi.nvidia.com/tensorrt-llm/ In this since I see the tags for only cp310 and cp312 I added the check
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
There are some changes that do not conform to Python style guidelines:
--- /home/runner/work/TensorRT/TensorRT/tests/py/dynamo/conversion/test_nccl_ops.py 2025-02-27 20:03:00.014038+00:00
+++ /home/runner/work/TensorRT/TensorRT/tests/py/dynamo/conversion/test_nccl_ops.py 2025-02-27 20:03:24.885031+00:00
@@ -22,11 +22,11 @@
from .harness import DispatchTestCase
class TestGatherNcclOpsConverter(DispatchTestCase):
- @parameterized.expand([(8)])
+ @parameterized.expand([8])
def test_nccl_ops(self, linear_layer_dim):
class DistributedGatherModel(nn.Module):
def __init__(self, input_dim):
super().__init__()
self.fc = torch.nn.Linear(input_dim, input_dim)
9ba407b
to
5f3fdac
Compare
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
There are some changes that do not conform to Python style guidelines:
--- /home/runner/work/TensorRT/TensorRT/tests/py/dynamo/conversion/test_nccl_ops.py 2025-02-27 20:05:38.023287+00:00
+++ /home/runner/work/TensorRT/TensorRT/tests/py/dynamo/conversion/test_nccl_ops.py 2025-02-27 20:06:02.662188+00:00
@@ -22,11 +22,11 @@
from .harness import DispatchTestCase
class TestGatherNcclOpsConverter(DispatchTestCase):
- @parameterized.expand([(8)])
+ @parameterized.expand([8])
def test_nccl_ops(self, linear_layer_dim):
class DistributedGatherModel(nn.Module):
def __init__(self, input_dim):
super().__init__()
self.fc = torch.nn.Linear(input_dim, input_dim)
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
There are some changes that do not conform to Python style guidelines:
--- /home/runner/work/TensorRT/TensorRT/tests/py/dynamo/conversion/test_nccl_ops.py 2025-02-27 20:05:54.405311+00:00
+++ /home/runner/work/TensorRT/TensorRT/tests/py/dynamo/conversion/test_nccl_ops.py 2025-02-27 20:06:21.454993+00:00
@@ -22,11 +22,11 @@
from .harness import DispatchTestCase
class TestGatherNcclOpsConverter(DispatchTestCase):
- @parameterized.expand([(8)])
+ @parameterized.expand([8])
def test_nccl_ops(self, linear_layer_dim):
class DistributedGatherModel(nn.Module):
def __init__(self, input_dim):
super().__init__()
self.fc = torch.nn.Linear(input_dim, input_dim)
5f3fdac
to
b66350e
Compare
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
There are some changes that do not conform to Python style guidelines:
--- /home/runner/work/TensorRT/TensorRT/tests/py/dynamo/conversion/test_nccl_ops.py 2025-04-15 19:58:05.267724+00:00
+++ /home/runner/work/TensorRT/TensorRT/tests/py/dynamo/conversion/test_nccl_ops.py 2025-04-15 19:58:36.145897+00:00
@@ -22,11 +22,11 @@
from .harness import DispatchTestCase
class TestGatherNcclOpsConverter(DispatchTestCase):
- @parameterized.expand([(8)])
+ @parameterized.expand([8])
def test_nccl_ops(self, linear_layer_dim):
class DistributedGatherModel(nn.Module):
def __init__(self, input_dim):
super().__init__()
self.fc = torch.nn.Linear(input_dim, input_dim)
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
There are some changes that do not conform to Python style guidelines:
--- /home/runner/work/TensorRT/TensorRT/tests/py/dynamo/conversion/test_nccl_ops.py 2025-04-15 21:00:13.719714+00:00
+++ /home/runner/work/TensorRT/TensorRT/tests/py/dynamo/conversion/test_nccl_ops.py 2025-04-15 21:00:40.093669+00:00
@@ -22,11 +22,11 @@
from .harness import DispatchTestCase
class TestGatherNcclOpsConverter(DispatchTestCase):
- @parameterized.expand([(8)])
+ @parameterized.expand([8])
def test_nccl_ops(self, linear_layer_dim):
class DistributedGatherModel(nn.Module):
def __init__(self, input_dim):
super().__init__()
self.fc = torch.nn.Linear(input_dim, input_dim)
6e893ed
to
77f2145
Compare
f30acb7
to
9c238ae
Compare
89d621d
to
27aa2f2
Compare
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Use a temp directory to save the wheel and unzipped wheel
37d4e90
to
e8bc3a4
Compare
a0cecf4
to
4193ca2
Compare
4589c76
to
1e2148d
Compare
42d4862
to
9cb3cab
Compare
dbfd7ee
to
15d681a
Compare
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Seems like its close are there any tests for the downloader? like verifying the correct file is downloaded and available?
The test t |
…ing in dynamo.compile TRT-LLM installation utilities and adding test cases adding the option in _compiler.py changes in the TRT-LLM loading tool- removing install_wget, install_unzip, install_mpi Further changes in error logging of the TRT-LLM installation tool moving the load_tensorrt_llm to dynamo/utils.py correcting misprint for TRT LLM load Using python lib for download to make it platform agnostic dll file path update for windows correcting the non critical lint error Including version in versions.txt
…action, variable for py_version
c8b8337
to
340182b
Compare
TRT-LLM download utility