Replies: 5 comments
-
|
I think the main argument against it is that it just adds a significant amount of complexity to the basic infrastructure of the stack and the surrounding test suite. The use case (being able to "embed the stack as a library") is nearly non-existent right now. |
Beta Was this translation helpful? Give feedback.
-
Beta Was this translation helpful? Give feedback.
-
|
Several teams in Red Hat leveraged Llama Stack as a library, mainly because it is much easier to deploy + run it that way in different environments. Running Llama Stack in separate process is usually fine, but there are more possible problems that have to be checked and managed (network issues, different versions, whole deployment process etc.). If I may I'd vote to keep it there. There's also one non-technical aspect: we were able to move some projects from using langchain/langraph etc. into Llama Stack, because "just" the inference call is different (sorry, simplification), everything else in the project can remain as is. This encourages the teams to try Llama Stack in the 1st place. |
Beta Was this translation helpful? Give feedback.
-
|
I'd like to add my voice in strong support of keeping LlamaStackAsLibraryClient. For our use case, the library client isn't just a tool for notebooks or local development; it's a deliberate architectural choice for our containerized deployments that significantly reduces operational complexity. While I understand the goal of focusing on a "hosted stack", this pattern of embedding the stack as a library directly translates to minimal operational overhead. It lowers the barrier not just for initial experimentation, as @mattf pointed out, but also for production deployment. |
Beta Was this translation helpful? Give feedback.
-
|
Thanks for the feedback here. It sounds like we should keep it and strive to make the underlying infrastructure of managing the library client easier. |
Beta Was this translation helpful? Give feedback.


Uh oh!
There was an error while loading. Please reload this page.
-
The history predates my time here but I think the reason for introducing
LlamaStackAsLibraryClientwas to have a good DevX in notebook environment. Stack has since evolved quite a bit and now might be a good time to revisit the necessity of this.LlamaStackAsLibraryClientno longer fits this narrative as there's no production scenario that makes use of it.Question for the group: do we still have a valid case for keeping LlamaStackAsLibraryClient? If we do want something like this, we could create a helper function that spins up the stack server in the background.
Beta Was this translation helpful? Give feedback.
All reactions