diff --git a/README.md b/README.md index 0dcd63121..133342415 100644 --- a/README.md +++ b/README.md @@ -27,7 +27,7 @@ TruLens-Eval has two key value propositions: * Anything that is tracked by the instrumentation can be evaluated! The process for building your evaluated and tracked LLM application with TruLens is shown below 👇 -![Architecture Diagram](https://www.trulens.org/Assets/image/TruLens_Architecture.png) +![Architecture Diagram](https://www.trulens.org/assets/images/TruLens_Architecture.png) ### Installation and setup diff --git a/docs/Assets/Scripts/app.js b/docs/assets/Scripts/app.js similarity index 100% rename from docs/Assets/Scripts/app.js rename to docs/assets/Scripts/app.js diff --git a/docs/Assets/favicon.svg b/docs/assets/favicon.svg similarity index 100% rename from docs/Assets/favicon.svg rename to docs/assets/favicon.svg diff --git a/docs/Assets/favicon/android-chrome-192x192.png b/docs/assets/favicon/android-chrome-192x192.png similarity index 100% rename from docs/Assets/favicon/android-chrome-192x192.png rename to docs/assets/favicon/android-chrome-192x192.png diff --git a/docs/Assets/favicon/android-chrome-512x512.png b/docs/assets/favicon/android-chrome-512x512.png similarity index 100% rename from docs/Assets/favicon/android-chrome-512x512.png rename to docs/assets/favicon/android-chrome-512x512.png diff --git a/docs/Assets/favicon/apple-touch-icon.png b/docs/assets/favicon/apple-touch-icon.png similarity index 100% rename from docs/Assets/favicon/apple-touch-icon.png rename to docs/assets/favicon/apple-touch-icon.png diff --git a/docs/Assets/favicon/browserconfig.xml b/docs/assets/favicon/browserconfig.xml similarity index 100% rename from docs/Assets/favicon/browserconfig.xml rename to docs/assets/favicon/browserconfig.xml diff --git a/docs/Assets/favicon/favicon-16x16.png b/docs/assets/favicon/favicon-16x16.png similarity index 100% rename from docs/Assets/favicon/favicon-16x16.png rename to docs/assets/favicon/favicon-16x16.png diff --git a/docs/Assets/favicon/favicon-32x32.png b/docs/assets/favicon/favicon-32x32.png similarity index 100% rename from docs/Assets/favicon/favicon-32x32.png rename to docs/assets/favicon/favicon-32x32.png diff --git a/docs/Assets/favicon/favicon.ico b/docs/assets/favicon/favicon.ico similarity index 100% rename from docs/Assets/favicon/favicon.ico rename to docs/assets/favicon/favicon.ico diff --git a/docs/Assets/favicon/mstile-144x144.png b/docs/assets/favicon/mstile-144x144.png similarity index 100% rename from docs/Assets/favicon/mstile-144x144.png rename to docs/assets/favicon/mstile-144x144.png diff --git a/docs/Assets/favicon/mstile-150x150.png b/docs/assets/favicon/mstile-150x150.png similarity index 100% rename from docs/Assets/favicon/mstile-150x150.png rename to docs/assets/favicon/mstile-150x150.png diff --git a/docs/Assets/favicon/mstile-310x150.png b/docs/assets/favicon/mstile-310x150.png similarity index 100% rename from docs/Assets/favicon/mstile-310x150.png rename to docs/assets/favicon/mstile-310x150.png diff --git a/docs/Assets/favicon/mstile-310x310.png b/docs/assets/favicon/mstile-310x310.png similarity index 100% rename from docs/Assets/favicon/mstile-310x310.png rename to docs/assets/favicon/mstile-310x310.png diff --git a/docs/Assets/favicon/mstile-70x70.png b/docs/assets/favicon/mstile-70x70.png similarity index 100% rename from docs/Assets/favicon/mstile-70x70.png rename to docs/assets/favicon/mstile-70x70.png diff --git a/docs/Assets/favicon/safari-pinned-tab.svg b/docs/assets/favicon/safari-pinned-tab.svg similarity index 100% rename from docs/Assets/favicon/safari-pinned-tab.svg rename to docs/assets/favicon/safari-pinned-tab.svg diff --git a/docs/Assets/favicon/site.webmanifest b/docs/assets/favicon/site.webmanifest similarity index 100% rename from docs/Assets/favicon/site.webmanifest rename to docs/assets/favicon/site.webmanifest diff --git a/docs/Assets/image/Chain_Explore.png b/docs/assets/images/Chain_Explore.png similarity index 100% rename from docs/Assets/image/Chain_Explore.png rename to docs/assets/images/Chain_Explore.png diff --git a/docs/Assets/image/Evaluations.png b/docs/assets/images/Evaluations.png similarity index 100% rename from docs/Assets/image/Evaluations.png rename to docs/assets/images/Evaluations.png diff --git a/docs/Assets/image/Leaderboard.png b/docs/assets/images/Leaderboard.png similarity index 100% rename from docs/Assets/image/Leaderboard.png rename to docs/assets/images/Leaderboard.png diff --git a/docs/Assets/image/Neural_Network_Explainability.png b/docs/assets/images/Neural_Network_Explainability.png similarity index 100% rename from docs/Assets/image/Neural_Network_Explainability.png rename to docs/assets/images/Neural_Network_Explainability.png diff --git a/docs/Assets/image/TruLens_Architecture.png b/docs/assets/images/TruLens_Architecture.png similarity index 100% rename from docs/Assets/image/TruLens_Architecture.png rename to docs/assets/images/TruLens_Architecture.png diff --git a/docs/trulens_eval/Assets/image/appui/apps.png b/docs/assets/images/appui/apps.png similarity index 100% rename from docs/trulens_eval/Assets/image/appui/apps.png rename to docs/assets/images/appui/apps.png diff --git a/docs/trulens_eval/Assets/image/appui/blank_session.png b/docs/assets/images/appui/blank_session.png similarity index 100% rename from docs/trulens_eval/Assets/image/appui/blank_session.png rename to docs/assets/images/appui/blank_session.png diff --git a/docs/trulens_eval/Assets/image/appui/running_session.png b/docs/assets/images/appui/running_session.png similarity index 100% rename from docs/trulens_eval/Assets/image/appui/running_session.png rename to docs/assets/images/appui/running_session.png diff --git a/docs/trulens_eval/Assets/image/Chain_Explore.png b/docs/trulens_eval/Assets/image/Chain_Explore.png deleted file mode 100644 index a0630e7bc..000000000 Binary files a/docs/trulens_eval/Assets/image/Chain_Explore.png and /dev/null differ diff --git a/docs/trulens_eval/Assets/image/Evaluations.png b/docs/trulens_eval/Assets/image/Evaluations.png deleted file mode 100644 index cbbaac15b..000000000 Binary files a/docs/trulens_eval/Assets/image/Evaluations.png and /dev/null differ diff --git a/docs/trulens_eval/Assets/image/Leaderboard.png b/docs/trulens_eval/Assets/image/Leaderboard.png deleted file mode 100644 index 9a91e7872..000000000 Binary files a/docs/trulens_eval/Assets/image/Leaderboard.png and /dev/null differ diff --git a/docs/trulens_eval/Assets/image/TruLens_Architecture.png b/docs/trulens_eval/Assets/image/TruLens_Architecture.png deleted file mode 100644 index c05555bfd..000000000 Binary files a/docs/trulens_eval/Assets/image/TruLens_Architecture.png and /dev/null differ diff --git a/docs/trulens_eval/gh_top_intro.md b/docs/trulens_eval/gh_top_intro.md index 0dff33007..c31341385 100644 --- a/docs/trulens_eval/gh_top_intro.md +++ b/docs/trulens_eval/gh_top_intro.md @@ -27,7 +27,7 @@ TruLens-Eval has two key value propositions: * Anything that is tracked by the instrumentation can be evaluated! The process for building your evaluated and tracked LLM application with TruLens is shown below 👇 -![Architecture Diagram](https://www.trulens.org/Assets/image/TruLens_Architecture.png) +![Architecture Diagram](https://www.trulens.org/assets/images/TruLens_Architecture.png) ### Installation and setup diff --git a/docs/trulens_eval/intro.md b/docs/trulens_eval/intro.md index 9368300bf..d325cf350 100644 --- a/docs/trulens_eval/intro.md +++ b/docs/trulens_eval/intro.md @@ -1,6 +1,6 @@ # Welcome to TruLens-Eval! -![TruLens](https://www.trulens.org/Assets/image/Neural_Network_Explainability.png) +![TruLens](https://www.trulens.org/assets/images/Neural_Network_Explainability.png) Evaluate and track your LLM experiments with TruLens. As you work on your models and prompts TruLens-Eval supports the iterative development and of a wide range of LLM applications by wrapping your application to log key metadata across the entire chain (or off chain if your project does not use chains) on your local machine. @@ -20,7 +20,7 @@ TruLens-Eval has two key value propositions: The process for building your evaluated and tracked LLM application with TruLens is below 👇 -![Architecture Diagram](https://www.trulens.org/Assets/image/TruLens_Architecture.png) +![Architecture Diagram](https://www.trulens.org/assets/images/TruLens_Architecture.png) ## Installation and Setup diff --git a/trulens_eval/README.md b/trulens_eval/README.md index 9368300bf..d325cf350 100644 --- a/trulens_eval/README.md +++ b/trulens_eval/README.md @@ -1,6 +1,6 @@ # Welcome to TruLens-Eval! -![TruLens](https://www.trulens.org/Assets/image/Neural_Network_Explainability.png) +![TruLens](https://www.trulens.org/assets/images/Neural_Network_Explainability.png) Evaluate and track your LLM experiments with TruLens. As you work on your models and prompts TruLens-Eval supports the iterative development and of a wide range of LLM applications by wrapping your application to log key metadata across the entire chain (or off chain if your project does not use chains) on your local machine. @@ -20,7 +20,7 @@ TruLens-Eval has two key value propositions: The process for building your evaluated and tracked LLM application with TruLens is below 👇 -![Architecture Diagram](https://www.trulens.org/Assets/image/TruLens_Architecture.png) +![Architecture Diagram](https://www.trulens.org/assets/images/TruLens_Architecture.png) ## Installation and Setup diff --git a/trulens_eval/examples/quickstart/colab/langchain_quickstart_colab.ipynb b/trulens_eval/examples/quickstart/colab/langchain_quickstart_colab.ipynb index b895c9bb5..ffa58baee 100644 --- a/trulens_eval/examples/quickstart/colab/langchain_quickstart_colab.ipynb +++ b/trulens_eval/examples/quickstart/colab/langchain_quickstart_colab.ipynb @@ -302,7 +302,7 @@ "\n", "Note: Average feedback values are returned and displayed in a range from 0 (worst) to 1 (best).\n", "\n", - "![Chain Leaderboard](https://www.trulens.org/Assets/image/Leaderboard.png)\n", + "![Chain Leaderboard](https://www.trulens.org/assets/images/Leaderboard.png)\n", "\n", "To dive deeper on a particular chain, click \"Select Chain\".\n", "\n", @@ -312,13 +312,13 @@ "\n", "The evaluations tab provides record-level metadata and feedback on the quality of your LLM application.\n", "\n", - "![Evaluations](https://www.trulens.org/Assets/image/Leaderboard.png)\n", + "![Evaluations](https://www.trulens.org/assets/images/Leaderboard.png)\n", "\n", "### Deep dive into full chain metadata\n", "\n", "Click on a record to dive deep into all of the details of your chain stack and underlying LLM, captured by tru_chain_recorder.\n", "\n", - "![Explore a Chain](https://www.trulens.org/Assets/image/Chain_Explore.png)\n", + "![Explore a Chain](https://www.trulens.org/assets/images/Chain_Explore.png)\n", "\n", "If you prefer the raw format, you can quickly get it using the \"Display full chain json\" or \"Display full record json\" buttons at the bottom of the page." ] diff --git a/trulens_eval/examples/quickstart/dashboard_appui.ipynb b/trulens_eval/examples/quickstart/dashboard_appui.ipynb index db92d239e..46ee10c3d 100644 --- a/trulens_eval/examples/quickstart/dashboard_appui.ipynb +++ b/trulens_eval/examples/quickstart/dashboard_appui.ipynb @@ -8,11 +8,11 @@ "\n", "This notebook describes how to run your apps from the streamlit dashboard. Following this notebook, you should be able to access your apps and interact with them within the streamlit dashboard under the **Apps** page (see screenshot below). Make sure to check the **Setting up** section below to get your app in the list of apps on that page.\n", "\n", - "![App Runner](https://www.trulens.org/Assets/image/appui/apps.png)\n", + "![App Runner](https://www.trulens.org/assets/images/appui/apps.png)\n", "\n", "Clicking *New session* under any of these apps will bring up an empty transcript of the interactions between the user (you) and the app (see screenshot below). Typing a message under *Your message* on the bottom of the window, and pressing enter, will run your app with that specified message as input, produce the app output, and add both to the chat transcript under the *Records* column.\n", "\n", - "![Blank Session](https://www.trulens.org/Assets/image/appui/blank_session.png)\n", + "![Blank Session](https://www.trulens.org/assets/images/appui/blank_session.png)\n", "\n", "Several other inputs are present on this page which control what about the produced transcript record to show alongside their inputs/outputs.\n", "\n", @@ -24,7 +24,7 @@ "\n", "An example of a running session with several selectors is shown in the following screenshot:\n", "\n", - "![Running Session](https://www.trulens.org/Assets/image/appui/running_session.png)\n", + "![Running Session](https://www.trulens.org/assets/images/appui/running_session.png)\n", "\n", "The session is preserved when navigating away from this page, letting you inspect the produced records in the **Evaluation** page, for example. To create a new session, you first need to end the existing one by pressing the \"End session\" button on top of the runner page." ] diff --git a/trulens_eval/examples/quickstart/langchain_quickstart.ipynb b/trulens_eval/examples/quickstart/langchain_quickstart.ipynb index 001d20ebd..5b6c550a2 100644 --- a/trulens_eval/examples/quickstart/langchain_quickstart.ipynb +++ b/trulens_eval/examples/quickstart/langchain_quickstart.ipynb @@ -265,7 +265,7 @@ "\n", "Note: Average feedback values are returned and displayed in a range from 0 (worst) to 1 (best).\n", "\n", - "![Chain Leaderboard](https://www.trulens.org/Assets/image/Leaderboard.png)\n", + "![Chain Leaderboard](https://www.trulens.org/assets/images/Leaderboard.png)\n", "\n", "To dive deeper on a particular chain, click \"Select Chain\".\n", "\n", @@ -275,13 +275,13 @@ "\n", "The evaluations tab provides record-level metadata and feedback on the quality of your LLM application.\n", "\n", - "![Evaluations](https://www.trulens.org/Assets/image/Leaderboard.png)\n", + "![Evaluations](https://www.trulens.org/assets/images/Leaderboard.png)\n", "\n", "### Deep dive into full chain metadata\n", "\n", "Click on a record to dive deep into all of the details of your chain stack and underlying LLM, captured by tru_chain_recorder.\n", "\n", - "![Explore a Chain](https://www.trulens.org/Assets/image/Chain_Explore.png)\n", + "![Explore a Chain](https://www.trulens.org/assets/images/Chain_Explore.png)\n", "\n", "If you prefer the raw format, you can quickly get it using the \"Display full chain json\" or \"Display full record json\" buttons at the bottom of the page." ] diff --git a/trulens_eval/examples/quickstart/py_script_quickstarts/all_tools.py b/trulens_eval/examples/quickstart/py_script_quickstarts/all_tools.py index d89fa525a..7ae09545c 100644 --- a/trulens_eval/examples/quickstart/py_script_quickstarts/all_tools.py +++ b/trulens_eval/examples/quickstart/py_script_quickstarts/all_tools.py @@ -166,7 +166,7 @@ # # Note: Average feedback values are returned and displayed in a range from 0 (worst) to 1 (best). # -# ![Chain Leaderboard](https://www.trulens.org/Assets/image/Leaderboard.png) +# ![Chain Leaderboard](https://www.trulens.org/assets/images/Leaderboard.png) # # To dive deeper on a particular chain, click "Select Chain". # @@ -176,13 +176,13 @@ # # The evaluations tab provides record-level metadata and feedback on the quality of your LLM application. # -# ![Evaluations](https://www.trulens.org/Assets/image/Leaderboard.png) +# ![Evaluations](https://www.trulens.org/assets/images/Leaderboard.png) # # ### Deep dive into full chain metadata # # Click on a record to dive deep into all of the details of your chain stack and underlying LLM, captured by tru_chain_recorder. # -# ![Explore a Chain](https://www.trulens.org/Assets/image/Chain_Explore.png) +# ![Explore a Chain](https://www.trulens.org/assets/images/Chain_Explore.png) # # If you prefer the raw format, you can quickly get it using the "Display full chain json" or "Display full record json" buttons at the bottom of the page. diff --git a/trulens_eval/examples/quickstart/py_script_quickstarts/langchain_quickstart.py b/trulens_eval/examples/quickstart/py_script_quickstarts/langchain_quickstart.py index 8fcecf53e..641bfd889 100644 --- a/trulens_eval/examples/quickstart/py_script_quickstarts/langchain_quickstart.py +++ b/trulens_eval/examples/quickstart/py_script_quickstarts/langchain_quickstart.py @@ -166,7 +166,7 @@ # # Note: Average feedback values are returned and displayed in a range from 0 (worst) to 1 (best). # -# ![Chain Leaderboard](https://www.trulens.org/Assets/image/Leaderboard.png) +# ![Chain Leaderboard](https://www.trulens.org/assets/images/Leaderboard.png) # # To dive deeper on a particular chain, click "Select Chain". # @@ -176,13 +176,13 @@ # # The evaluations tab provides record-level metadata and feedback on the quality of your LLM application. # -# ![Evaluations](https://www.trulens.org/Assets/image/Leaderboard.png) +# ![Evaluations](https://www.trulens.org/assets/images/Leaderboard.png) # # ### Deep dive into full chain metadata # # Click on a record to dive deep into all of the details of your chain stack and underlying LLM, captured by tru_chain_recorder. # -# ![Explore a Chain](https://www.trulens.org/Assets/image/Chain_Explore.png) +# ![Explore a Chain](https://www.trulens.org/assets/images/Chain_Explore.png) # # If you prefer the raw format, you can quickly get it using the "Display full chain json" or "Display full record json" buttons at the bottom of the page. diff --git a/trulens_eval/generated_files/all_tools.ipynb b/trulens_eval/generated_files/all_tools.ipynb index 48bcf103e..1826ab33a 100644 --- a/trulens_eval/generated_files/all_tools.ipynb +++ b/trulens_eval/generated_files/all_tools.ipynb @@ -265,7 +265,7 @@ "\n", "Note: Average feedback values are returned and displayed in a range from 0 (worst) to 1 (best).\n", "\n", - "![Chain Leaderboard](https://www.trulens.org/Assets/image/Leaderboard.png)\n", + "![Chain Leaderboard](https://www.trulens.org/assets/images/Leaderboard.png)\n", "\n", "To dive deeper on a particular chain, click \"Select Chain\".\n", "\n", @@ -275,13 +275,13 @@ "\n", "The evaluations tab provides record-level metadata and feedback on the quality of your LLM application.\n", "\n", - "![Evaluations](https://www.trulens.org/Assets/image/Leaderboard.png)\n", + "![Evaluations](https://www.trulens.org/assets/images/Leaderboard.png)\n", "\n", "### Deep dive into full chain metadata\n", "\n", "Click on a record to dive deep into all of the details of your chain stack and underlying LLM, captured by tru_chain_recorder.\n", "\n", - "![Explore a Chain](https://www.trulens.org/Assets/image/Chain_Explore.png)\n", + "![Explore a Chain](https://www.trulens.org/assets/images/Chain_Explore.png)\n", "\n", "If you prefer the raw format, you can quickly get it using the \"Display full chain json\" or \"Display full record json\" buttons at the bottom of the page." ] diff --git a/trulens_eval/generated_files/all_tools.py b/trulens_eval/generated_files/all_tools.py index bd523f4b3..f3e9c3103 100644 --- a/trulens_eval/generated_files/all_tools.py +++ b/trulens_eval/generated_files/all_tools.py @@ -96,7 +96,7 @@ # # Note: Average feedback values are returned and printed in a range from 0 (worst) to 1 (best). # -# ![Chain Leaderboard](https://www.trulens.org/Assets/image/Leaderboard.png) +# ![Chain Leaderboard](https://www.trulens.org/assets/images/Leaderboard.png) # # To dive deeper on a particular chain, click "Select Chain". # @@ -106,13 +106,13 @@ # # The evaluations tab provides record-level metadata and feedback on the quality of your LLM application. # -# ![Evaluations](https://www.trulens.org/Assets/image/Leaderboard.png) +# ![Evaluations](https://www.trulens.org/assets/images/Leaderboard.png) # # ### Deep dive into full chain metadata # # Click on a record to dive deep into all of the details of your chain stack and underlying LLM, captured by tru_chain_recorder. # -# ![Explore a Chain](https://www.trulens.org/Assets/image/Chain_Explore.png) +# ![Explore a Chain](https://www.trulens.org/assets/images/Chain_Explore.png) # # If you prefer the raw format, you can quickly get it using the "Display full chain json" or "Display full record json" buttons at the bottom of the page. diff --git a/trulens_explain/README.md b/trulens_explain/README.md index 88780104a..a5937f27c 100644 --- a/trulens_explain/README.md +++ b/trulens_explain/README.md @@ -1,6 +1,6 @@ # Welcome to TruLens! -![TruLens](https://www.trulens.org/Assets/image/Neural_Network_Explainability.png) +![TruLens](https://www.trulens.org/assets/images/Neural_Network_Explainability.png) TruLens is a cross-framework library for deep learning explainability. It provides a uniform abstraction over a number of different frameworks. It provides a uniform abstraction layer over TensorFlow, PyTorch, and Keras and allows input and internal explanations.