You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Description
All our attempts to create a conda environment, and leveraging it inside the triton serving container have failed. We're following the instructions for conda-pack on the Python backend snippet.
We're running a local test on a mac with no support to nvidia gpus. The eventual target is sagemaker instances.
Are you using the Triton container or did you build it yourself?
Using 24.02-pyt-python-py3 with additional environment variables for SageMaker endpoints
Describe the models (framework, inputs, outputs), ideally include the model configuration file (if using an ensemble include the model configuration file for that as well).
This is a simple Python backend, running models.py that just imports the needed packages. For example:
importtransformers
Expected behavior
A clear and concise description of what you expected to happen.
Creating a conda environment outside of our serving container, packaging it using conda-pack and leveraging it with Triton Serving Container.
Note: the combination of Python version, conda-pack and Triton server is not rigid, we can try any combination that works well together.
Hi, I am running into a similar issue. Did you find a fix?
@VirginieBfd I haven't heard anything back from the Triton team. Are you also running on SageMaker endpoints? I believe that the error is driven by creating the environment on a different OS than the one running the Triton container. No solid leads yet.
@Tabrizian we've found the root cause.
As advised by conda-pack, the OS used to create an environment must match the hosting OS. In addition, these environments must be created within the a container that matches the hosting container. We installed miniconda on top of the NGC Triton container, created a conda environment and packed it from within the container.
See this reply on a recent Triton issue, it was what unblocked us.
Description
All our attempts to create a conda environment, and leveraging it inside the triton serving container have failed. We're following the instructions for conda-pack on the Python backend snippet.
We're running a local test on a mac with no support to nvidia gpus. The eventual target is sagemaker instances.
This issue looks similar but didn't help.
Triton Information
What version of Triton are you using?
Are you using the Triton container or did you build it yourself?
Using 24.02-pyt-python-py3 with additional environment variables for SageMaker endpoints
To Reproduce
Steps to reproduce the behavior.
Describe the models (framework, inputs, outputs), ideally include the model configuration file (if using an ensemble include the model configuration file for that as well).
This is a simple Python backend, running models.py that just imports the needed packages. For example:
Expected behavior
A clear and concise description of what you expected to happen.
Creating a conda environment outside of our serving container, packaging it using conda-pack and leveraging it with Triton Serving Container.
Note: the combination of Python version, conda-pack and Triton server is not rigid, we can try any combination that works well together.
Traceback
Click me for a complete traceback
The text was updated successfully, but these errors were encountered: