Currently, I want to deploy a Triton server to Vertex AI endpoint. However I received this error message.
"failed to start Vertex AI service: Invalid argument - Expect the model repository contains only a single model if default model is not specified"
Is this mean that the Triton server deploy only support one model? It is different from what I have read in this document about concurrent model execution
https://cloud.google.com/vertex-ai/docs/predictions/using-nvidia-triton
User | Count |
---|---|
2 | |
1 | |
1 | |
1 | |
1 |