I have deployed llama model in us-east using Vertex AI. The model took 30 minutes to deploy. Status is Active but when i sending JSON on Test your Model screen, it gives me Internal Server Error.
I have tried multiple deployments bu same error everytime
Please help!!
The "Internal Server Error" you're encountering when testing your deployed Llama model in Vertex AI suggests that there might be an issue with the deployment or the request you're sending to the model.
Make sure that your Llama model is correctly deployed on Vertex AI, and the deployment configuration is set up correctly, including specifying the correct runtime version, hardware configuration, and other relevant settings. Also, ensure that the JSON request format is correct. The request you provided seems fine, but it's crucial to double-check that it matches the expected input format for your Llama model.
Try sending a simpler request without the "max_length" and "top_k" parameters to see if the model can handle basic input without errors. This can help isolate the issue.
"Internal Server Error" is a generic error message that could result from various issues, including infrastructure problems, configuration errors, or issues on the service side. Therefore, detailed logs and Google Cloud support can be valuable resources for identifying and resolving the problem.
User | Count |
---|---|
2 | |
1 | |
1 | |
1 | |
1 |