Hi,
I've created a page with a webhook exists as a cloud function. Through the cloud function, I'm sending the user prompt to the fine-tuned model in vertex ai and return the response to the dialog flow cx. But the latency of returning the response to the dialog flow cx is more. What are the ways to reduce the time taken to get the solution from the llm?
Thanks in advance..,
User | Count |
---|---|
2 | |
1 | |
1 | |
1 | |
1 |