Hosting LocalAI in cloud platforms such as AWS, Azure or GCP #160
Replies: 2 comments 3 replies
-
Hey, what is the performance on lambda? how long to you need to wait for the answer? |
Beta Was this translation helpful? Give feedback.
-
From my record, I deployed the Deploying I didn't test for AWS Fargate. Kindly share your experience here if you did. Thanks. |
Beta Was this translation helpful? Give feedback.
-
I hosted the gpt4all-j model at AWS Lambda Function successfully (as it allowed memory size up to 10GB, container image size up to 10GB and up to 15 minutes timeout via function url) using the Python binding at https://pypi.org/project/gpt4all-j/. For low traffic app, I think it make sense to host the API in the cloud platforms, don't you think so?
We can use the AWS lambda function for experimental or development purpose also (which is low traffic by nature), so that it is pay for the usage, not the idle compute time.
From the cost perspective, I think it is overkill to host the LocalAI APIs on a dedicated server. I would like to hear from you if you found more cost effective way to host the LocalAI APIs.
Lastly, could we host the LocalAI APIs as AWS Lambda Function?
I hope to hear from you soon. Appreciate your sharing.
Thank you.
Beta Was this translation helpful? Give feedback.
All reactions