[Question]: Configuration of the VM instance to host librechat in the cloud for a corporate setting #5541
Replies: 4 comments 3 replies
-
Hi, thanks for your question. A GPU would only be useful if running local models in the same server. Try at least 4 GB of RAM and 2 vCPU for that many users. |
Beta Was this translation helpful? Give feedback.
-
Hi Danny, Thanks for the quick reply!
} |
Beta Was this translation helpful? Give feedback.
-
Hi, The file upload issue worked when I introduced timeout-related changes in the nginx.conf file. Just the client_max_body_size attribute was not sufficient. The following configuration changes helped resolve the issue: nginx
} |
Beta Was this translation helpful? Give feedback.
-
Hi Danny, I request your help in verifying if my understanding of the flow/architecture of the application correct with a self hosted RAG and embeddings model:- LibreChat Architecture Flow1. File Upload and Initial Handling
2. File Processing and Embedding (Offline-ish, but triggered by upload)
3. User Interaction (Real-time - Chat)
4. RAG Pipeline (if triggered)
5. LLM Interaction
6. Final Response
|
Beta Was this translation helpful? Give feedback.
-
What is your question?
Hi Community,
Thank you for the fantastic contribution. I am looking forward to your insights on what VM settings/configurations should be used to host the Libre Chat in a VM for a corporate setting for up to 500 users. The idea is also to host the RAG within the VM itself. Will a GPU too be useful? What configurations etc.?
Thank you!
More Details
Any pointers on this would be appreciated.
What is the main subject of your question?
No response
Screenshots
No response
Code of Conduct
Beta Was this translation helpful? Give feedback.
All reactions