mirror of
https://github.com/Mintplex-Labs/anything-llm
synced 2026-04-25 17:15:37 +02:00
update bare metal docs
This commit is contained in:
@@ -112,4 +112,36 @@ cd $HOME/anything-llm/collector
|
||||
(NODE_ENV=production node index.js) &> /logs/collector.log &
|
||||
```
|
||||
|
||||
## Using Nginx?
|
||||
|
||||
If you are using Nginx, you can use the following example configuration to proxy the requests to the server. Chats for streaming require **websocket** connections, so you need to ensure that the Nginx configuration is set up to support websockets. You can do this with a simple reverse proxy configuration.
|
||||
|
||||
```nginx
|
||||
server {
|
||||
# Enable websocket connections for agent protocol.
|
||||
location ~* ^/api/agent-invocation/(.*) {
|
||||
proxy_pass http://0.0.0.0:3001;
|
||||
proxy_http_version 1.1;
|
||||
proxy_set_header Upgrade $http_upgrade;
|
||||
proxy_set_header Connection "Upgrade";
|
||||
}
|
||||
|
||||
listen 80;
|
||||
server_name [insert FQDN here];
|
||||
location / {
|
||||
# Prevent timeouts on long-running requests.
|
||||
proxy_connect_timeout 605;
|
||||
proxy_send_timeout 605;
|
||||
proxy_read_timeout 605;
|
||||
send_timeout 605;
|
||||
keepalive_timeout 605;
|
||||
|
||||
# Enable readable HTTP Streaming for LLM streamed responses
|
||||
proxy_buffering off;
|
||||
proxy_cache off;
|
||||
|
||||
# Proxy your locally running service
|
||||
proxy_pass http://0.0.0.0:3001;
|
||||
}
|
||||
}
|
||||
```
|
||||
Reference in New Issue
Block a user