Weβre overhauling Dgraphβs docs to make them clearer and more approachable. If
you notice any issues during this transition or have suggestions, please
let us know.
There might be times when youβll want to set up a load balancer to accomplish
goals such as increasing the utilization of your database by sending queries
from the app to multiple database server replicas. You can follow these steps to
get started with that.
Setting up NGINX load balancer using Docker Compose
Download ZIP
Download the contents of this gistβs ZIP file and extract it to a directory
called graph-nginx
, as follows:
mkdir dgraph-nginx
cd dgraph-nginx
wget -O dgraph-nginx.zip https://gist.github.com/danielmai/0cf7647b27c7626ad8944c4245a9981e/archive/5a2f1a49ca2f77bc39981749e4783e3443eb3ad9.zip
unzip -j dgraph-nginx.zip
Two files will be created: docker-compose.yml
and nginx.conf
.
Start Dgraph cluster
Start a 6-node Dgraph cluster (3 Dgraph Zero, 3 Dgraph Alpha, replication
setting 3) by starting the Docker Compose config:
Setting up NGINX load balancer with Dgraph running directly on the host machine
You can start your Dgraph cluster directly on the host machine (for example,
with systemd) as follows:
Install NGINX using the following apt-get
command
After you have set up your Dgraph cluster, install the latest stable NGINX. On
Debian and Ubuntu systems use the following command:
Make sure that your Dgraph cluster is up and running (it this case we will refer
to a 6 node cluster). After installing NGINX, you can configure it for load
balancing. You do this by specifying which types of connections to listen to,
and where to redirect them. Create a new configuration file called
load-balancer.conf
:
sudo vim /etc/nginx/conf.d/load-balancer.conf
and edit it to read as follows:
upstream alpha_grpc {
server alpha1:9080;
server alpha2:9080;
server alpha3:9080;
}
upstream alpha_http {
server alpha1:8080;
server alpha2:8080;
server alpha3:8080;
}
# $upstream_addr is the ip:port of the Dgraph Alpha defined in the upstream
# Example: 172.25.0.2, 172.25.0.7, 172.25.0.5 are the IP addresses of alpha1, alpha2, and alpha3
# /var/log/nginx/access.log will contain these logs showing "localhost to <upstream address>"
# for the different backends. By default, NGINX load balancing is round robin.
log_format upstreamlog '[$time_local] $remote_addr - $remote_user - $server_name $host to: $upstream_addr: $request $status upstream_response_time $upstream_response_time msec $msec request_time $request_time';
server {
listen 9080 http2;
access_log /var/log/nginx/access.log upstreamlog;
location / {
grpc_pass grpc://alpha_grpc;
}
}
server {
listen 8080;
access_log /var/log/nginx/access.log upstreamlog;
location / {
proxy_pass http://alpha_http;
}
}
Next, disable the default server configuration; on Debian and Ubuntu systems
youβll need to remove the default symbolic link from the sites-enabled
folder.
rm /etc/nginx/sites-enabled/default
Now you can restart nginx
:
In a different shell, run the dgraph increment
(docs) tool against the NGINX gRPC load balancer
(nginx:9080
):
docker-compose exec alpha1 dgraph increment --alpha nginx:9080 --num=10
If you have Dgraph installed on your host machine, then you can also run this
from the host:
dgraph increment --alpha localhost:9080 --num=10
The increment tool uses the Dgraph Go client to establish a gRPC connection
against the --alpha
flag and transactionally increments a counter predicate
--num
times.
Check logs
In the NGINX access logs (in the docker-compose
up shell window), or if you
are not using Docker Compose you can tail logs from /var/log/nginx/access.log
.
Youβll see access logs like the following:
With gRPC load balancing, each request can hit a different Alpha node. This
can increase read throughput.
nginx_1 | [15/Jan/2020:03:12:02 +0000] 172.20.0.9 - - - nginx to: 172.20.0.7:9080: POST /api.Dgraph/Query HTTP/2.0 200 upstream_response_time 0.008 msec 1579057922.135 request_time 0.009
nginx_1 | [15/Jan/2020:03:12:02 +0000] 172.20.0.9 - - - nginx to: 172.20.0.2:9080: POST /api.Dgraph/Query HTTP/2.0 200 upstream_response_time 0.012 msec 1579057922.149 request_time 0.013
nginx_1 | [15/Jan/2020:03:12:02 +0000] 172.20.0.9 - - - nginx to: 172.20.0.5:9080: POST /api.Dgraph/Query HTTP/2.0 200 upstream_response_time 0.008 msec 1579057922.162 request_time 0.012
nginx_1 | [15/Jan/2020:03:12:02 +0000] 172.20.0.9 - - - nginx to: 172.20.0.7:9080: POST /api.Dgraph/Query HTTP/2.0 200 upstream_response_time 0.012 msec 1579057922.176 request_time 0.013
nginx_1 | [15/Jan/2020:03:12:02 +0000] 172.20.0.9 - - - nginx to: 172.20.0.2:9080: POST /api.Dgraph/Query HTTP/2.0 200 upstream_response_time 0.012 msec 1579057922.188 request_time 0.011
nginx_1 | [15/Jan/2020:03:12:02 +0000] 172.20.0.9 - - - nginx to: 172.20.0.5:9080: POST /api.Dgraph/Query HTTP/2.0 200 upstream_response_time 0.016 msec 1579057922.202 request_time 0.013
These logs show that traffic is being load balanced to the following upstream
addresses defined in alpha_grpc in nginx.conf:
nginx to: 172.20.0.7
nginx to: 172.20.0.2
nginx to: 172.20.0.5
Load balancing methods
By default, NGINX load balancing is done round-robin. By the way There are other
load-balancing methods available such as least connections or IP hashing. To use
a different method than round-robin, specify the desired load-balancing method
in the upstream section of load-balancer.conf
.
# use least connection method
upstream alpha_grpc {
least_conn;
server alpha1:9080;
server alpha2:9080;
server alpha3:9080;
}
upstream alpha_http {
least_conn;
server alpha1:8080;
server alpha2:8080;
server alpha3:8080;
}
Responses are generated using AI and may contain mistakes.