Skip to main content

Start RisingWave using Docker Compose

This topic describes how to start RisingWave using Docker Compose on a single machine. With this option, data is persisted in your preferred storage service, and observability is enabled for enhanced monitoring and analysis.

In this option, RisingWave functions as an all-in-one service. All components of RisingWave, including the compute node, frontend node, meta node, and compactor node, are put into a single process. They are executed in different threads, eliminating the need to start each component as a separate process.

However, please be aware that certain critical features, such as failover and resource management, are not implemented in this mode. Therefore, this option is not recommended for production deployments. For production deployments, please consider RisingWave Cloud, Kubernetes with Helm, or Kubernetes with Operator.

This option uses a pre-defined Docker Compose configuration file to set up a RisingWave cluster.

The cluster also incorporates these third-party components:

  • Grafana
  • Etcd
  • MinIO
  • Prometheus

Download the source file

As a prerequisite, you need to install Docker Desktop in your environment. Ensure that it is running before launching the cluster.

Then, clone the risingwave repository.

git clone https://github.com/risingwavelabs/risingwave.git

Open the repository in a terminal and run the following command to navigate to the docker directory.

cd docker

Start a RisingWave cluster

You can now start a RisingWave cluster. You can use the following storage options as the storage backend of RisingWave:

For each of the options, we have a Docker Compose configuration file that you can use after the necessary configurations.

MinIO

This is the default option. To start a standalone RisingWave cluster with MinIO as the storage backend, run the following command:

docker compose up -d
Command not found?

The default command-line syntax in Compose V2 starts with docker compose. See details in the Docker docs.

If you're using Compose V1, use docker-compose instead.

I'd like to start RisingWave components separately in a multi-node cluster.

You can also start a multi-node cluster where all components of RisingWave, including the compute node, frontend node, meta node, and compactor node, are started as separate processes.

By default, this mode uses MinIO as the storage backend of RisingWave.

To start a multi-node RisingWave cluster with MinIO as the storage backend, run the following command:

docker compose -f docker-compose-distributed.yml up

S3 or S3-compatible storage

To use S3 as the storage backend, configure your AWS credential information in /docker/aws.env.

To use S3-compatible storage options like Tencent Cloud COS, you need to configure the endpoint via the RW_S3_ENDPOINT parameter in /docker/aws.env. Don't include the bucket name in the endpoint.

In docker-compose-with-s3.yml, specify the bucket name via the hummock+s3 parameter.

- "hummock+s3://<bucket-name>"

Run this command to start the RisingWave cluster:

docker compose -f docker-compose-with-s3.yml up

Google Cloud Storage, Alibaba Cloud OSS, or Azure Blob Storage

Configure the credentials for the cloud service you want to use in /docker/multiple_object_storage.env.

In the corresponding docker-compose-with-service_name.yml file (for example, docker-compose-with-gcs.yml for Google Cloud Storage), specify the bucket name via the hummock+<service_name> parameter.

 - "hummock+<service_name>://<bucket-name>"

Run the following command to start the RisingWave cluster with one of the cloud storage services that you choose.

docker compose -f docker-compose-with-service_name.yml up

Remember to replace the docker-compose-with-service_name.yml with the full file name of the corresponding configuration file.

HDFS

Mount your HADOOP_HOME in the compactor node, computer node, and meta node volumes.

In /docker-compose-with-hdfs.yml, specify the cluster name via the hummock+hdfs parameter.

- "hummock+hdfs://<cluster_name>"

Run the following command to start a RisingWave cluster:

docker compose -f docker-compose-with-hdfs.yml up

Huawei Cloud OBS

To use Huawei Cloud OBS as the storage backend, you need to configure your OBS credential information in the multiple_object_storage.env file. Uncomment and set the following environment variables: OBS_ENDPOINT, OBS_ACCESS_KEY_ID, and OBS_SECRET_ACCESS_KEY. Don't include the bucket name in the endpoint.

In the docker-compose-with-obs.yml file, specify the bucket name via the hummock+obs parameter. Replace <bucket-name> with the name of your OBS bucket.

--state-store hummock+obs://<bucket-name>

To start the RisingWave cluster with OBS as the storage backend, run the following command:

docker compose -f docker-compose-with-obs.yml up

Connect to RisingWave

After RisingWave is up and running, you need to connect to it via the Postgres interactive terminal psql so that you can issue queries to RisingWave and see the query results. If you don't have psql installed, install psql first.

psql -h localhost -p 4566 -d dev -U root

Notes about the psql options:

  • The -h option is used to specify the hostname or IP address of the PostgreSQL server to connect to.
  • The -p option is used to specify the port number that the server is listening on.
  • The -d option is used to specify the name of the database to connect to.
  • The -U option is used to specify the name of the database user to connect as.
  • By default, the PostgreSQL server uses the root user account to authenticate connections to the dev database. Note that this user account does not require a password to connect.

Manage your RisingWave cluster

When the cluster is running, you can monitor the status of RisingWave and the additional components and make adjustments when necessary.

  • RisingWave Dashboard

    Access the RisingWave Dashboard at http://127.0.0.1:5691/. RisingWave Dashboard displays an overview of the cluster, as well as sources, sinks, tables, materialized views, and indexes available on the cluster.

  • Grafana

    Access Grafana at http://127.0.0.1:3001/ and search for risingwave_dashboard. In this dashboard, you can view the internal metrics such as node count, memory consumption, throughput, and latencies. You can use these metrics to troubleshoot and optimize the cluster performance.

  • MinIO

    Access the MinIO instance at http://127.0.0.1:9400/. Use the following credentials to log in.

    • User name: hummockadmin

    • Password: hummockadmin

  • Prometheus

    Access Prometheus at http://127.0.0.1:9500/. No credentials are needed. You can use Prometheus for real-time alerting.

Common issues

One of the common issues you may encounter is insufficient storage space. For example:

Error { code: "XMinioStorageFull", message: "Storage backend has reached its minimum free drive threshold. Please delete a few objects to proceed."

This issue typically occurs on macOS when using Docker Desktop. Docker Desktop runs within the macOS Hypervisor, where all the data, including logs, images, and volumes, is stored. The macOS Hypervisor has a default limit on disk capacity. If you encounter this error, you can resolve it by cleaning up unused containers or images. Another option is to increase the disk image size limit by following these steps: Click on the Docker Desktop icon in the menu bar, then go to Preferences > Resources > Advanced, and adjust the slider for disk image size to allocate more space for Docker images. If you are using a different platform, please ensure sufficient space is available on the local disk.

Help us make this doc better!

Was this page helpful?

Happy React is loading...