• Stars
    star
    648
  • Rank 69,480 (Top 2 %)
  • Language
    Shell
  • License
    GNU General Publi...
  • Created over 10 years ago
  • Updated 3 months ago

Reviews

There are no reviews yet. Be the first to send feedback to the community and the maintainers!

Repository Details

Dockerfile for postgis

Scenario Tests deploy-image

Table of Contents

docker-postgis

A simple docker container that runs PostGIS

Visit our page on the docker hub at: https://hub.docker.com/r/kartoza/postgis/

There are a number of other docker postgis containers out there. This one differentiates itself by:

  • Provides SSL support out of the box and enforces SSL client connections
  • Connections are restricted to the docker subnet
  • A default database gis is created for you so you can use this container 'out of the box' when it runs with e.g. QGIS
  • Streaming replication and logical replication support included (turned off by default)
  • Ability to create multiple database when starting the container.
  • Ability to create multiple schemas when starting the container.
  • Enable multiple extensions in the database when setting it up.
  • Gdal drivers automatically registered for pg raster.
  • Support for out-of-db rasters.

We will work to add more security features to this container in the future with the aim of making a PostGIS image that is ready to be used in a production environment (though probably not for heavy load databases).

There is a nice 'from scratch' tutorial on using this docker image on Alex Urquhart's blog here - if you are just getting started with docker, PostGIS and QGIS, we recommend that you read it and try out the instructions specified on the blog.

Tagged versions

The following convention is used for tagging the images we build:

kartoza/postgis:[POSTGRES_MAJOR_VERSION]-[POSTGIS_MAJOR_VERSION].[POSTGIS_MINOR_RELEASE]

So for example:

kartoza/postgis:14-3.1 Provides PostgreSQL 14.0, PostGIS 3.1

Note: We highly recommend that you use tagged versions because successive minor versions of PostgreSQL write their database clusters into different database directories - which will cause your database to appear to be empty if you are using persistent volumes for your database storage.

Getting the image

There are various ways to get the image onto your system:

The preferred way (but using most bandwidth for the initial image) is to get our docker trusted build like this,

docker pull kartoza/postgis:image_version

Building the image

Self build using Repository checkout

To build the image yourself do:

docker build -t kartoza/postgis git://github.com/kartoza/docker-postgis

Alternatively clone the repository and build against any preferred branch

git clone git://github.com/kartoza/docker-postgis
git checkout branch_name

Then do:

docker build -t kartoza/postgis .

Or build against a specific PostgreSQL version

docker build --build-arg POSTGRES_MAJOR_VERSION=13 --build-arg POSTGIS_MAJOR=3 -t kartoza/postgis:POSTGRES_MAJOR_VERSION .

Alternative base distributions builds

There are build args for DISTRO (=debian), IMAGE_VERSION (=buster) and IMAGE_VARIANT (=slim) which can be used to control the base image used (but it still needs to be Debian based and have PostgreSQL official apt repo).

For example making Ubuntu 20.04 based build (for better arm64 support) Edit the .env file to change the build arguments,

DISTRO=ubuntu 
IMAGE_VERSION=focal 
IMAGE_VARIANT="" 

Then run the script

./build.sh

Locales

By default, the image build will include all locales to cover any value for locale settings such as DEFAULT_COLLATION, DEFAULT_CTYPE or DEFAULT_ENCODING.

You can use the build argument: GENERATE_ALL_LOCALE=0

This will build with the default locate and speed up the build considerably.

Environment variables

Cluster Initializations

With a minimum setup, our image will use an initial cluster located in the DATADIR environment variable. If you want to use persistence, mount these locations into your volume/host. By default, DATADIR will point to /var/lib/postgresql/{major-version}. You can instead mount the parent location like this:

-v data-volume:/var/lib/postgresql

This default cluster will be initialized with default locale settings C.UTF-8. If, for instance, you want to create a new cluster with your own settings (not using the default cluster). You need to specify different empty directory, like this

-v data-volume:/opt/postgres/data \
-e DATADIR:/opt/postgres/data \
-e DEFAULT_ENCODING="UTF8" \
-e DEFAULT_COLLATION="id_ID.utf8" \
-e DEFAULT_CTYPE="id_ID.utf8" \
-e PASSWORD_AUTHENTICATION="md5" \
-e INITDB_EXTRA_ARGS="<some more initdb command args>" \
-v pgwal-volume:/opt/postgres/pg_wal \
-e POSTGRES_INITDB_WALDIR=/opt/postgres/pg_wal

The containers will use above parameters to initialize a new db cluster in the specified directory. If the directory is not empty, then the initialization parameter will be ignored.

These are some initialization parameters that will only be used to initialize a new cluster. If the container uses an existing cluster, it is ignored (for example, when the container restarts).

  • DEFAULT_ENCODING: cluster encoding
  • DEFAULT_COLLATION: cluster collation
  • DEFAULT_CTYPE: cluster ctype
  • WAL_SEGSIZE: WAL segsize option
  • PASSWORD_AUTHENTICATION : PASSWORD AUTHENTICATION
  • INITDB_EXTRA_ARGS: extra parameter that will be passed down to initdb command
  • POSTGRES_INITDB_WALDIR: parameter to tell Postgres about the initial waldir location. Note: You must always mount persistent volume to this location. Postgres will expect that the directory will always be available, even though it doesn't need the environment variable anymore. If you didn't persist this location, Postgres will not be able to find the pg_wal directory and consider the instance to be broken.

In addition to that, we have another parameter: RECREATE_DATADIR that can be used to force database re-initializations. If this parameter is specified as TRUE it will act as explicit consent to delete DATADIR and create new db cluster.

  • RECREATE_DATADIR: Force database re-initialization in the location DATADIR

If you used RECREATE_DATADIR and successfully created a new cluster. Remember that you should remove this parameter afterwards. Because, if it was not omitted, it will always recreate new db cluster after every container restarts.

Postgres Encoding

The database cluster is initialized with the following encoding settings

-E "UTF8" --lc-collate="en_US.UTF-8" --lc-ctype="en_US.UTF-8"

or

-E "UTF8" --lc-collate="C.UTF-8" --lc-ctype="C.UTF-8"

If you use default DATADIR location.

If you need to set up a database cluster with other encoding parameters you need to pass the environment variables when you initialize the cluster.

  • -e DEFAULT_ENCODING="UTF8"
  • -e DEFAULT_COLLATION="en_US.UTF-8"
  • -e DEFAULT_CTYPE="en_US.UTF-8"

Initializing a new cluster can be done by using different DATADIR location and mounting an empty volume. Or use parameter RECREATE_DATADIR to forcefully delete the current cluster and create a new one. Make sure to remove parameter RECREATE_DATADIR after creating the cluster.

See the postgres documentation about encoding for more information.

PostgreSQL extensions

The container ships with some default extensions i.e. postgis,hstore,postgis_topology,postgis_raster,pgrouting

You can use the environment variable POSTGRES_MULTIPLE_EXTENSIONS to activate a subset or multiple extensions i.e.

-e POSTGRES_MULTIPLE_EXTENSIONS=postgis,hstore,postgis_topology,postgis_raster,pgrouting`

Note: Some extensions require extra configurations to get them running properly otherwise they will cause the container to exit. Users should also consult documentation relating to that specific extension i.e. timescaledb, pg_cron, pgrouting

You can also install tagged version of extensions i.e

POSTGRES_MULTIPLE_EXTENSIONS=postgis,pgrouting:3.4.0

where pgrouting:3.4.0 The extension name is fixed with the version name with the delimiter being a colon.

Note In some cases, some versions of extensions might not be available for install. To enable them you can do the following inside the container:

wget --directory-prefix /usr/share/postgresql/15/extension/ https://raw.githubusercontent.com/postgres/postgres/master/contrib/hstore/hstore--1.1--1.2.sql

Then proceed to install it the normal way.

Shared preload libraries

Some PostgreSQL extensions require shared_preload_libraries to be specified in the conf files. Using the environment variable SHARED_PRELOAD_LIBRARIES you can pass comma separated values that correspond to the extensions defined using the environment variable POSTGRES_MULTIPLE_EXTENSIONS.

The default libraries that are loaded are pg_cron,timescaledb if the image is built with timescale support otherwise only pg_cron is loaded. You can pass the env variable,

  -e SHARED_PRELOAD_LIBRARIES='pg_cron,timescaledb'

Note You cannot pass the environment variable SHARED_PRELOAD_LIBRARIES without specifying the PostgreSQL extension that correspond to the SHARED_PRELOAD_LIBRARIES. This will cause the container to exit immediately.

Basic configuration

You can use the following environment variables to pass a username, password and/or default database name(or multiple databases comma separated).

  • -e POSTGRES_USER=<PGUSER>

  • -e POSTGRES_PASS=<PGPASSWORD>

    Note: You should use a strong passwords. If you are using docker-compose make sure docker can interpolate the password. Example using a password with a $ you will need to escape it ie $$

  • -e POSTGRES_DBNAME=<PGDBNAME>

  • -e SSL_CERT_FILE=/your/own/ssl_cert_file.pem

  • -e SSL_KEY_FILE=/your/own/ssl_key_file.key

  • -e SSL_CA_FILE=/your/own/ssl_ca_file.pem

  • -e DEFAULT_ENCODING="UTF8"

  • -e DEFAULT_COLLATION="en_US.UTF-8"

  • -e DEFAULT_CTYPE="en_US.UTF-8"

  • -e POSTGRES_TEMPLATE_EXTENSIONS=true

  • -e ACCEPT_TIMESCALE_TUNING=TRUE Useful to tune PostgreSQL conf based on timescaledb-tune. Defaults to FALSE.

  • -e TIMESCALE_TUNING_PARAMS Useful to configure none default settings to use when running ACCEPT_TIMESCALE_TUNING=TRUE. This defaults to empty so that we can use the default settings provided by the timescaledb-tune. Example,

    docker run -it --name timescale -e ACCEPT_TIMESCALE_TUNING=TRUE \
      -e POSTGRES_MULTIPLE_EXTENSIONS=postgis,hstore,postgis_topology,postgis_raster,pgrouting,timescaledb \
      -e TIMESCALE_TUNING_PARAMS="-cpus=4" kartoza/postgis:14-3.1

Note: ACCEPT_TIMESCALE_TUNING environment variable will overwrite all configurations based on the timescale configurations

Specifies whether extensions will also be installed in template1 database.

Schema Initialization

  • -e SCHEMA_NAME=<PGSCHEMA> You can pass a comma separated value of schema names which will be created when the database initializes. The default behavior is to create the schema in the first database specified in the environment variable POSTGRES_DBNAME. If you need to create matching schemas in all the databases that will be created you use the environment variable ALL_DATABASES=TRUE.

Configures archive mode

This image uses the initial PostgreSQL values which disables the archiving option by default. When ARCHIVE_MODE is changed to on, the archiving command will copy WAL files to /opt/archivedir

More info: 19.5. Write Ahead Log

  • -e ARCHIVE_MODE=off
  • -e ARCHIVE_COMMAND="test ! -f /opt/archivedir/%f && cp %p /opt/archivedir/%f" More info
  • -e ARCHIVE_CLEANUP_COMMAND="pg_archivecleanup /opt/archivedir %r"
  • -e RESTORE_COMMAND='cp /opt/archivedir/%f "%p"'

Configure WAL level

  • -e WAL_LEVEL=replica

    More info. Maximum size to let the WAL grow to between automatic WAL checkpoints.

  • -e WAL_SIZE=4GB

  • -e MIN_WAL_SIZE=2048MB

  • -e WAL_SEGSIZE=1024

  • -e MAINTAINANCE_WORK_MEM=128MB

Configure networking

You can open up the PG port by using the following environment variable. By default, the container will allow connections only from the docker private subnet.

  • -e ALLOW_IP_RANGE=<0.0.0.0/0> By default

Postgres conf is set up to listen to all connections and if a user needs to restrict which IP address PostgreSQL listens to you can define it with the following environment variable. The default is set to listen to all connections,

  • -e IP_LIST=<*>

Additional configuration

You can also define any other configuration to add to extra.conf, separated by '\n' e.g.:

  • -e EXTRA_CONF="log_destination = 'stderr'\nlogging_collector = on"

You can alternatively mount an extra config file into the setting's folder i.e

docker run --name "postgis" -v /data/extra.conf:/settings/extra.conf -p 25432:5432 -d -t kartoza/postgis

The /setting folder stores the extra configuration and is copied to the proper directory on runtime. The environment variable EXTRA_CONF_DIR controls the location of the mounted folder.

Then proceed to run the following:

 docker run --name "postgis" -e EXTRA_CONF_DIR=/etc/conf_settings -v /data:/etc/conf_settings -p 25432:5432 -d -t kartoza/postgis

If you want to reinitialize the data directory from scratch, you need to do:

  1. Do backup, move data, etc. Any preparations before deleting your data directory.
  2. Set environment variables RECREATE_DATADIR=TRUE. Restart the service
  3. The service will delete your DATADIR directory and start re-initializing your data directory from scratch.

Lockfile

During container startup, some lockfile are generated which prevent re-initialization of some settings. These lockfile are by default stored in the /settings folder, but a user can control where to store these files using the environment variable CONF_LOCKFILE_DIR Example

-e CONF_LOCKFILE_DIR=/opt/conf_lockfiles \
-v /data/lock_files:/opt/conf_lockfiles 
 -v /data/lock_files:/opt/conf_lockfiles 
-v /data/lock_files:/opt/conf_lockfiles 
 -v /data/lock_files:/opt/conf_lockfiles 
-v /data/lock_files:/opt/conf_lockfiles 

Note If you change the environment variable to point to another location when you restart the container the settings are reinitialized again.

Docker secrets

To avoid passing sensitive information in environment variables, _FILE can be appended to some of the variables to read from files present in the container. This is particularly useful in conjunction with Docker secrets, as passwords can be loaded from /run/secrets/<secret_name> e.g.:

  • -e POSTGRES_PASS_FILE=/run/secrets/<pg_pass_secret>

For more information see https://docs.docker.com/engine/swarm/secrets/.

Currently, POSTGRES_PASS, POSTGRES_USER, POSTGRES_DB, SSL_CERT_FILE, SSL_KEY_FILE, SSL_CA_FILE are supported.

Running the container

Rootless mode

You can run the container in rootless mode. This can be achieved by setting the env variable RUN_AS_ROOT=false. By default, this setting is set to true to allow the container to run as root for backward compatibility with older images.

With RUN_AS_ROOT=false you can additionally set the following environment variables to enable you to pass user id and group id into the container.

POSTGRES_UID=1000
POSTGRES_GID=1000
USER=postgresuser
GROUP_NAME=postgresusers

If you do not pass the UID and GID, the container will use the defaults specified in the container.

Using the terminal

To create a running container do:

docker run --name "postgis" -p 25432:5432 -d -t kartoza/postgis

Note: If you do not pass the env variable POSTGRES_PASS a random password will be generated and will be visible from the logs or within the container in /tmp/PGPASSWORD.txt.

Convenience docker-compose.yml

For convenience, we provide a docker-compose.yml that will run a copy of the database image and also our related database backup image (see https://github.com/kartoza/docker-pg-backup).

The docker-compose recipe will expose PostgreSQL on port 25432 (to prevent potential conflicts with any local database instance you may have),

Example usage:

docker-compose up -d

Note: The docker-compose recipe above will not persist your data on your local disk, only in a docker volume.

Connect via psql

Connect with psql (make sure you first install postgresql client tools on your host / client):

psql -h localhost -U docker -p 25432 -l

Note: Default postgresql user is 'docker'. If you do not pass the env variable POSTGRES_PASS a random strong password will be generated and can be accessed within the startup logs.

You can then go on to use any normal postgresql commands against the container.

Under ubuntu LTS the postgresql client can be installed like this:

sudo apt-get install postgresql-client-${POSTGRES_MAJOR_VERSION}

Where POSTGRES_MAJOR_VERSION corresponds to a specific PostgreSQL version i.e 12

Running SQL scripts on container startup.

In some instances users want to run some SQL scripts to populate the database. The environment variable POSTGRES_DB allows us to specify multiple database that can be created on startup. When running scripts they will only be executed against the first database ie POSTGRES_DB=gis,data,sample. The SQL script will be executed against the gis database. Additionally, a lock file is generated in /docker-entrypoint-initdb.d, which will prevent the scripts from getting executed after the first container startup. Provide IGNORE_INIT_HOOK_LOCKFILE=true to execute the scripts on every container start.

By default, the lockfile is generated in /docker-entrypoint-initdb.d but it can be overwritten by passing the environment variable SCRIPTS_LOCKFILE_DIR which can point to another location i.e

-e SCRIPTS_LOCKFILE_DIR=/data/ \
-v /data:/data

Currently, you can pass .sql, .sql.gz and .sh files as mounted volumes.

docker run -d -v `pwd`/setup-db.sql:/docker-entrypoint-initdb.d/setup-db.sql kartoza/postgis

Storing data on the host rather than the container.

Docker volumes can be used to persist your data.

mkdir -p ~/postgres_data
docker run -d -v $HOME/postgres_data:/var/lib/postgresql kartoza/postgis

You need to ensure the postgres_data directory has sufficient permissions for the docker process to read / write it.

Postgres SSL setup

There are three modalities in which you can work with SSL:

  1. Optional: using the shipped snakeoil certificates
  2. Forced SSL: forced using the shipped snakeoil certificates
  3. Forced SSL with Certificate Exchange: using SSL certificates signed by a certificate authority

By default, the image is delivered with an unsigned SSL certificate. This helps to have an encrypted connection to clients and avoid eavesdropping but does not help to mitigate Man In The Middle (MITM) attacks.

You need to provide your own, signed private key to avoid this kind of attacks (and make sure clients connect with verify-ca or verify-full sslmode).

Although SSL is enabled by default, connection to PostgreSQL with other clients i.e (PSQL or QGIS) still doesn't enforce SSL encryption. To force SSL connection between clients you need to use the environment variable,

FORCE_SSL=TRUE

The following example sets up a container with custom ssl private key and certificate:

docker run -p 25432:5432 -e FORCE_SSL=TRUE -e SSL_DIR="/etc/ssl_certificates" -e SSL_CERT_FILE='/etc/ssl_certificates/fullchain.pem' -e SSL_KEY_FILE='/etc/ssl_certificates/privkey.pem' -e SSL_CA_FILE='/etc/ssl_certificates/root.crt' -v /tmp/postgres/letsencrypt:/etc/ssl_certificates --name ssl -d kartoza/postgis:13-3.1

The environment variable SSL_DIR allows a user to specify the location where custom SSL certificates will be located. The environment variable currently defaults to SSL_DIR=/ssl_certificates

See the postgres documentation about SSL for more information.

Forced SSL: forced using the shipped snakeoil certificates

If you are using the default certificates provided by the image when connecting to the database you will need to set SSL Mode to any value besides verify-full or verify-ca.

The pg_hba.con will have entries like:

hostssl all all 0.0.0.0/0 scram-sha-256 clientcert=0

where PASSWORD_AUTHENTICATION=scram-sha-256 and ALLOW_IP_RANGE=0.0.0.0/0

Forced SSL with Certificate Exchange: using SSL certificates signed by a certificate authority

When setting up the database you need to define the following environment variables.

  • SSL_CERT_FILE
  • SSL_KEY_FILE
  • SSL_CA_FILE

Example:

docker run -p 5432:5432 -e FORCE_SSL=TRUE -e SSL_CERT_FILE='/ssl_certificates/fullchain.pem' -e SSL_KEY_FILE='/ssl_certificates/privkey.pem' -e SSL_CA_FILE='/ssl_certificates/root.crt' --name ssl -d kartoza/postgis:13-3.1

On the host machine where you need to connect to the database you also need to copy the SSL_CA_FILE file to the location /home/$user/.postgresql/root.crt or define an environment variable pointing to location of the SSL_CA_FILE example: PGSSLROOTCERT=/etc/letsencrypt/root.crt

The pg_hba.conf will have entries like:

hostssl all all 0.0.0.0/0 cert

where ALLOW_IP_RANGE=0.0.0.0/0

SSL connection inside the docker container using openssl certificates

Generate the certificates inside the container

CERT_DIR=/ssl_certificates
mkdir $CERT_DIR
openssl req -x509 -newkey rsa:4096 -keyout ${CERT_DIR}/privkey.pem -out \
      ${CERT_DIR}/fullchain.pem -days 3650 -nodes -sha256 -subj '/CN=localhost'

cp $CERT_DIR/fullchain.pem $CERT_DIR/root.crt
chmod -R 0700 ${CERT_DIR}
chown -R postgres ${CERT_DIR}

Set up your ssl config to point to the new location,

ssl = true
ssl_cert_file = '/ssl_certificates/fullchain.pem'
ssl_key_file = '/ssl_certificates/privkey.pem'
ssl_ca_file = '/ssl_certificates/root.crt' 

Then connect to the database using the psql command:

psql "dbname=gis port=5432 user=docker host=localhost sslmode=verify-full sslcert=/etc/letsencrypt/fullchain.pem sslkey=/etc/letsencrypt/privkey.pem sslrootcert=/etc/letsencrypt/root.crt"

Postgres Replication Setup

The image supports replication out of the box. By default, replication is turned off. The two main replication methods allowed are,

  • Streaming replication
  • Logical replication

Database permissions and password authentication

Replication uses a dedicated user REPLICATION_USER. The role ${REPLICATION_USER} uses the default group role pg_read_all_data. You can read more about this from the PostgreSQL documentation

Note: When setting up replication you need to specify the password using the environment variable REPLICATION_PASS. If you do not specify it a random strong password will be generated. This is visible in the startup logs as well as a text file within the container in /tmp/REPLPASSWORD.txt.

Streaming replication

Replication allows you to maintain two or more synchronized copies of a database, with a single master copy and one or more replicant copies. The animation below illustrates this - the layer with the red boundary is accessed from the master database and the layer with the green fill is accessed from the replicant database. When edits to the master layer are saved, they are automatically propagated to the replicant. Note also that the replicant is read-only.

docker run --name "streaming-replication" -e REPLICATION=true -e WAL_LEVEL='replica' -d -p 25432:5432 kartoza/postgis:14.3.2

Note If you do not pass the env variable REPLICATION_PASS a random password will be generated and will be visible from the logs or within the container in /tmp/REPLPASSWORD.txt

qgis

This image is provided with replication abilities. We can categorize an instance of the container as master or replicant. A master instance means that a particular container has a role as a single point of database write. A replicant instance means that a particular container will mirror database content from a designated master. This replication scheme allows us to sync databases. However, a replicant is only for read-only transaction, thus we can't write new data to it. The whole database cluster will be replicated.

Database permissions

Since we are using a role ${REPLICATION_USER}, we need to ensure that it has access to all the tables in a particular schema. So if a user adds another schema called data to the database gis he also has to update the permission for the user with the following SQL assuming the ${REPLICATION_USER} is called replicator,

ALTER DEFAULT PRIVILEGES IN SCHEMA data GRANT SELECT ON TABLES TO replicator;

Note You need to set up a strong password for replication otherwise the default password for ${REPLICATION_USER} will default to random generated string.

To experiment with the streaming replication abilities, you can see a docker-compose.yml. There are several environment variables that you can set, such as:

Master settings:

  • ALLOW_IP_RANGE: A pg_hba.conf domain format which will allow specified host(s) to connect into the container. This is needed to allow the slave to connect into master, so specifically these settings should allow slave address. It is also needed to allow clients on other hosts to connect to either the slave or the master.
  • REPLICATION_USER User to initiate streaming replication
  • REPLICATION_PASS Password for a user with streaming replication role

Slave settings:

  • REPLICATE_FROM: This should be the domain name or IP address of the master instance. It can be anything from the docker resolved name like that written in the sample, or the IP address of the actual machine where you expose master. This is useful to create cross machine replication, or cross stack/server.
  • REPLICATE_PORT: This should be the port number of master postgres instance. Will default to 5432 (default postgres port), if not specified.
  • DESTROY_DATABASE_ON_RESTART: Default is True. Set to 'False' to prevent this behavior. A replicant will always destroy its current database on restart, because it will try to sync again from master and avoid inconsistencies.
  • PROMOTE_MASTER: Default none. If set to any value then the current replicant will be promoted to master. In some cases when the master container has failed, we might want to use our replicant as master for a while. However, the promoted replicant will break consistencies and is not able to revert to replicant anymore, unless it is destroyed and re-synced with the new master.
  • REPLICATION_USER User to initiate streaming replication
  • REPLICATION_PASS Password for a user with streaming replication role

To run the example streaming_replication, follow these instructions:

Do a manual image build by executing the build.sh script

./build.sh

Go into the replication_examples/streaming_replication directory and experiment with the following Make command to run both master and slave services.

make up

To shut down services, execute:

make down

To view logs for master and slave respectively, use the following command:

make master-log
make node-log

You can try experiment with several scenarios to see how replication works

Sync changes from master to replicant

You can use any postgres database tools to create new tables in master, by connecting using POSTGRES_USER and POSTGRES_PASS credentials using exposed port. In the streaming_replication example, the master database was exposed on port 7777. Or you can do it via command line, by entering the shell:

make master-shell

Then make any database changes using psql.

After that, you can see that the replicant follows the changes by inspecting the slave database. You can, again, use database management tools using connection credentials, hostname, and ports for replicant. Or you can do it via command line, by entering the shell:

make node-shell

Then view your changes using psql.

Promoting replicant to master

You will notice that you cannot make changes in replicant, because it is read-only. If somehow you want to promote it to master, you can specify PROMOTE_MASTER: 'True' into slave environment and set DESTROY_DATABASE_ON_RESTART: 'False'.

After this, you can make changes to your replicant, but master and replicant will not be in sync anymore. This is useful if the replicant needs to take over a failover master. However, it is recommended to take additional action, such as creating a backup from the slave so a dedicated master can be created again.

Preventing replicant database destroy on restart

You can optionally set DESTROY_DATABASE_ON_RESTART: 'False' after successful sync to prevent the database from being destroyed on restart. With this setting you can shut down your replicant and restart it later, and it will continue to sync using the existing database (as long as there are no consistencies conflicts).

However, you should note that this option doesn't mean anything if you didn't persist your database volume. Because if it is not persisted, then it will be lost on restart because docker will recreate the container.

Logical replication

To activate the following you need to use the environment variable

WAL_LEVEL=logical to get a running instance like

docker run --name "logical-replication" -e WAL_LEVEL=logical -d  kartoza/postgis:13.0

For a detailed example see the docker-compose in the folder replication_examples/logical_replication.

Docker image versions

All instructions mentioned in the README are valid for the latest running image. Other docker images might have a few missing features than the ones in the latest image. We mainly do not back port changes to current stable images that are being used in production. However, if you feel that some changes included in the latest tagged version of the image are essential for the previous image you can cherry-pick the changes against that specific branch and we will test and merge.

Support

If you require more substantial assistance from kartoza (because our work and interaction on docker-postgis is pro bono), please consider taking out a Support Level Agreement.

Credits

April 2022

More Repositories

1

docker-geoserver

A docker recipe for GeoServer
Shell
610
star
2

docker-pg-backup

A cron job that will back up databases running in a docker postgres container
Shell
450
star
3

docker-osm

A docker compose project to setup an OSM PostGIS database with automatic updates from OSM periodically
C
285
star
4

docker-qgis-desktop

A docker project that will run your QGIS desktop app inside a docker container
Shell
96
star
5

docker-qgis-server

A dockerfile that contains a running QGIS server
QML
69
star
6

docker-mapproxy

A docker recipe for http://mapproxy.org/
Shell
62
star
7

QGIS-Legend-Patches

A collection of patches to use in your legends in QGIS!
Python
49
star
8

osgs

A rich, integrated, and opinionated GIS Stack with a focus on configurability and ease of use, built from Open Source Components
Makefile
40
star
9

leaflet-wms-legend

A simple leaflet WMS legend widget
JavaScript
37
star
10

docker-mapserver

Mapserver in Docker
HTML
37
star
11

prj.app

A django app for creating visual changelogs for software releases
Python
30
star
12

fbf-project

Project data and resources for WB Forecast Based Financing work
TSQL
23
star
13

docker-tilemill

A docker project for tilemill
Shell
21
star
14

docker-ssh

A simple docker container that runs ssh
Shell
20
star
15

kartoza-rancher-catalogue

This is our catalogue of Rancher recipes
19
star
16

osm-reporter

osm-reporter
JavaScript
19
star
17

awesome-geodata

A curated list of awesome geospatial data sources and services
19
star
18

maps.kartoza.com

Maps and web site for http://maps.kartoza.com
Python
18
star
19

docker-helpers

Helper commands for docker because typing long docker commands is boring...
Shell
15
star
20

django-bims

JavaScript
12
star
21

charts

Kartoza Helm charts for Kubernetes
Smarty
11
star
22

geocontext

A django app to retrieve context for spatial feature
Python
9
star
23

QGISTrainingWorkshop

This is a hands on workshop for QGIS, InaSAFE and OSM
QML
9
star
24

docker-qgis-orchestration

Orchestration scripts for QGIS docker services
9
star
25

docker-mezzanine

A docker project for mezzanine (http://mezzanine.jupo.org/)
JavaScript
9
star
26

geonode_qgis_server

Django app to make geonode work with QGIS Server
Python
9
star
27

SpeciesExplorer

A QGIS (3 or greater) plugin for quickly retrieving species occurrence data from GBIF.
Python
9
star
28

docker-django

A django ready docker container with nginx and uwsgi for serving up your site
Shell
8
star
29

docker-geogig

A docker container for using geogig (formerly known as geogit)
Shell
7
star
30

otf-project

Python
7
star
31

geosafe

InaSAFE package for Geonode
Python
7
star
32

docker-tao-web

Docker appliance for the tao examination platform (see http://taotesting.com/)
Shell
6
star
33

stream_feature_extractor

A QGIS plugin to extract stream features (wells, sinks, confluences etc.) from a stream network
QML
6
star
34

topostyle

publish and style South African 1:50000 topographical map vector data from NGI
QML
6
star
35

TheKartozaHandbook

Organisational Handbook and Technical Docs Repository for Kartoza. Here is where we highlight the procedures, principles and processes related to Development, DevOps, and GIS in line with the organisation's best practices
Nix
6
star
36

django-wms-client

A django application that provides generic WMS client support.
Python
5
star
37

docker-sftp-backup

Docker project to back up your data to an sftp server using rotating backup system
Python
5
star
38

docker-geosafe

Geosafe and GeoNode_QGIS-server deployment in docker
Python
5
star
39

flask_user_map

A simple flask for creating user community maps
JavaScript
5
star
40

IGRAC-GGIS

HTML
4
star
41

ckanext-dalrrd-emc-dcpr

Electronic Metadata Catalog for South Africa's Department of Agriculture, Land Reform and Rural Development
Python
4
star
42

QGISReportsWorkshop

Workshop presented to the Swiss QGIS User group March 2022
Python
4
star
43

SAEOSS-Portal

The SAEOSS Portal has been proposed as a system of software components functioning together as the national central earth observation geospatial repository, with a view to metadata and open geospatial API standards compliance as well as user impact maximization
CSS
4
star
44

parcel_plugin

QML
3
star
45

docker-geonode

A docker project for geonode
Python
3
star
46

docker-osm-examples

And directory for examples of docker-osm usage
QML
2
star
47

trail-mapper-ios

Mobile app for mapping trails and trail conditions
Swift
2
star
48

docker-django-base

A base docker image for django projects.
Shell
2
star
49

RandomDEMGenerator

An experiment to create random DEMS
Python
2
star
50

IGRAC-Mobile

TypeScript
2
star
51

sg-diagram-downloader

A tool for QGIS that will download SG (South African Surveyor General) diagrams
Python
2
star
52

docker-tilestream

A docker project for tilestream
Shell
2
star
53

geodata-mart

Geoprocessing as a Service
JavaScript
2
star
54

IGRAC-WellAndMonitoringDatabase

A PostgreSQL schema implementing the groundwater interchange schema defined in GWML2
JavaScript
2
star
55

django-training

Training couse materials for django
HTML
2
star
56

feti

Further Education and Training Institute Website
PLpgSQL
2
star
57

miniSASS

miniSASS website for GroundTruth
JavaScript
2
star
58

docker-tegola-maputnik

A demonstrator project showing how to run tegola vector tile server and maputnik tile style editor in docker.
Shell
2
star
59

qgis_dataset_qa_workbench

A QGIS3 plugin for assisting in dataset Quality Assurance workflows
Python
2
star
60

watchkeeper

A security alerts platform for iMMAP
Python
1
star
61

QGISCheatSheet

Our handy guide for QGIS!
1
star
62

trail-mapper-django

A django app for the Kartoza trail mapper app
CSS
1
star
63

LEDET_BIMS

Limpopo Biodiversity Information Management System
PLpgSQL
1
star
64

timesheet-project

Python
1
star
65

least-cost-electrification-somalia

Global Electrification Platform for Global
HTML
1
star
66

node-red-flows

Example flows using node red
1
star
67

Woolpert

Repo for the Woolpert project
JavaScript
1
star
68

rir-dashboard

An open platform for risk informed response planning.
JavaScript
1
star
69

geoserver-deploy

Example deployment of geoserver and postgis using docker and fig
1
star
70

kobo-docker-builder

Build script compilation for Kobotoolbox
Shell
1
star
71

utilities

miscelleneous scripts
1
star
72

clean-cooking-platform

JavaScript
1
star
73

internship-program

Internship program repository containing all the course materials and assignment submissions
1
star
74

qgis-server-monitor

An iOS app to monitor if QGIS servers are online
Swift
1
star
75

GeoContextQGISPlugin

A QGIS Plugin for the GeoContext API
Python
1
star
76

kartoza-docker-wordpress

The wordpress site powering kartoza.com, in docker
Makefile
1
star
77

ingc_geonode_theme

INGC Geonode Theme
CSS
1
star
78

docker-metacat

A docker project to set up KNB Metacat and dependencies
Shell
1
star
79

qgis-chat-plugin

A QGIS plugin for chatting on Gitter
Python
1
star
80

docker-qgis-base

A base repo upon which docker-qgis-desktop and docker-qgis-server are built
1
star
81

osm-cartography

Our collaborative work to provide an awesome QGIS cartography for OSM
QML
1
star
82

etherpad

Docker project to provide ether pad - an online document collaboration tool
1
star
83

osm-africa

An Africa project for tilemill derived from https://github.com/mapbox/osm-bright
1
star
84

freshwaterbiodiversity.org

HTML
1
star
85

django-react-base

A boilerplate project for quickly setting up a Django backend with a React frontend
Python
1
star
86

tilemaker-workflows

Internal repo for working with global OSM tiles and tilemaker
Lua
1
star
87

jakarta-flood-maps

A web based flood mapping application for jakarta
Python
1
star
88

InWARDS-Dashboard-Desktop

An application written in Electron.js for visualising stream flow conditions.
Vue
1
star