We provide an NGINX Docker image based on OpenResty for the PWA deployment.
The Docker image can be built by running a Docker build with the Dockerfile
located in the nginx
folder.
Mandatory environment variables:
UPSTREAM_PWA
value in the form of http://<IP>:<PORT>
For HTTP, the server will run on default port 80.
For HTTPS, the server will run on default port 443.
We are using an OpenResty Docker image.
Therefore, we inherit all their configuration capabilities.
For further information, please refer to the official OpenResty Docker image page
You can switch on HTTPS for the NGINX container to execute a production-like setup locally, or for demo purposes by changing ENV SSL=0
to ENV SSL=1
and adjusting the port mapping in docker-compose.yml
.
No need to supply a certificate and a key.
They are automatically generated inside the running container.
The certificate is self-signed and will not work in your browser.
You have to confirm the security exception.
As developer convenience you can volume mount an internal folder to your host system to effectively trust the generated certificate.
Please check the NGINX logs for the following output.
You can now export the local CA by adjusting your docker-compose.yml /home/your-user/ca-dir:/root/.local/share/mkcert/rootCA.pem
For deploying to test environments that are not to be indexed by search bots or are not to be accessible by the public, the NGINX container can be set up with basic authentication.
To do so, supply a single user-password combination as environment variable, i.e., BASIC_AUTH=<user>:<password>
.
You can also whitelist IPs by supplying a YAML list to the environment variable BASIC_AUTH_IP_WHITELIST
:
nginx:
environment:
BASIC_AUTH: 'developer:!InterShop00!'
BASIC_AUTH_IP_WHITELIST: |
- 172.22.0.1
- 1.2.3.4
Entries of the IP whitelist are added to the NGINX config as allow
statements, which also supports IP ranges.
Please refer to the linked NGINX documentation on how to configure this.
After globally activating basic authentication for your setup, you can also disable it selectively per site.
See Multi-Site Configurations for examples on how to do that.
If the NGINX container is run without further configuration, the default Angular CLI environment properties are not overridden.
Multiple PWA channels can be set up by supplying a YAML configuration listing all domains the PWA should work for.
For more information on the multi-site syntax, refer to Multi-Site Configurations
The configuration can be supplied by setting the environment variable MULTI_CHANNEL
.
Alternatively, the source can be supplied by setting MULTI_CHANNEL_SOURCE
in any supported format by gomplate.
If no environment variables for multi-channel configuration are provided, the configuration will fall back to the content of nginx/multi-channel.yaml
, which can also be customized.
Warning
Multi-Channel configuration with context paths does not work in conjunction with service workers.
An extended list of examples can be found in the Multi-Site Configurations guide.
Often, NGINX receives requests from advertising networks or various user agents that append unused query parameters when making a request, for example, utm_source
.
These parameters can lead to inefficient caching, because even if the same URL is requested multiple times, the cached version will not be used if the URL is accessed with different query parameters.
To prevent this, you can define any number of blacklisted parameters that will be ignored by NGINX during caching.
As with multi-site handling above, the configuration can be supplied by setting the environment variable CACHING_IGNORE_PARAMS
.
Alternatively, the source can be supplied by setting CACHING_IGNORE_PARAMS_SOURCE
in any supported format by gomplate.
Be aware that the supplied list of parameters must be declared under a params
property.
If no environment variables for ignoring parameters are provided, the configuration will fall back to the content of nginx/caching-ignore-params.yaml
, which can also be customized.
Please refer to Concept - XML Sitemaps on how to configure ICM to generate PWA sitemap files where the sitemap XML name has to start with the prefix sitemap_
, e.g., sitemap_pwa
.
http://foo.com/en/sitemap_pwa.xml
To make above sitemap index file available under your deployment, you need to add the environment variable ICM_BASE_URL
to your NGINX container.
Let ICM_BASE_URL
point to your ICM backend installation, e.g., https://develop.icm.intershop.de
.
On a local development system you need to add it to docker-compose.yml
, e.g.,
nginx:
environment:
ICM_BASE_URL: 'https://develop.icm.intershop.de'
When the container is started it will process cache-ignore and multi-channel templates as well as sitemap proxy rules like this:
location /sitemap_ {
proxy_pass https://develop.icm.intershop.de/INTERSHOP/static/WFS/inSPIRED-inTRONICS-Site/rest/inSPIRED-inTRONICS/en_US/sitemaps/pwa/sitemap_;
}
The process will utilize your Multi-Site Configuration.
Be sure to include application
if you deviate from standard rest
application.
The Multi-Site Configuration has to include the correct channel value (e.g., inSPIRED-inTRONICS_Business-Site
and inSPIRED-inTRONICS-Site
instead of default
) because it is used to generate the correct sitemap URL path, e.g.,
nginx:
environment:
MULTI_CHANNEL: |
.+:
- baseHref: /en
channel: inSPIRED-inTRONICS_Business-Site
lang: en_US
- baseHref: /de
channel: inSPIRED-inTRONICS_Business-Site
lang: de_DE
- baseHref: /fr
channel: inSPIRED-inTRONICS_Business-Site
lang: fr_FR
- baseHref: /b2c
channel: inSPIRED-inTRONICS-Site
theme: b2c
The PWA can be configured with multiple identity providers.
In some use cases a specific identity provider must be selected when a certain route is requested.
For example, a punchout user should be logged in by the punchout identity provider requesting a punchout route.
For all other possible routes the default identity provider must be selected.
This can be done by setting the environment variable OVERRIDE_IDENTITY_PROVIDER
.
nginx:
environment:
OVERRIDE_IDENTITY_PROVIDERS: |
.+:
- path: /punchout
type: Punchout
This setting will generate rewrite rules for the URL paths for all given domains.
Alternatively, the source can be supplied by setting OVERRIDE_IDENTITY_PROVIDERS_SOURCE
in any supported format by gomplate.
If no environment variable is set, this feature is disabled.
Note
The alternative identity providers need to be configured for the SSR container via IDENTITY_PROVIDERS
to be used in the NGINX OVERRIDE_IDENTITY_PROVIDER
configuration.
Important
Overriding identity providers by path via NGINX configuration will only work with enabled SSR. SSR is enabled by default and must not be disabled via SSR: 0
.
Important
To configure additional headers, the PWA Helm Chart version 0.8.0 or above has to be used.
For some security or functional reasons it is necessary to add additional headers to page responses.
To make such headers configurable, the environment variable ADDITIONAL_HEADERS
has been introduced.
nginx:
environment:
ADDITIONAL_HEADERS: |
headers:
- header-a: 'value-a'
- header-b: 'value-b'
Alternatively, the source can be supplied by setting ADDITIONAL_HEADERS_SOURCE
in any supported format by gomplate.
For every entry NGINX will add this header to every possible response.
To make the additional headers available during build-time, the value for the environment variable ADDITIONAL_HEADERS
can be put into the additional-headers.yaml file.
Built-in features can be enabled and disabled:
SSR=off
effectively disables SSR rendering for browsers (default on
)CACHE=off
disables caching (default on
)COMPRESSION=off
disables compression (default on
)DEVICE_DETECTION=off
disables user-agent detection (default on
)PROMETHEUS=on
enables Prometheus metrics exports on port 9113
(default off
)SSL=on
to switch on HTTPS. See HTTPS or SSL above for further explanation.DEBUG=on
to log extra information like path matching.LOG_ALL=off
to restrict logging to errors.New features can be supplied in the folder nginx/features
.
A file named <feature>.conf
is included if the environment variable <feature>
is set to on
, 1
, true
or yes
(case insensitive).
Otherwise, the feature is disabled and an optional file <feature>-off.conf
is included in the configuration.
The feature name must only contain word characters (letters, numbers, and underscore).
If the cache feature is switched off, all caching for pre-rendered pages is disabled.
The cache duration for pre-rendered pages can be customized using CACHE_DURATION_NGINX_OK
(for successful responses) and CACHE_DURATION_NGINX_NF
(for 404 responses).
The value supplied must be in the time
format that is supported by NGINX proxy_cache_valid.
Each NGINX has its own cache, so in a deployment with multiple NGINX (for redundancy) the cache hit rate is significantly lower than it could be.
With the shared Redis cache the different NGINX instances push the cache to a shared Redis service and retrieve it from there.
This way each NGINX profits from already rendered SSR results and the overall performance of such a deployment increases.
Important
To configure the shared Redis cache, PWA Helm Chart version 0.8.0 or above has to be used.
Multiple NGINX instances can share the same Redis cache if this feature is activated.
To use the shared Redis cache, the environment variable REDIS_URI
must be provided with a valid Redis entrypoint.
The Redis URI has the following format: redis://USERNAME:PASSWORD@HOST:PORT/DB
To use a secure connection, use rediss://
instead of redis://
.
The parameters USERNAME
, PASSWORD
, and DB
are optional.
The current implementation supports only Redis deployments with a single entrypoint.
Redis Cluster setup is not supported because the redirect handling is not implemented.
Connecting to Redis Sentinel is also not supported.
For production setups, we recommend using a Redis cloud service.
The cache duration for pre-rendered pages can be customized using CACHE_DURATION_NGINX_OK
.
The value is transferred to srcache_default_expire.
Using CACHE_DURATION_NGINX_NF
is not supported with Redis cache.
404 pages are cached with the same duration as successful responses.
Because the cache is no longer embedded into the NGINX container, cache clearing has to be done separately.
Use the redis-cli
to send a flushdb
command to the Redis instance.
This can be done using Docker with the following command:
docker run --rm -it bitnami/redis redis-cli -u <REDIS_URI> flushdb
For development environments a local Redis can be started with the example docker-compose.yml
configuration.
redis:
image: bitnami/redis
container_name: redis
environment:
- ALLOW_EMPTY_PASSWORD=yes
---
nginx:
environment:
CACHE: 1
REDIS_URI: redis://redis:6379
Passing extra command-line flags to the Redis service for configuration (see Redis configuration) can be done via Docker command
.
redis:
...
command: /opt/bitnami/scripts/redis/run.sh --loglevel debug
An additional Redis service is not intended for production environments where a Redis cloud service should be used.
For that reason, also the PWA Helm chart does not support deploying an own Redis service with the PWA.
Only the redis.uri
can be configured for a Helm deployment.
All Brotli configuration directives can be set in
compression.conf
.
The on-the-fly Brotli compression level is set to the recommended value of 4
instead of 6
(default).
This setting provides a balance between the compression rate and CPU usage.
A higher level would achieve better compression but would also consume more CPU resources.
The two NGINX modules
ngx_http_brotli_filter_module.so
– for compressing responses on-the-flyngx_http_brotli_static_module.so
- for serving pre-compressed filesare built in the Dockerfile
using an NGINX archive in a version which matches the openresty Docker image version and are referenced in nginx.conf.tmpl
.
The archive needs to be used because it includes the configure
command.
The ./configure
arguments are taken from the current openresty configuration using nginx -V
(--add-module
arguments are excluded), see ngx_brotli
.
The modules can also be built using an openresty archive, but in this case the built process is taking longer:
wget
the openresty archive version which matches the Docker image versionmake modules
with make && make install
NGINX environment variables need to be configured in the cache.extraEnvVars
section of the PWA Helm Chart, e.g.,
cache:
extraEnvVars:
- name: NGINX_WORKER_PROCESSES
value: auto
parameter | format | default | comment |
---|---|---|---|
NGINX_WORKER_PROCESSES | string | auto |
overwrite the default worker_processes configuration |
NGINX_WORKER_CONNECTIONS | string | 1024 |
overwrite the default worker_connections configuration |