# Caching and Clustering
The tutorial below covers the basic steps to setup and validate a clustered instance of the Entando App Engine.
NOTE
When building your deployment architecture, it is important to review your goals, hardware, networking and application-specific setup. You must also optimize App Engine deployment for your environment. The configurations and tests below do not address every type of application or deployment, but they can be used as building blocks to create a deployment architecture that works for your application.
# Storage Requirements for Clustered Entando Apps
In order to scale an Entando Application across multiple nodes, you must provide a storage class that supports
a ReadWriteMany
access policy. There are many ways to accomplish this, including using dedicated storage providers like GlusterFS. Cloud Kubernetes providers also offer clustered storage options specific to their implementation, such as Google Cloud File in GKE or Azure Files in AKS.
If your default class doesn't support ReadWriteMany
, add the following operator properties to the ConfigMap of your Helm template to use two different storage classes for clustered vs non-clustered storage:
entando.k8s.operator.default.clustered.storage.class: [your clustered RWX storage class]
entando.k8s.operator.default.non.clustered.storage.class: [your RWO storage class]
Set both values to the appropriate storage class for your configuration.
TIP
A ReadWriteOnce (RWO)
policy that ensures instances are all scheduled to the same node will scale an Entando Application without clustered storage and is accomplished via taints on other nodes. Be aware of the pros and cons of scheduling instances to the same node so you can maximize utilization of node resources and recover from an unreachable application instance. If the node terminates or is shutdown, your application will be down while Kubernetes reschedules the pods to a different node.
# Clustering
This tutorial reviews setting up a clustered Entando App Engine in entando-de-app
. The goal is to deploy a clustered instance of the App Engine and verify the high availability and scalable deployment of the application.
# Prerequisites
- An existing deployment of an Entando App or the ability to create one.
- If you haven't created a deployment or don't have a YAML file for an Entando deployment, follow the Quickstart instructions.
- The Entando deployment must use a Relational Database Management System (RDBMS) to organize data in a table structure. Clustered instances will not work correctly with in-memory databases.
# Creating a Clustered App Instance
- Create an Entando deployment via the Helm template or edit an existing deployment YAML file.
- Scale your Entando server application:
kubectl scale deployment quickstart-server-deployment -n entando --replicas=2
- Run
kubectl get pods -n [your namespace]
to view the pods in your deployment. - You should have two
quickstart-server-deployment
pods in your namespace. - The logs of the
quickstart-server-deployment
in either pod contains logging information related to different instances joining the cluster and how data is balanced between instances. See the screenshot below for an example (actual logs will vary).
# Validating Clustered Instances
This is an advanced exercise and not required or recommended for most deployment scenarios. The steps below validate that the clustered instances are working in your environment and that you have created a high availability deployment.
- Create a clustered instance or have an existing clustered Entando App instance available for testing.
- Get the URL for your
entando-de-app
withkubectl get ingress -n [your namespace]
. - Open the URL in the browser of your choice and ensure that the application is working.
- Open a new browser window in incognito or private browsing mode to ensure that no data is cached and you're receiving a copy of the running application. Do not navigate to the app. In the next steps, you'll delete a pod in your cluster and verify that your application is still getting served. Kubernetes will automatically restore the desired number of replicas, so you'll need to perform the validation test inside of approximately 1 minute (environments vary) before the new replica is launched.
- Delete one of the server deployment pods in your clustered instance with
kubectl delete [your-pod-name] -n [your namespace]
.- There are other ways to do this. You could also shell into the server-container and manually kill the running app process with
kill -9 357
. - If you want to test at the hardware level, you could manually terminate a node in your cluster (ensuring that the pods are scheduled to different nodes).
- There are other ways to do this. You could also shell into the server-container and manually kill the running app process with
- In your private/incognito browser window, open the URL to your
entando-de-app
. - Check that the application continues to render while the pod you deleted is no longer present.
- Wait for Kubernetes to restore your deleted pod.
- Check that the application continues to render after the pod is restored.
# Caching Validation
Validating the shared cache can be done in a similar process to the clustered instance validation. The high level steps are:
- Deploy a clustered instance (see creating a clustered instance tutorial).
- Create data with the App Builder (pages, page templates, content etc.), using the external route for the application.
- Refer to the logs to note which instance processed the request.
- Terminate that instance.
- Fetch the recently created data and verify that the data are returned.
# Configuring and Deploying with Redis
In this section, an Entando App Engine instance is deployed using Redis as a cache for data served by the App Engine. For more information on the cache configuration for the App Engine, see the Caching and Clustering documentation.
# Deploy Redis to Kubernetes
- Create the Redis deployment and expose the endpoints:
kubectl create deployment redis --image=redis:6
kubectl expose deployment redis --port=6379 --target-port=6379 -n [your namespace]
- Install the Redis CLI for your environment per https://redis.io/topics/rediscli (opens new window).
- Get the IP for your Redis deployment:
kubectl get service -n [your namespace]
- Validate your deployment:
redis-cli -h 10.43.99.198 -p 6379 ping
- Should respond PONG.
redis-cli -h 10.43.99.198 -p 6379 incr mycounter
- Should increment each time.
# Configure implementation
- Use git to clone the
entando-de-app
repository:
git clone https://github.com/entando/entando-de-app
- Fetch the tags, checkout the release tag and create a branch for your customization:
git fetch --tags
git checkout tags/v6.3.68 -b 6.3.2-redis
Open the pom.xml file of the
entando-de-app
.Add the Redis caching plugin to the pom.xml:
<dependency>
<groupId>org.entando.entando.plugins</groupId>
<artifactId>entando-plugin-jpredis</artifactId>
<type>war</type>
</dependency>
- Save the pom.
- Build and push a custom image for the
entando-de-app
with Building a Docker Image Tutorial. - Download the operator configuration deployment file:
curl -L -C - -O https://raw.githubusercontent.com/entando/entando-releases/v6.5.4/dist/ge-1-1-6/namespace-scoped-deployment/orig/namespace-resources.yaml > namespace-resources.yaml
- Update the deployment YAML file to point to your custom
entando-de-app
image with Redis. The line to change is in theConfigMap
and is noted below:
entando-de-app-wildfly: >-
{"version":"6.3.10","executable-type":"jvm","registry":"docker.io","organization":"entando"}
- Deploy your edited file with
kubectl
. For example:
kubectl apply -f namespace-resources.yaml
- Run the Helm template to generate the application configuration file
my-clustered-app.yaml
:
helm template quickstart ./ > my-clustered-app.yaml
- Add environment variables for your deployed Redis instance to
EntandoApp
inmy-clustered-app.yaml
. The variables to create are:
REDIS_ADDRESS
REDIS_PASSWORD
REDIS_ACTIVE
The following example pertains to the EntandoApp only, not a complete deployment. Utilize this as a reference to create your configuration in a complete deployment.
NOTE: This example uses a Secret for the
REDIS_PASSWORD
, which is recommended. You can also hardcode the password in the YAML for testing purposes, but the use of clear text passwords in deployment files is not recommended. Create and use a Secret for the password as a best practice.
apiVersion: "entando.org/v1"
kind: "EntandoApp"
metadata:
name: "quickstart"
annotations:
entando.org/processing-instruction: defer
spec:
dbms: embedded
replicas: 1
standardServerImage: wildfly
ingressPath: /entando-de-app
environmentVariables:
- name: REDIS_ADDRESS
value: [your redis URI. For example redis://localhost:6379]
- name: REDIS_PASSWORD
valueFrom:
secretKeyRef:
key: password
name: quickstart1-redis-secret
optional: false
- name: REDIS_ACTIVE
value: "true"
- Deploy your file:
kubectl apply -f [your-clustered-app.yaml]