(4/8) Deploy NextCloud on Kuberbetes The self-hosted Dropbox
This article is part of the series Build your very own self-hosting platform with Raspberry Pi and Kubernetes
- Install Raspbian Operating-System and prepare the system for Kubernetes
- Install and configure a Kubernetes cluster with k3s to self-host applications
- Deploy NextCloud on Kuberbetes: The self-hosted Dropbox
- Self-host your Media Center On Kubernetes with Plex, Sonarr, Radarr, Transmission and Jackett
- Self-host Pi-Hole on Kubernetes and block ads and trackers at the network level
- Self-host your password manager with Bitwarden
- Deploy Prometheus and Grafana to monitor a Kubernetes cluster
Now we have prepared our RaspberryPi cluster to receive Kubernetes as a self-hosting platform, it's time to start installing applications !
NextCloud is an file hosting open-source software similar to Dropbox. Unlike Dropbox, NextCloud is not available as a SaaS but only on-premise which means anyone is allowed to install and operate it on their own private server. NextCloud offers individuals and organisations to gain control over their private data with a safe and secure solutions. NextCloud also provides a long list of add-ons working alongside the file sharing solution such as: Calendar & Contacts management, Audio/Video conferencing, Task Management, Photos albums and more.
In this article we will learn how to safely install NextCloud on a Kubernetes environment and configure both Desktop and Mobile access from anywhere.
In order to run entirely the tutorial, we will need:
- A running Kubernetes cluster (see previous articles if you haven't set this up yet)
- A domain name in order to access our NextCloud instance from outside our network. (replace
by your domain)
- Have a external static IP (usually the case by default)
- Access to your router admin console to port-forward an incoming request to our Kubernetes Ingress service.
We are going to isolate all the Kubernetes objects related to NextCloud in the namespace
To create a namespace, run the following command:
$ kubectl create namespace nextcloud
The first step consists in setting up a volume to store our NextCloud data (files and database). If you followed the previous articles to install and configure a self-hosting platform using RaspberryPi and Kubernetes, you remember we have on each worker a NFS client pointing to a SSD on
1. Deploy the Persistent Volume (PV)
The Persistent Volume specify the name, the size, the location and the access modes of the volume:
- The name of the PV is
- The size allocated is 50GB
- The location is
- The access is ReadWriteOnce
Create the following file and apply it to the k8 cluster.
## nextcloud.persistentvolume.yml --- apiVersion: v1 kind: PersistentVolume metadata: name: "nextcloud-ssd" labels: type: "local" spec: storageClassName: "manual" capacity: storage: "50Gi" accessModes: - ReadWriteOnce hostPath: path: "/mnt/ssd/nextcloud" ---
$ kubectl apply -f nextcloud.persistentvolume.yml persistentvolume/nextcloud-ssd created
You can verify the PV exists with the following command:
$ kubectl get pv NAME CAPACITY ACCESS MODES RECLAIM POLICY STATUS CLAIM STORAGECLASS REASON AGE nextcloud-ssd 50Gi RWO Retain Available manual 34s
2. Create the Persistent Volume Claim (PVC)
The Persistent Volume Claim is used to map a Persistent Volume to a deployment or stateful set. Unlike the PV, the PVC belongs to a namespace.
Create the following file and apply it to the k8 cluster.
## nextcloud.persistentvolumeclaim.yml --- apiVersion: v1 kind: PersistentVolumeClaim metadata: namespace: "nextcloud" name: "nextcloud-ssd" spec: storageClassName: "manual" accessModes: - ReadWriteOnce resources: requests: storage: "50Gi" ---
$ kubectl apply -f nextcloud.persistentvolumeclaim.yml persistentvolumeclaim/nextcloud-ssd created
You can verify the PVC exists with the following command:
$ kubectl get pvc -n nextcloud NAME STATUS VOLUME CAPACITY ACCESS MODES STORAGECLASS AGE nextcloud-ssd Bound nextcloud-ssd 50Gi RWO manual 26s
In the next part, we are now going to deploy NextCloud using the stable/nextcloud Helm chart.
1. Download the Chart values of the chart locally
Run the following command to download the Chart values into the local file
$ helm show values stable/nextcloud >> nextcloud.values.yml
If you open the file, you will see the default configuration values to setup NextCloud. Instead of using the flag
--set property=value like before, we will use the file
nextcloud.values.yml to make all the changes.
2. Update the values
We now need to update a few properties before installing the Helm chart. Open the file
nextcloud.values.yml and change the following properties (replace the information surrounded by
## nextcloud.values.yml nextcloud: host: "nextcloud.<domain.com>" # Host to reach NextCloud username: "admin" # Admin password: "<PASSWORD>" # Admin Password (...) persistence: enabled: true # Change to true existingClaim: "nextcloud-ssd" # Persistent Volume Claim created earlier accessMode: ReadWriteOnce size: "50Gi"
Take a look at the file if you want to make more customisation to NextCloud:
- Configure emails
- Configure an external database or deploy a MariaDB as part of the Chart (improve performance)
- Configure Redis server (improve performance)
3. Install the Chart
In the part, we will install the Helm chart under the namespace
nextcloud.values.yml as configuration file.
$ helm install nextcloud stable/nextcloud \ --namespace nextcloud \ --values nextcloud.values.yml
After a couple of minutes, check if the pod and service is up and running:
$ kubectl get pods -n nextcloud NAME READY STATUS RESTARTS AGE IP NODE NOMINATED NODE READINESS GATES nextcloud-78f5564f89-854jr 1/1 Running 0 15m 10.42.0.16 kube-master <none> <none>
$ kubectl get services -n nextcloud -o wide NAME TYPE CLUSTER-IP EXTERNAL-IP PORT(S) AGE SELECTOR nextcloud ClusterIP 10.43.188.121 <none> 8080/TCP 51m app.kubernetes.io/name=nextcloud
You can check the logs with the following command:
$ kubectl logs -f nextcloud-78f5564f89-854jr -n nextcloud
Otherwise check the folder
This step is configured before the ingress in order to be able to issue a certificate automatically when we deploy the ingress
The next part consist to enable the connections to NextCloud from outside so you can access your data from anywhere.
1. Port Forwarding
First you need to go to your router setup and add a port-forwarding rule to map any incoming requests on port 80 or port 443 to be forwarded to
192.168.0.240 (the LoadBalancer IP of the Nginx).
VirginHub - Port-Forwarding
2. Map the subdomain
nextcloud.<domain.com> to your home router
First you need to find out what's your router external IP, run this command or go to whatismyip.com.
$ curl ipecho.net/plain x.x.x.x
Then, we need to configure our subdomain to make sure
nextcloud.<domain.com> resolves to our external static IP. Go to your domain provider console / DNS management add a record:
- Type: A
- Name: nextcloud (subdomain)
- Value: x.x.x.x (external satic IP)
At this point, the application (pod
nextcloud-78f5564f89-854jr) is only accessible within the cluster on port 8080. To make it accessible from outside the cluster (on our network), we need to deploy an Ingress mapping service:port to a route of the Nginx proxy.
Because, this route will also be exposed over the Internet, we will also issue a certificate to encrypt the traffic with SSL.
1. Create the ingress config file
Create the file
## nextcloud.ingress.yml --- apiVersion: extensions/v1beta1 kind: Ingress metadata: namespace: "nextcloud" # Same namespace as the deployment name: "nextcloud-ingress" # Name of the ingress (see kubectl get ingress -A) annotations: kubernetes.io/ingress.class: "nginx" cert-manager.io/cluster-issuer: "letsencrypt-prod" # Encrypt using the ClusterIssuer deployed while setting up Cert-Manager nginx.ingress.kubernetes.io/proxy-body-size: "50m" # Increase the size of the maximum allowed size of the client request body spec: tls: - hosts: - "nextcloud.<domain.com>" # Host to access nextcloud secretName: "nextcloud-prod-tls" # Name of the certifciate (see kubectl get certificate -A) rules: - host: "nextcloud.<domain.com>" # Host to access nextcloud http: paths: - path: / # We will access NextCloud via the URL https://nextcloud.<domain.com>/ backend: serviceName: "nextcloud" # Mapping to the service (see kubectl get services -n nextcloud) servicePort: 8080 # Mapping to the port (see kubectl get services -n nextcloud) ---
2. Deploy the Ingress
Now deploy the ingress:
$ kubectl apply -f nextcloud.ingress.yml ingress.extensions/nextcloud-ingress created
3. Check the certificate issuance
After you deployed the ingress, a certificate should be issued, check the
certificate (it might take a couple of minutes to be READY):
$ kubectl get certificaterequest -n nextcloud -o wide NAMESPACE NAME READY ISSUER STATUS AGE nextcloud nextcloud-prod-tls-9929727 True letsencrypt-prod Certificate fetched from issuer successfully 17m $ kubectl get certificate -n nextcloud -o wide NAMESPACE NAME READY SECRET ISSUER STATUS AGE nextcloud nextcloud-prod-tls True nextcloud-prod-tls letsencrypt-prod Certificate is up to date and has not expired 17m
Alright, still with me :) You can now try to access your NextCloud instance using your browser, mobile or the Android/iOS app from home or outside via https://nextcloud.
Connect with the user admin and the password configured in the file
- Kauri original title: (4/8) Deploy NextCloud on Kuberbetes The self-hosted Dropbox
- Kauri original link: https://kauri.io/48-deploy-nextcloud-on-kuberbetes-the-selfhosted-d/f958350b22794419b09fc34c7284b02e/a
- Kauri original author: Grégoire Jeanmart (@gregjeanmart)
- Kauri original Publication date: 2020-03-31
- Kauri original tags: self-hosting, nextcloud, kubernetes, dropbox, open-source, file-hosting, helm
- Kauri original hash: QmaKwvWjA1eTYcnE4o1gwTFvpWkegaYZZ8JRzJhzXjEyiK
- Kauri original checkpoint: unknown