Deckhouse Platform in Google Cloud

Configure cluster

This template is used for system apps domains within the cluster, e.g., Grafana for %s.example.com will be available as grafana.example.com.
This prefix is used for names of cluster objects created by Deckhouse (virtual machines, networks, security policies, etc.).
This key is passed to the cloud provider during the virtual machine creation process.

Select layout

Layout — the way how resources are located in the cloud. There are several pre-defined layouts.

resources

  • A dedicated VPC is created for the cluster. All cluster nodes have public IP addresses.
  • Peering can be configured between the cluster VPC and other VPCs.

resources

  • A separate VPC with Cloud NAT is created for the cluster.
  • Nodes in the cluster do not have public IP addresses.
  • Public IP addresses can be allocated to master and static nodes.
    • In this case, one-to-one NAT is used to translate the public IP address to the node’s IP address (note that CloudNAT is not used in such a case).
  • If the master does not have a public IP, then an additional instance with a public IP (aka bastion host) is required for installation tasks and accessing the cluster.
  • Peering can also be configured between the cluster VPC and other VPCs.

Select preset

Preset — the structure of nodes in the cluster. There are several pre-defined presets.

  • The cluster consists of one master node and one worker node.
  • Kubernetes Control Plane and Deckhouse controller run on the master node.
  • Deckhouse deploys other components (Ingress Controller, Prometheus, cert-manager, etc.) on the worker node.
  • Your applications should run on the worker node.
  • Highly Available Kubernetes Control Plane.
  • The cluster consists of three master nodes and two worker nodes.
  • Kubernetes Control Plane and Deckhouse controller run on master nodes.
  • Deckhouse deploys other components (Ingress Controller, Prometheus, cert-manager, etc.) on the worker nodes.
  • Your applications should run on the worker nodes.
  • Highly Available Kubernetes Control Plane.
  • The cluster consists of three master nodes, two system nodes, several frontend nodes, and one worker node.
  • Kubernetes Control Plane and Deckhouse controller run on master nodes.
  • Deckhouse deploys system components (Prometheus, cert-manager, etc.) on system nodes.
  • Deckhouse deploys Ingress Controller on frontend nodes. The number of frontend nodes depends on the number of availability zones in a cloud provider.
  • Your applications should run on the worker node.