Skip to main content

google_container_regional_node_pool resource


A google_container_regional_node_pool is used to test a Google RegionalNodePool resource


describe google_container_regional_node_pool(project: 'chef-gcp-inspec', location: 'europe-west2', cluster: 'inspec-gcp-regional-cluster', name: 'inspec-gcp-regional-node-pool') do
  it { should exist }
  its('initial_node_count') { should eq '1'}

describe google_container_regional_node_pool(project: 'chef-gcp-inspec', location: 'europe-west2', cluster: 'inspec-gcp-regional-cluster', name: 'nonexistent') do
  it { should_not exist }


Properties that can be accessed from the google_container_regional_node_pool resource:

The name of the node pool.
The node configuration of the pool.
The name of a Google Compute Engine machine type (e.g. n1-standard-1). If unspecified, the default machine type is n1-standard-1.
Size of the disk attached to each node, specified in GB. The smallest allowed disk size is 10GB. If unspecified, the default disk size is 100GB.
The set of Google API scopes to be made available on all of the node VMs under the “default” service account. The following scopes are recommended, but not required, and by default are not included: is required for mounting persistent storage on your nodes. is required for communicating with (the Google Container Registry). If unspecified, no scopes are added, unless Cloud Logging or Cloud Monitoring are enabled, in which case their required scopes will be added.
The Google Cloud Platform Service Account to be used by the node VMs. If no Service Account is specified, the “default” service account is used.
The metadata key/value pairs assigned to instances in the cluster. Keys must conform to the regexp [a-zA-Z0-9-_]+ and be less than 128 bytes in length. These are reflected as part of a URL in the metadata server. Additionally, to avoid ambiguity, keys must not conflict with any other metadata keys for the project or be one of the four reserved keys: “instance-template”, “kube-env”, “startup-script”, and “user-data” Values are free-form strings, and only have meaning as interpreted by the image running in the instance. The only restriction placed on them is that each value’s size must be less than or equal to 32 KB. The total size of all keys and values must be less than 512 KB. An object containing a list of “key”: value pairs. Example: { “name”: “wrench”, “mass”: “1.3kg”, “count”: “3” }.
The image type to use for this node. Note that for a given image type, the latest version of it will be used.
The map of Kubernetes labels (key/value pairs) to be applied to each node. These will added in addition to any default label(s) that Kubernetes may apply to the node. In case of conflict in label keys, the applied set may differ depending on the Kubernetes version – it’s best to assume the behavior is undefined and conflicts should be avoided. For more information, including usage and the valid values, see: An object containing a list of “key”: value pairs. Example: { “name”: “wrench”, “mass”: “1.3kg”, “count”: “3” }.
The number of local SSD disks to be attached to the node. The limit for this value is dependant upon the maximum number of disks available on a machine per zone. See: for more information.
The list of instance tags applied to all nodes. Tags are used to identify valid sources or targets for network firewalls and are specified by the client during cluster or node pool creation. Each tag within the list must comply with RFC1035.
Whether the nodes are created as preemptible VM instances. See: for more information about preemptible VM instances.
A list of hardware accelerators to be attached to each node
The number of the accelerator cards exposed to an instance.
The accelerator type resource name
Type of the disk attached to each node (e.g. ‘pd-standard’ or ‘pd-ssd’) If unspecified, the default disk type is ‘pd-standard’
Minimum CPU platform to be used by this instance. The instance may be scheduled on the specified or newer CPU platform
List of kubernetes taints to be applied to each node.
Key for taint
Value for taint
Effect for taint
The initial node count for the pool. You must ensure that your Compute Engine resource quota is sufficient for this number of instances. You must also have available firewall and routes quota.
Status of nodes in this pool instance
Additional information about the current status of this node pool instance
The version of the Kubernetes of this node.
Autoscaler configuration for this NodePool. Autoscaler is enabled only if a valid configuration is present.
Is autoscaling enabled for this node pool.
Minimum number of nodes in the NodePool. Must be >= 1 and <= maxNodeCount.
Maximum number of nodes in the NodePool. Must be >= minNodeCount. There has to enough quota to scale up the cluster.
Management configuration for this NodePool.
A flag that specifies whether node auto-upgrade is enabled for the node pool. If enabled, node auto-upgrade helps keep the nodes in your node pool up to date with the latest release version of Kubernetes.
A flag that specifies whether the node auto-repair is enabled for the node pool. If enabled, the nodes in this node pool will be monitored and, if they fail health checks too many times, an automatic repair action will be triggered.
Specifies the Auto Upgrade knobs for the node pool.
This field is set when upgrades are about to commence with the approximate start time for the upgrades, in RFC3339 text format.
This field is set when upgrades are about to commence with the description of the upgrade.
The constraint on the maximum number of pods that can be run simultaneously on a node in the node pool.

max_pods_per_node : Constraint enforced on the max num of pods per node.

Which conditions caused the current node pool state.
Machine-friendly representation of the condition
The pod CIDR block size per node in this node pool.
The cluster this node pool belongs to.
The location where the node pool is deployed

GCP Permissions

Ensure the Kubernetes Engine API is enabled for the current project.