You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: docs/dyn/container_v1.projects.locations.clusters.nodePools.html
+19-16Lines changed: 19 additions & 16 deletions
Original file line number
Diff line number
Diff line change
@@ -153,10 +153,10 @@ <h3>Method Details</h3>
153
153
"autoprovisioned": True or False, # Can this node pool be deleted automatically.
154
154
"enabled": True or False, # Is autoscaling enabled for this node pool.
155
155
"locationPolicy": "A String", # Location policy used when scaling up a nodepool.
156
-
"maxNodeCount": 42, # Maximum number of nodes for one location in the NodePool. Must be >= min_node_count. There has to be enough quota to scale up the cluster.
157
-
"minNodeCount": 42, # Minimum number of nodes for one location in the NodePool. Must be >= 1 and <= max_node_count.
158
-
"totalMaxNodeCount": 42, # Maximum number of nodes in the node pool. Must be greater than total_min_node_count. There has to be enough quota to scale up the cluster. The total_*_node_count fields are mutually exclusive with the *_node_count fields.
159
-
"totalMinNodeCount": 42, # Minimum number of nodes in the node pool. Must be greater than 1 less than total_max_node_count. The total_*_node_count fields are mutually exclusive with the *_node_count fields.
156
+
"maxNodeCount": 42, # Maximum number of nodes for one location in the node pool. Must be >= min_node_count. There has to be enough quota to scale up the cluster.
157
+
"minNodeCount": 42, # Minimum number of nodes for one location in the node pool. Must be greater than or equal to 0 and less than or equal to max_node_count.
158
+
"totalMaxNodeCount": 42, # Maximum number of nodes in the node pool. Must be greater than or equal to total_min_node_count. There has to be enough quota to scale up the cluster. The total_*_node_count fields are mutually exclusive with the *_node_count fields.
159
+
"totalMinNodeCount": 42, # Minimum number of nodes in the node pool. Must be greater than or equal to 0 and less than or equal to total_max_node_count. The total_*_node_count fields are mutually exclusive with the *_node_count fields.
160
160
},
161
161
"bestEffortProvisioning": { # Best effort provisioning. # Enable best effort provisioning for nodes
162
162
"enabled": True or False, # When this is enabled, cluster/node pool creations will ignore non-fatal errors like stockout to best provision as many nodes as possible right now and eventually bring up all target number of nodes
@@ -248,6 +248,7 @@ <h3>Method Details</h3>
248
248
"localSsdCount": 42, # Number of local NVMe SSDs to use. The limit for this value is dependent upon the maximum number of disk available on a machine per zone. See: https://cloud.google.com/compute/docs/disks/local-ssd for more information. A zero (or unset) value has different meanings depending on machine type being used: 1. For pre-Gen3 machines, which support flexible numbers of local ssds, zero (or unset) means to disable using local SSDs as ephemeral storage. 2. For Gen3 machines which dictate a specific number of local ssds, zero (or unset) means to use the default number of local ssds that goes with that machine type. For example, for a c3-standard-8-lssd machine, 2 local ssds would be provisioned. For c3-standard-8 (which doesn't support local ssds), 0 will be provisioned. See https://cloud.google.com/compute/docs/disks/local-ssd#choose_number_local_ssds for more info.
249
249
},
250
250
"localSsdCount": 42, # The number of local SSD disks to be attached to the node. The limit for this value is dependent upon the maximum number of disks available on a machine per zone. See: https://cloud.google.com/compute/docs/disks/local-ssd for more information.
251
+
"localSsdEncryptionMode": "A String", # Specifies which method should be used for encrypting the Local SSDs attahced to the node.
"variantConfig": { # LoggingVariantConfig specifies the behaviour of the logging component. # Logging variant configuration.
253
254
"variant": "A String", # Logging variant deployed on nodes.
@@ -575,10 +576,10 @@ <h3>Method Details</h3>
575
576
"autoprovisioned": True or False, # Can this node pool be deleted automatically.
576
577
"enabled": True or False, # Is autoscaling enabled for this node pool.
577
578
"locationPolicy": "A String", # Location policy used when scaling up a nodepool.
578
-
"maxNodeCount": 42, # Maximum number of nodes for one location in the NodePool. Must be >= min_node_count. There has to be enough quota to scale up the cluster.
579
-
"minNodeCount": 42, # Minimum number of nodes for one location in the NodePool. Must be >= 1 and <= max_node_count.
580
-
"totalMaxNodeCount": 42, # Maximum number of nodes in the node pool. Must be greater than total_min_node_count. There has to be enough quota to scale up the cluster. The total_*_node_count fields are mutually exclusive with the *_node_count fields.
581
-
"totalMinNodeCount": 42, # Minimum number of nodes in the node pool. Must be greater than 1 less than total_max_node_count. The total_*_node_count fields are mutually exclusive with the *_node_count fields.
579
+
"maxNodeCount": 42, # Maximum number of nodes for one location in the node pool. Must be >= min_node_count. There has to be enough quota to scale up the cluster.
580
+
"minNodeCount": 42, # Minimum number of nodes for one location in the node pool. Must be greater than or equal to 0 and less than or equal to max_node_count.
581
+
"totalMaxNodeCount": 42, # Maximum number of nodes in the node pool. Must be greater than or equal to total_min_node_count. There has to be enough quota to scale up the cluster. The total_*_node_count fields are mutually exclusive with the *_node_count fields.
582
+
"totalMinNodeCount": 42, # Minimum number of nodes in the node pool. Must be greater than or equal to 0 and less than or equal to total_max_node_count. The total_*_node_count fields are mutually exclusive with the *_node_count fields.
582
583
},
583
584
"bestEffortProvisioning": { # Best effort provisioning. # Enable best effort provisioning for nodes
584
585
"enabled": True or False, # When this is enabled, cluster/node pool creations will ignore non-fatal errors like stockout to best provision as many nodes as possible right now and eventually bring up all target number of nodes
@@ -670,6 +671,7 @@ <h3>Method Details</h3>
670
671
"localSsdCount": 42, # Number of local NVMe SSDs to use. The limit for this value is dependent upon the maximum number of disk available on a machine per zone. See: https://cloud.google.com/compute/docs/disks/local-ssd for more information. A zero (or unset) value has different meanings depending on machine type being used: 1. For pre-Gen3 machines, which support flexible numbers of local ssds, zero (or unset) means to disable using local SSDs as ephemeral storage. 2. For Gen3 machines which dictate a specific number of local ssds, zero (or unset) means to use the default number of local ssds that goes with that machine type. For example, for a c3-standard-8-lssd machine, 2 local ssds would be provisioned. For c3-standard-8 (which doesn't support local ssds), 0 will be provisioned. See https://cloud.google.com/compute/docs/disks/local-ssd#choose_number_local_ssds for more info.
671
672
},
672
673
"localSsdCount": 42, # The number of local SSD disks to be attached to the node. The limit for this value is dependent upon the maximum number of disks available on a machine per zone. See: https://cloud.google.com/compute/docs/disks/local-ssd for more information.
674
+
"localSsdEncryptionMode": "A String", # Specifies which method should be used for encrypting the Local SSDs attahced to the node.
"variantConfig": { # LoggingVariantConfig specifies the behaviour of the logging component. # Logging variant configuration.
675
677
"variant": "A String", # Logging variant deployed on nodes.
@@ -863,10 +865,10 @@ <h3>Method Details</h3>
863
865
"autoprovisioned": True or False, # Can this node pool be deleted automatically.
864
866
"enabled": True or False, # Is autoscaling enabled for this node pool.
865
867
"locationPolicy": "A String", # Location policy used when scaling up a nodepool.
866
-
"maxNodeCount": 42, # Maximum number of nodes for one location in the NodePool. Must be >= min_node_count. There has to be enough quota to scale up the cluster.
867
-
"minNodeCount": 42, # Minimum number of nodes for one location in the NodePool. Must be >= 1 and <= max_node_count.
868
-
"totalMaxNodeCount": 42, # Maximum number of nodes in the node pool. Must be greater than total_min_node_count. There has to be enough quota to scale up the cluster. The total_*_node_count fields are mutually exclusive with the *_node_count fields.
869
-
"totalMinNodeCount": 42, # Minimum number of nodes in the node pool. Must be greater than 1 less than total_max_node_count. The total_*_node_count fields are mutually exclusive with the *_node_count fields.
868
+
"maxNodeCount": 42, # Maximum number of nodes for one location in the node pool. Must be >= min_node_count. There has to be enough quota to scale up the cluster.
869
+
"minNodeCount": 42, # Minimum number of nodes for one location in the node pool. Must be greater than or equal to 0 and less than or equal to max_node_count.
870
+
"totalMaxNodeCount": 42, # Maximum number of nodes in the node pool. Must be greater than or equal to total_min_node_count. There has to be enough quota to scale up the cluster. The total_*_node_count fields are mutually exclusive with the *_node_count fields.
871
+
"totalMinNodeCount": 42, # Minimum number of nodes in the node pool. Must be greater than or equal to 0 and less than or equal to total_max_node_count. The total_*_node_count fields are mutually exclusive with the *_node_count fields.
870
872
},
871
873
"bestEffortProvisioning": { # Best effort provisioning. # Enable best effort provisioning for nodes
872
874
"enabled": True or False, # When this is enabled, cluster/node pool creations will ignore non-fatal errors like stockout to best provision as many nodes as possible right now and eventually bring up all target number of nodes
@@ -958,6 +960,7 @@ <h3>Method Details</h3>
958
960
"localSsdCount": 42, # Number of local NVMe SSDs to use. The limit for this value is dependent upon the maximum number of disk available on a machine per zone. See: https://cloud.google.com/compute/docs/disks/local-ssd for more information. A zero (or unset) value has different meanings depending on machine type being used: 1. For pre-Gen3 machines, which support flexible numbers of local ssds, zero (or unset) means to disable using local SSDs as ephemeral storage. 2. For Gen3 machines which dictate a specific number of local ssds, zero (or unset) means to use the default number of local ssds that goes with that machine type. For example, for a c3-standard-8-lssd machine, 2 local ssds would be provisioned. For c3-standard-8 (which doesn't support local ssds), 0 will be provisioned. See https://cloud.google.com/compute/docs/disks/local-ssd#choose_number_local_ssds for more info.
959
961
},
960
962
"localSsdCount": 42, # The number of local SSD disks to be attached to the node. The limit for this value is dependent upon the maximum number of disks available on a machine per zone. See: https://cloud.google.com/compute/docs/disks/local-ssd for more information.
963
+
"localSsdEncryptionMode": "A String", # Specifies which method should be used for encrypting the Local SSDs attahced to the node.
"variantConfig": { # LoggingVariantConfig specifies the behaviour of the logging component. # Logging variant configuration.
963
966
"variant": "A String", # Logging variant deployed on nodes.
@@ -1222,10 +1225,10 @@ <h3>Method Details</h3>
1222
1225
"autoprovisioned": True or False, # Can this node pool be deleted automatically.
1223
1226
"enabled": True or False, # Is autoscaling enabled for this node pool.
1224
1227
"locationPolicy": "A String", # Location policy used when scaling up a nodepool.
1225
-
"maxNodeCount": 42, # Maximum number of nodes for one location in the NodePool. Must be >= min_node_count. There has to be enough quota to scale up the cluster.
1226
-
"minNodeCount": 42, # Minimum number of nodes for one location in the NodePool. Must be >= 1 and <= max_node_count.
1227
-
"totalMaxNodeCount": 42, # Maximum number of nodes in the node pool. Must be greater than total_min_node_count. There has to be enough quota to scale up the cluster. The total_*_node_count fields are mutually exclusive with the *_node_count fields.
1228
-
"totalMinNodeCount": 42, # Minimum number of nodes in the node pool. Must be greater than 1 less than total_max_node_count. The total_*_node_count fields are mutually exclusive with the *_node_count fields.
1228
+
"maxNodeCount": 42, # Maximum number of nodes for one location in the node pool. Must be >= min_node_count. There has to be enough quota to scale up the cluster.
1229
+
"minNodeCount": 42, # Minimum number of nodes for one location in the node pool. Must be greater than or equal to 0 and less than or equal to max_node_count.
1230
+
"totalMaxNodeCount": 42, # Maximum number of nodes in the node pool. Must be greater than or equal to total_min_node_count. There has to be enough quota to scale up the cluster. The total_*_node_count fields are mutually exclusive with the *_node_count fields.
1231
+
"totalMinNodeCount": 42, # Minimum number of nodes in the node pool. Must be greater than or equal to 0 and less than or equal to total_max_node_count. The total_*_node_count fields are mutually exclusive with the *_node_count fields.
1229
1232
},
1230
1233
"clusterId": "A String", # Deprecated. The name of the cluster to upgrade. This field has been deprecated and replaced by the name field.
1231
1234
"name": "A String", # The name (project, location, cluster, node pool) of the node pool to set autoscaler settings. Specified in the format `projects/*/locations/*/clusters/*/nodePools/*`.
0 commit comments