Skip to content

Commit 7567331

Browse files
author
Bill Prin
committed
hegemonic review
1 parent f2c39df commit 7567331

File tree

1 file changed

+3
-4
lines changed

1 file changed

+3
-4
lines changed

bigtable/metricscaler/metricscaler.py

Lines changed: 3 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -27,8 +27,7 @@
2727
The minimum number of nodes to use. The default minimum is 3. If you have a
2828
lot of data, the rule of thumb is to not go below 2.5 TB per node for SSD
2929
clusters, and 8 TB for HDD. The bigtable.googleapis.com/disk/bytes_used
30-
metric is useful in figuring out the minimum number.
31-
of nodes.
30+
metric is useful in figuring out the minimum number of nodes.
3231
"""
3332

3433
_MAX_NODE_COUNT = 30
@@ -85,15 +84,15 @@ def scale_bigtable(bigtable_instance, bigtable_cluster, scale_up):
8584
new_node_count = min(current_node_count + 3, _MAX_NODE_COUNT)
8685
cluster.serve_nodes = new_node_count
8786
cluster.update()
88-
print('Scaled up from {} up to {} nodes.'.format(
87+
print('Scaled up from {} to {} nodes.'.format(
8988
current_node_count, new_node_count))
9089
else:
9190
if current_node_count > _MIN_NODE_COUNT:
9291
new_node_count = max(
9392
current_node_count - _SIZE_CHANGE_STEP, _MIN_NODE_COUNT)
9493
cluster.serve_nodes = new_node_count
9594
cluster.update()
96-
print('Scaled down from {} up to {} nodes.'.format(
95+
print('Scaled down from {} to {} nodes.'.format(
9796
current_node_count, new_node_count))
9897
# [END bigtable_scale]
9998

0 commit comments

Comments
 (0)