1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
#![forbid(unsafe_code)]
use anyhow::{anyhow, bail, format_err, Result};
use diem_logger::{info, warn};
use rusoto_autoscaling::{
AutoScalingGroupNamesType, Autoscaling, AutoscalingClient, SetDesiredCapacityType,
};
use rusoto_core::Region;
use rusoto_sts::WebIdentityProvider;
#[allow(clippy::collapsible_if)]
pub async fn set_asg_size(
desired_capacity: i64,
buffer_percent: f64,
asg_name: &str,
wait_for_completion: bool,
scaling_down: bool,
) -> Result<()> {
let buffer = if scaling_down {
0
} else {
((desired_capacity as f64 * buffer_percent) / 100_f64).ceil() as i64
};
info!(
"Scaling to desired_capacity : {}, buffer: {}, asg_name: {}",
desired_capacity, buffer, asg_name
);
let set_desired_capacity_type = SetDesiredCapacityType {
auto_scaling_group_name: asg_name.to_string(),
desired_capacity: desired_capacity + buffer,
honor_cooldown: Some(false),
};
let credentials_provider = WebIdentityProvider::from_k8s_env();
let dispatcher = rusoto_core::HttpClient::new()
.map_err(|e| anyhow!("Failed to create request dispatcher, met Error:{}", e))?;
let asc = AutoscalingClient::new_with(dispatcher, credentials_provider, Region::UsWest2);
diem_retrier::retry_async(diem_retrier::fixed_retry_strategy(10_000, 60), || {
let asc = asc.clone();
let set_desired_capacity_type = set_desired_capacity_type.clone();
Box::pin(async move {
asc.set_desired_capacity(set_desired_capacity_type)
.await
.map_err(|e| {
warn!("set_desired_capacity failed: {}, retrying", e);
format_err!("set_desired_capacity failed: {}", e)
})
})
})
.await?;
if !wait_for_completion {
return Ok(());
}
diem_retrier::retry_async(diem_retrier::fixed_retry_strategy(10_000, 60), || {
let asc_clone = asc.clone();
Box::pin(async move {
let mut total = 0;
let mut current_token = None;
loop {
let current_token_clone = current_token.clone();
let auto_scaling_group_names_type = AutoScalingGroupNamesType {
auto_scaling_group_names: Some(vec![asg_name.to_string()]),
max_records: Some(100),
next_token: current_token_clone,
};
let asgs = asc_clone
.describe_auto_scaling_groups(auto_scaling_group_names_type)
.await?;
if asgs.auto_scaling_groups.is_empty() {
bail!("asgs.auto_scaling_groups.is_empty()");
}
let asg = &asgs.auto_scaling_groups[0];
if scaling_down {
total += asg
.instances
.clone()
.ok_or_else(|| format_err!("instances not found for auto_scaling_group"))?
.len() as i64;
} else {
total += asg
.instances
.clone()
.ok_or_else(|| format_err!("instances not found for auto_scaling_group"))?
.iter()
.filter(|instance| instance.lifecycle_state == "InService")
.count() as i64;
}
if asgs.next_token.is_none() {
break;
}
current_token = asgs.next_token;
}
info!(
"Waiting for scaling to complete. Current size: {}, Min Desired Size: {}",
total, desired_capacity
);
if scaling_down {
if total > desired_capacity {
bail!(
"Waiting for scale-down to complete. Current size: {}, Min Desired Size: {}",
total,
desired_capacity
);
} else {
info!("Scale down completed");
Ok(())
}
} else if total < desired_capacity {
bail!(
"Waiting for scale-up to complete. Current size: {}, Min Desired Size: {}",
total,
desired_capacity
);
} else {
info!("Scale up completed");
Ok(())
}
})
})
.await
}