pulumi“错误删除自动启动启动配置”与Aws.eks.s.nodegroup
我在运行pulumi up
时会遇到以下错误。
aws:ec2:LaunchConfiguration (spot-ng-01-nodeLaunchConfiguration):
error: deleting urn:pulumi:staging::xx-api::eks:index:NodeGroup$aws:ec2/launchConfiguration:LaunchConfiguration::spot-ng-01-nodeLaunchConfiguration: 1 error occurred:
* error deleting Autoscaling Launch Configuration (spot-ng-01-nodeLaunchConfiguration-3a59b7e): ResourceInUse: Cannot delete launch configuration spot-ng-01-nodeLaunchConfiguration-3a59b7e because it is attached to AutoScalingGroup spot-ng-01-d1815eb6-NodeGroup-UBM7XABBGVNU
status code: 400, request id: fc55d507-0884-4c50-aeba-33831646a914
这是所讨论的资源,但没有更新代码。
new eks.NodeGroup("spot-ng-01", {
cluster: cluster,
spotPrice: "0.1",
instanceType: "t3.xlarge",
taints,
labels: { spot: "true" },
version: "1.21",
maxSize: 60,
minSize: 1,
nodeSubnetIds: options.vpc.privateSubnetIds,
instanceProfile: new aws.iam.InstanceProfile("spot-ng-profile-01", { role: role.name }),
nodeAssociatePublicIpAddress: false,
nodeSecurityGroup: clusterSG,
clusterIngressRule: cluster.eksClusterIngressRule,
autoScalingGroupTags: {
Name: "spot",
"k8s.io/cluster-autoscaler/enabled": "true",
[`k8s.io/cluster-autoscaler/${clusterName}`]: "true",
},
});
即使在运行pulumi刷新
之后,我仍然会出现错误。
I am getting the following error while running pulumi up
, I am getting a templateBody
update in the preview for aws:cloudformation:Stack spot-ng-01-nodes
.
aws:ec2:LaunchConfiguration (spot-ng-01-nodeLaunchConfiguration):
error: deleting urn:pulumi:staging::xx-api::eks:index:NodeGroup$aws:ec2/launchConfiguration:LaunchConfiguration::spot-ng-01-nodeLaunchConfiguration: 1 error occurred:
* error deleting Autoscaling Launch Configuration (spot-ng-01-nodeLaunchConfiguration-3a59b7e): ResourceInUse: Cannot delete launch configuration spot-ng-01-nodeLaunchConfiguration-3a59b7e because it is attached to AutoScalingGroup spot-ng-01-d1815eb6-NodeGroup-UBM7XABBGVNU
status code: 400, request id: fc55d507-0884-4c50-aeba-33831646a914
This is the resource in question, but the code was not updated.
new eks.NodeGroup("spot-ng-01", {
cluster: cluster,
spotPrice: "0.1",
instanceType: "t3.xlarge",
taints,
labels: { spot: "true" },
version: "1.21",
maxSize: 60,
minSize: 1,
nodeSubnetIds: options.vpc.privateSubnetIds,
instanceProfile: new aws.iam.InstanceProfile("spot-ng-profile-01", { role: role.name }),
nodeAssociatePublicIpAddress: false,
nodeSecurityGroup: clusterSG,
clusterIngressRule: cluster.eksClusterIngressRule,
autoScalingGroupTags: {
Name: "spot",
"k8s.io/cluster-autoscaler/enabled": "true",
[`k8s.io/cluster-autoscaler/${clusterName}`]: "true",
},
});
Even after running pulumi refresh
, I still get the error.
如果你对这篇内容有疑问,欢迎到本站社区发帖提问 参与讨论,获取更多帮助,或者扫码二维码加入 Web 技术交流群。

绑定邮箱获取回复消息
由于您还没有绑定你的真实邮箱,如果其他用户或者作者回复了您的评论,将不能在第一时间通知您!
发布评论
评论(1)
解决方案需要手动干预,它可能不是最好的,但它解决了问题。
Pulumi创建了另一个启动仪,我制作了有关自动分解组使用的新启动仪。然后我运行
pulumi up
,它能够删除卡住的启动configuration。然后运行pulumi刷新
。The solution required manual intervention, it might not be the best but it solved the issue.
Another LaunchConfiguration was created by pulumi, I made this new LaunchConfiguration used by the AutoscalingGroup in question. then I ran
pulumi up
and it was able to delete the LaunchConfiguration that was stuck. Then ranpulumi refresh
.