Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

AKS Fails to build due to disabled feature: Microsoft.ContainerService/AutoUpgradePreview #2584

Closed
sossickd opened this issue Oct 5, 2021 · 9 comments

Comments

@sossickd
Copy link

sossickd commented Oct 5, 2021

Since the release of the latest azurerm terraform provider 2.79.x the following error appears when building a new AKS cluster.

Plan: 1 to add, 0 to change, 0 to destroy.
azurerm_kubernetes_cluster.main: Creating...

Error: creating Managed Kubernetes Cluster "Example-AKS" (Resource Group "kubernetes-prod"): containerservice.ManagedClustersClient#CreateOrUpdate: Failure sending request: StatusCode=0 -- Original Error: Code="BadRequest" Message="Feature Microsoft.ContainerService/AutoUpgradePreview is not enabled. Please see https://aka.ms/aks/previews for how to enable features."

Downgrading the azurerm provider to 2.78.0 fixes the issue.

The following MR caused the issue on the azurerm terraform provider.

@tombuildsstuff mentioned in the related issue on the azurerm terraform provider github project.

"Whilst this is a change in Terraform, unfortunately this is an AKS bug - since submitting the default value to the API shouldn't require this preview feature to be enabled (as it's marked as the default value in the API, too). We've raised this with the AKS Team on Friday, but I'd suggest opening an issue on the AKS Repository too where someone from that team should be able to take a look."

@ghost ghost added the triage label Oct 5, 2021
@ghost
Copy link

ghost commented Oct 5, 2021

Hi sossickd, AKS bot here 👋
Thank you for posting on the AKS Repo, I'll do my best to get a kind human from the AKS team to assist you.

I might be just a bot, but I'm told my suggestions are normally quite good, as such:

  1. If this case is urgent, please open a Support Request so that our 24/7 support team may help you faster.
  2. Please abide by the AKS repo Guidelines and Code of Conduct.
  3. If you're having an issue, could it be described on the AKS Troubleshooting guides or AKS Diagnostics?
  4. Make sure your subscribed to the AKS Release Notes to keep up to date with all that's new on AKS.
  5. Make sure there isn't a duplicate of this issue already reported. If there is, feel free to close this one and '+1' the existing issue.
  6. If you have a question, do take a look at our AKS FAQ. We place the most common ones there!

@ghost
Copy link

ghost commented Oct 7, 2021

Triage required from @Azure/aks-pm

@palma21
Copy link
Member

palma21 commented Oct 7, 2021

Seems like the TF provider 2.79.x is setting the automatic_upgrade_channel and passing it's default value even if a user doesn't specify it. Like any other preview feature you can't pass the property at all (regardless of the value) unless you have the preview flag is enabled, otherwise it would call preview code.

hashicorp/terraform-provider-azurerm@48a5a47

This should likely be handled the same way as other preview features and not setting it/passing it by default.

@tombuildsstuff
Copy link

@palma21 FWIW we raised this issue with the AKS Team on Friday since this API behaviour differs from some other Azure API's where it's possible to set this a field to the default value (e.g. none) without being opted into the feature.

A null value and none aren't treated the same on the AKS side. The value when this is set to null isn't returned in the API response unless it's explicitly configured, however, there is a default behaviour for this field. Notably we've had reports that the default value isn't set to none in some cases either, for example here it's set to the value "undefined" in the API response which isn't defined in the Swagger spec. Within Terraform we need a means of reconciling these to a single default value.

Unfortunately whilst this is a bug on our side (and it's unfortunate that we didn't spot setting this to none required the Preview feature, which is something we've got a plan to fix longer-term) - since there are multiple possible default values for this field (null, none and undefined), removing this default value would reintroduce a separate bug into the plan (that this field is shown as having no value when there's multiple defaults available in the API) - so to be honest I don't think that's necessarily the right fix here.

Were there a single default value here then we could reliably not send this value in the request, however, since there's multiple possible default values and the AKS API isn't returning the value for this field from the API if it's unset, but is defaulting this fields behaviour (and there's multiple possible default values) - in practice we need to do so to coalesce this on our side. One of the core tenants of Terraform is that the plan says what it'll do (and folks use tools like Sentinel & OpenPolicyAgent to validate the Terraform Configurations match what they expect) - so from our side it's imperative that we can determine the default value for these fields.

This is causing issues for users today, this issue shows a user provisioning the same cluster using the same settings at different points in time and getting different results from the AKS API - which is ultimately why we've had to set this to none here to disable this preview functionality.

If it's not possible to determine the single default value for a given Preview feature, and that can't be explicitly disabled without being opted into the Preview - given the requirements here we may have to consider omitting support for new Preview functionality until that's fixed (and that's not something we want to do / in the best interest of users) - so ultimately I think that both of these bugs need to be fixed in the API here.

Happy to chat further about this, but this is an issue affecting all preview functionality within AKS, this is exacerbated by having multiple default values, which we've no means of reconciling down to a single value in the Provider with the information returned from the API unfortunately.

@y3sy
Copy link

y3sy commented Oct 8, 2021

@sossickd I opened an issue with the Terraform azurerm provider this morning hashicorp/terraform-provider-azurerm#13661 on this exact issue, and since there is a workaround (omit the autoUpgradeProfile property from the generated JSON) I expect the provider team to fix this.

juandiegopalomino added a commit to run-x/opta that referenced this issue Oct 8, 2021
Honestly, I'm very disappointed with Azure:
Azure/AKS#2584
juandiegopalomino added a commit to run-x/opta that referenced this issue Oct 8, 2021
Honestly, I'm very disappointed with Azure:
Azure/AKS#2584
NitinAgg added a commit to run-x/opta that referenced this issue Oct 11, 2021
* Attempting more graceful usage of git dependency (#393)

Import is not within a try-catch and only imported if the code for its usage is invoked.

* Nicer plan displayer (#391)

* Nicer plan displayer

1. Using tables to show data about changes
2. Only 2 security levels "LOW RISK" and "HIGH RISK". Low risk represents no expected dataloss and minimal, recoverable downtime

* more colors

* typo

* addressing cr

* max node count increase should be benign

* addressing cr

* Preemptive GCP instances. (#394)

* Fix unnecessary logs and other minor issues (#395)

* fix-stuff

* revert-pipfile

* fix-dict

* fix-lint

* fix-lint

* Update nightly.yaml

* Checking b/w 2 versions will now verify no dataloss (#398)

* Feature/dashlocal (#397)

Added --local functionality for Opta local runs.

* Check in ci that pdb is not mentioned anywhere in the code (#401)

* Check in ci that pdb is not mentioned anywhere in the code

* wip

* wip

* Tiny fix for our runx module (#405)

* Jd/fixing displayer (#406)

* plan displayer handle delete

* destroy was not a valid value

* adding deletions to test

* Fixed local yaml quote issue (#404)

* Jd/fixing aws destroy (#407)

* AWS destroy facing issue b/c cli can't clean up auto created sec group

terraform-aws-modules/terraform-aws-vpc#283

* wip

* terraform fmt

* Adding Secondary Gcp NodePool Opta module (#403)

* Adding Secondary Gcp NodePool Opta module

* Update the GCP Node Pool name

* Update gcp-env.yml example and fix terraform lint

* Add IAM Member permissions

* Supporting a list of helm values files (#409)

Also, now we check for relative path and support it

* Updating to linkerd v2.10.2 (#408)

* Updating to linkerd v2.10.2

OK, looks like all of the important work was already taken care of earlier as we skipped the outbound mysql and postgres ports already.

More good news: the linkerd visualization components are now their own separate charts so linkerd resource overhead will drop a lot.

https://linkerd.io/2.10/tasks/upgrading-2.10-ports-and-protocols/#
https://linkerd.io/2.10/tasks/upgrade/#upgrade-notice-stable-2-10-0

* terraform fmt

* Fixing gcp dns delegation

* Fixed relative path issues for yaml files (#411)

* Disabling ssl for gcp postgres (#415)

Doing this because in order to enable ssl for postgres a user would need to download the ssl CA/key files and include them in all outgoing connections, which makes psql incredibly cumbersome to use. So for now, we won't be supporting it.

* Forgot to update regula to not complain about missing gcp postgres ssl (#416)

* Fixing azure destroy (#414) (#418)

* Fixing azure destroy

Sometime in the past week Azure terraform started failing when trying to destroy the acr key vault key because we disabled purge and yet destroy causes a purge. Found the new toggle to just do a soft delete on purge and confirmed that it worked.

* Disabling regular rule for purge

* Reverting azure provider to version 2.78.0

Honestly, I'm very disappointed with Azure:
Azure/AKS#2584

* Fixed terraform local working dir (#413)

Co-authored-by: Juan Diego Palomino <[email protected]>
Co-authored-by: Nilesh Sarupriya <[email protected]>
Co-authored-by: Sachin Agarwal <[email protected]>
NitinAgg added a commit to run-x/opta that referenced this issue Oct 11, 2021
* Add yaml syntax highlighting (#402)

* Release 0.15.0 (#410)

* Attempting more graceful usage of git dependency (#393)

Import is not within a try-catch and only imported if the code for its usage is invoked.

* Nicer plan displayer (#391)

* Nicer plan displayer

1. Using tables to show data about changes
2. Only 2 security levels "LOW RISK" and "HIGH RISK". Low risk represents no expected dataloss and minimal, recoverable downtime

* more colors

* typo

* addressing cr

* max node count increase should be benign

* addressing cr

* Preemptive GCP instances. (#394)

* Fix unnecessary logs and other minor issues (#395)

* fix-stuff

* revert-pipfile

* fix-dict

* fix-lint

* fix-lint

* Update nightly.yaml

* Checking b/w 2 versions will now verify no dataloss (#398)

* Feature/dashlocal (#397)

Added --local functionality for Opta local runs.

* Check in ci that pdb is not mentioned anywhere in the code (#401)

* Check in ci that pdb is not mentioned anywhere in the code

* wip

* wip

* Tiny fix for our runx module (#405)

* Jd/fixing displayer (#406)

* plan displayer handle delete

* destroy was not a valid value

* adding deletions to test

* Fixed local yaml quote issue (#404)

* Jd/fixing aws destroy (#407)

* AWS destroy facing issue b/c cli can't clean up auto created sec group

terraform-aws-modules/terraform-aws-vpc#283

* wip

* terraform fmt

* Adding Secondary Gcp NodePool Opta module (#403)

* Adding Secondary Gcp NodePool Opta module

* Update the GCP Node Pool name

* Update gcp-env.yml example and fix terraform lint

* Add IAM Member permissions

* Supporting a list of helm values files (#409)

Also, now we check for relative path and support it

* Updating to linkerd v2.10.2 (#408)

* Updating to linkerd v2.10.2

OK, looks like all of the important work was already taken care of earlier as we skipped the outbound mysql and postgres ports already.

More good news: the linkerd visualization components are now their own separate charts so linkerd resource overhead will drop a lot.

https://linkerd.io/2.10/tasks/upgrading-2.10-ports-and-protocols/#
https://linkerd.io/2.10/tasks/upgrade/#upgrade-notice-stable-2-10-0

* terraform fmt

* Fixing gcp dns delegation

* Fixed relative path issues for yaml files (#411)

* Disabling ssl for gcp postgres (#415)

Doing this because in order to enable ssl for postgres a user would need to download the ssl CA/key files and include them in all outgoing connections, which makes psql incredibly cumbersome to use. So for now, we won't be supporting it.

* Forgot to update regula to not complain about missing gcp postgres ssl (#416)

* Fixing azure destroy (#414) (#418)

* Fixing azure destroy

Sometime in the past week Azure terraform started failing when trying to destroy the acr key vault key because we disabled purge and yet destroy causes a purge. Found the new toggle to just do a soft delete on purge and confirmed that it worked.

* Disabling regular rule for purge

* Reverting azure provider to version 2.78.0

Honestly, I'm very disappointed with Azure:
Azure/AKS#2584

* Fixed terraform local working dir (#413)

Co-authored-by: Juan Diego Palomino <[email protected]>
Co-authored-by: Nilesh Sarupriya <[email protected]>
Co-authored-by: Sachin Agarwal <[email protected]>

Co-authored-by: Anthony Campolo <[email protected]>
Co-authored-by: Nitin Aggarwal <[email protected]>
Co-authored-by: Juan Diego Palomino <[email protected]>
Co-authored-by: Nilesh Sarupriya <[email protected]>
antonioT90 pushed a commit to pagopa/selfcare-infra that referenced this issue Oct 12, 2021
antonioT90 added a commit to pagopa/selfcare-infra that referenced this issue Oct 14, 2021
* removed portal and management from apim

* cleaned dev vars

* fixed certificate reference from keyvault

* aks-infra

* setting maxPods

* defining alert metrics

* aks_node_count to 1 in PROD temporarily

* max_pods defined as variable

* provider version rollback due to azure bug Azure/AKS#2584

Co-authored-by: anttorre <[email protected]>
Co-authored-by: Pasquale De Vita <[email protected]>
@FilBot3
Copy link

FilBot3 commented Oct 14, 2021

Any word on this?

I know this: hashicorp/terraform-provider-azurerm#11396 (comment) is currently one way of bypassing this then setting automatic_upgrade_channel to stable would keep you on stable.

Is Microsoft working on getting this "fixed" or updated?

@paulgmiller
Copy link
Member

Allowing ugrade channel none when preview flag is not set is in and currently rolling out. Expect it to reach all regions next week but that is an expectation not a promise.

nsarupr added a commit to run-x/opta that referenced this issue Oct 21, 2021
* Attempting more graceful usage of git dependency (#393)

Import is not within a try-catch and only imported if the code for its usage is invoked.

* Nicer plan displayer (#391)

* Nicer plan displayer

1. Using tables to show data about changes
2. Only 2 security levels "LOW RISK" and "HIGH RISK". Low risk represents no expected dataloss and minimal, recoverable downtime

* more colors

* typo

* addressing cr

* max node count increase should be benign

* addressing cr

* Preemptive GCP instances. (#394)

* Fix unnecessary logs and other minor issues (#395)

* fix-stuff

* revert-pipfile

* fix-dict

* fix-lint

* fix-lint

* Update nightly.yaml

* Checking b/w 2 versions will now verify no dataloss (#398)

* Feature/dashlocal (#397)

Added --local functionality for Opta local runs.

* Check in ci that pdb is not mentioned anywhere in the code (#401)

* Check in ci that pdb is not mentioned anywhere in the code

* wip

* wip

* Tiny fix for our runx module (#405)

* Jd/fixing displayer (#406)

* plan displayer handle delete

* destroy was not a valid value

* adding deletions to test

* Fixed local yaml quote issue (#404)

* Jd/fixing aws destroy (#407)

* AWS destroy facing issue b/c cli can't clean up auto created sec group

terraform-aws-modules/terraform-aws-vpc#283

* wip

* terraform fmt

* Adding Secondary Gcp NodePool Opta module (#403)

* Adding Secondary Gcp NodePool Opta module

* Update the GCP Node Pool name

* Update gcp-env.yml example and fix terraform lint

* Add IAM Member permissions

* Supporting a list of helm values files (#409)

Also, now we check for relative path and support it

* Updating to linkerd v2.10.2 (#408)

* Updating to linkerd v2.10.2

OK, looks like all of the important work was already taken care of earlier as we skipped the outbound mysql and postgres ports already.

More good news: the linkerd visualization components are now their own separate charts so linkerd resource overhead will drop a lot.

https://linkerd.io/2.10/tasks/upgrading-2.10-ports-and-protocols/#
https://linkerd.io/2.10/tasks/upgrade/#upgrade-notice-stable-2-10-0

* terraform fmt

* Fixing gcp dns delegation

* Fixed relative path issues for yaml files (#411)

* Disabling ssl for gcp postgres (#415)

Doing this because in order to enable ssl for postgres a user would need to download the ssl CA/key files and include them in all outgoing connections, which makes psql incredibly cumbersome to use. So for now, we won't be supporting it.

* Forgot to update regula to not complain about missing gcp postgres ssl (#416)

* Fixing azure destroy (#414) (#418)

* Fixing azure destroy

Sometime in the past week Azure terraform started failing when trying to destroy the acr key vault key because we disabled purge and yet destroy causes a purge. Found the new toggle to just do a soft delete on purge and confirmed that it worked.

* Disabling regular rule for purge

* Reverting azure provider to version 2.78.0

Honestly, I'm very disappointed with Azure:
Azure/AKS#2584

* Fixed terraform local working dir (#413)

Co-authored-by: Juan Diego Palomino <[email protected]>
Co-authored-by: Nilesh Sarupriya <[email protected]>
Co-authored-by: Sachin Agarwal <[email protected]>
bigbitbus added a commit to run-x/opta that referenced this issue Oct 29, 2021
* Attempting more graceful usage of git dependency (#393)

Import is not within a try-catch and only imported if the code for its usage is invoked.

* Nicer plan displayer (#391)

* Nicer plan displayer

1. Using tables to show data about changes
2. Only 2 security levels "LOW RISK" and "HIGH RISK". Low risk represents no expected dataloss and minimal, recoverable downtime

* more colors

* typo

* addressing cr

* max node count increase should be benign

* addressing cr

* Preemptive GCP instances. (#394)

* Fix unnecessary logs and other minor issues (#395)

* fix-stuff

* revert-pipfile

* fix-dict

* fix-lint

* fix-lint

* Update nightly.yaml

* Checking b/w 2 versions will now verify no dataloss (#398)

* Feature/dashlocal (#397)

Added --local functionality for Opta local runs.

* Check in ci that pdb is not mentioned anywhere in the code (#401)

* Check in ci that pdb is not mentioned anywhere in the code

* wip

* wip

* Tiny fix for our runx module (#405)

* Jd/fixing displayer (#406)

* plan displayer handle delete

* destroy was not a valid value

* adding deletions to test

* Fixed local yaml quote issue (#404)

* Jd/fixing aws destroy (#407)

* AWS destroy facing issue b/c cli can't clean up auto created sec group

terraform-aws-modules/terraform-aws-vpc#283

* wip

* terraform fmt

* Adding Secondary Gcp NodePool Opta module (#403)

* Adding Secondary Gcp NodePool Opta module

* Update the GCP Node Pool name

* Update gcp-env.yml example and fix terraform lint

* Add IAM Member permissions

* Supporting a list of helm values files (#409)

Also, now we check for relative path and support it

* Updating to linkerd v2.10.2 (#408)

* Updating to linkerd v2.10.2

OK, looks like all of the important work was already taken care of earlier as we skipped the outbound mysql and postgres ports already.

More good news: the linkerd visualization components are now their own separate charts so linkerd resource overhead will drop a lot.

https://linkerd.io/2.10/tasks/upgrading-2.10-ports-and-protocols/#
https://linkerd.io/2.10/tasks/upgrade/#upgrade-notice-stable-2-10-0

* terraform fmt

* Fixing gcp dns delegation

* Fixed relative path issues for yaml files (#411)

* Disabling ssl for gcp postgres (#415)

Doing this because in order to enable ssl for postgres a user would need to download the ssl CA/key files and include them in all outgoing connections, which makes psql incredibly cumbersome to use. So for now, we won't be supporting it.

* Forgot to update regula to not complain about missing gcp postgres ssl (#416)

* Fixing azure destroy (#414) (#418)

* Fixing azure destroy

Sometime in the past week Azure terraform started failing when trying to destroy the acr key vault key because we disabled purge and yet destroy causes a purge. Found the new toggle to just do a soft delete on purge and confirmed that it worked.

* Disabling regular rule for purge

* Reverting azure provider to version 2.78.0

Honestly, I'm very disappointed with Azure:
Azure/AKS#2584

* Fixed terraform local working dir (#413)

Co-authored-by: Juan Diego Palomino <[email protected]>
Co-authored-by: Nilesh Sarupriya <[email protected]>
Co-authored-by: Sachin Agarwal <[email protected]>
bigbitbus added a commit to run-x/opta that referenced this issue Nov 1, 2021
* Attempting more graceful usage of git dependency (#393)

Import is not within a try-catch and only imported if the code for its usage is invoked.

* Nicer plan displayer (#391)

* Nicer plan displayer

1. Using tables to show data about changes
2. Only 2 security levels "LOW RISK" and "HIGH RISK". Low risk represents no expected dataloss and minimal, recoverable downtime

* more colors

* typo

* addressing cr

* max node count increase should be benign

* addressing cr

* Preemptive GCP instances. (#394)

* Fix unnecessary logs and other minor issues (#395)

* fix-stuff

* revert-pipfile

* fix-dict

* fix-lint

* fix-lint

* Update nightly.yaml

* Checking b/w 2 versions will now verify no dataloss (#398)

* Feature/dashlocal (#397)

Added --local functionality for Opta local runs.

* Check in ci that pdb is not mentioned anywhere in the code (#401)

* Check in ci that pdb is not mentioned anywhere in the code

* wip

* wip

* Tiny fix for our runx module (#405)

* Jd/fixing displayer (#406)

* plan displayer handle delete

* destroy was not a valid value

* adding deletions to test

* Fixed local yaml quote issue (#404)

* Jd/fixing aws destroy (#407)

* AWS destroy facing issue b/c cli can't clean up auto created sec group

terraform-aws-modules/terraform-aws-vpc#283

* wip

* terraform fmt

* Adding Secondary Gcp NodePool Opta module (#403)

* Adding Secondary Gcp NodePool Opta module

* Update the GCP Node Pool name

* Update gcp-env.yml example and fix terraform lint

* Add IAM Member permissions

* Supporting a list of helm values files (#409)

Also, now we check for relative path and support it

* Updating to linkerd v2.10.2 (#408)

* Updating to linkerd v2.10.2

OK, looks like all of the important work was already taken care of earlier as we skipped the outbound mysql and postgres ports already.

More good news: the linkerd visualization components are now their own separate charts so linkerd resource overhead will drop a lot.

https://linkerd.io/2.10/tasks/upgrading-2.10-ports-and-protocols/#
https://linkerd.io/2.10/tasks/upgrade/#upgrade-notice-stable-2-10-0

* terraform fmt

* Updated release helper script (#412)

Prettier, more compact, less repetitive. Scrolling still not there but the updates makes that far less likely

* Fixing azure destroy (#414)

* Fixing azure destroy

Sometime in the past week Azure terraform started failing when trying to destroy the acr key vault key because we disabled purge and yet destroy causes a purge. Found the new toggle to just do a soft delete on purge and confirmed that it worked.

* Disabling regular rule for purge

* Reverting azure provider to version 2.78.0

Honestly, I'm very disappointed with Azure:
Azure/AKS#2584

* Fixing gcp dns delegation

* Disabling ssl for gcp postgres (#415)

Doing this because in order to enable ssl for postgres a user would need to download the ssl CA/key files and include them in all outgoing connections, which makes psql incredibly cumbersome to use. So for now, we won't be supporting it.

* Forgot to update regula to not complain about missing gcp postgres ssl (#416)

* Merge main to dev (#420)

* Add yaml syntax highlighting (#402)

* Release 0.15.0 (#410)

* Attempting more graceful usage of git dependency (#393)

Import is not within a try-catch and only imported if the code for its usage is invoked.

* Nicer plan displayer (#391)

* Nicer plan displayer

1. Using tables to show data about changes
2. Only 2 security levels "LOW RISK" and "HIGH RISK". Low risk represents no expected dataloss and minimal, recoverable downtime

* more colors

* typo

* addressing cr

* max node count increase should be benign

* addressing cr

* Preemptive GCP instances. (#394)

* Fix unnecessary logs and other minor issues (#395)

* fix-stuff

* revert-pipfile

* fix-dict

* fix-lint

* fix-lint

* Update nightly.yaml

* Checking b/w 2 versions will now verify no dataloss (#398)

* Feature/dashlocal (#397)

Added --local functionality for Opta local runs.

* Check in ci that pdb is not mentioned anywhere in the code (#401)

* Check in ci that pdb is not mentioned anywhere in the code

* wip

* wip

* Tiny fix for our runx module (#405)

* Jd/fixing displayer (#406)

* plan displayer handle delete

* destroy was not a valid value

* adding deletions to test

* Fixed local yaml quote issue (#404)

* Jd/fixing aws destroy (#407)

* AWS destroy facing issue b/c cli can't clean up auto created sec group

terraform-aws-modules/terraform-aws-vpc#283

* wip

* terraform fmt

* Adding Secondary Gcp NodePool Opta module (#403)

* Adding Secondary Gcp NodePool Opta module

* Update the GCP Node Pool name

* Update gcp-env.yml example and fix terraform lint

* Add IAM Member permissions

* Supporting a list of helm values files (#409)

Also, now we check for relative path and support it

* Updating to linkerd v2.10.2 (#408)

* Updating to linkerd v2.10.2

OK, looks like all of the important work was already taken care of earlier as we skipped the outbound mysql and postgres ports already.

More good news: the linkerd visualization components are now their own separate charts so linkerd resource overhead will drop a lot.

https://linkerd.io/2.10/tasks/upgrading-2.10-ports-and-protocols/#
https://linkerd.io/2.10/tasks/upgrade/#upgrade-notice-stable-2-10-0

* terraform fmt

* Fixing gcp dns delegation

* Fixed relative path issues for yaml files (#411)

* Disabling ssl for gcp postgres (#415)

Doing this because in order to enable ssl for postgres a user would need to download the ssl CA/key files and include them in all outgoing connections, which makes psql incredibly cumbersome to use. So for now, we won't be supporting it.

* Forgot to update regula to not complain about missing gcp postgres ssl (#416)

* Fixing azure destroy (#414) (#418)

* Fixing azure destroy

Sometime in the past week Azure terraform started failing when trying to destroy the acr key vault key because we disabled purge and yet destroy causes a purge. Found the new toggle to just do a soft delete on purge and confirmed that it worked.

* Disabling regular rule for purge

* Reverting azure provider to version 2.78.0

Honestly, I'm very disappointed with Azure:
Azure/AKS#2584

* Fixed terraform local working dir (#413)

Co-authored-by: Juan Diego Palomino <[email protected]>
Co-authored-by: Nilesh Sarupriya <[email protected]>
Co-authored-by: Sachin Agarwal <[email protected]>

Co-authored-by: Anthony Campolo <[email protected]>
Co-authored-by: Nitin Aggarwal <[email protected]>
Co-authored-by: Juan Diego Palomino <[email protected]>
Co-authored-by: Nilesh Sarupriya <[email protected]>

* Add a prompt for configuration file. (#419)

* Add a prompt for configuration file.

* Added Test cases.

* Refactoring

* Update comment

Co-authored-by: Nitin Aggarwal <[email protected]>

* Improvements to logs and helper strings (#423)

* improvements

* undo-pipfile

* ci

* Deleting that old debugger I made ages ago and was hidden since February (#425)

* Adding the tags for alb ingress to vpc (#426)

https://kubernetes-sigs.github.io/aws-load-balancer-controller/v2.2/deploy/subnet_discovery/
Adding tags hurts nothing so should be zero problem

* Feat/nilesh/runx 800 better creds error (#424)

* Compare System Configured Credentials (AWS/GCP)

* Comment fix.

* Update Error Message

* Updating helm module docs to be less confusing (#428)

* Add support for Multiple Instance in aws-documentdb (#427)

* Add support for Multiple Instance in aws-documentdb

* Updated Test Cases. Validate Document DB Instance count.

* Lint fixes.

* Opta Module Uniqueness check (#429)

* Opta Module uniqueness check

* Adding the Uniqueness check for required Opta modules.

* Adding the Uniqueness check for required Opta modules.

* Added Uniqueness check for external-ssl-cert

* Persistent storage option for k8s services (#430)

* Persistent storage option for k8s services

* Terraform fmt

* Addressing cr

* Addressing cr

* Unified helm chart for k8s service (#433)

Turns out that the only difference inside the chart was the service account annotations, which we can just put placeholders for lol

* Fixing a bug where due to a silly fix gcp k8s base needed gcp dns (#435)

* Retool example (#431)

* retool

* updates

* Enhanced amplitude event properties (#438)

* Enhanced amplitude event properties

1. All will have parent name if applicable
2. There will be a module_* count for certain modules we wish to keep track of
3. There is a new event for recording the end of an apply.

* Addressing cr

* fixing tests

* Fixing tests

* addressing cr

* addressing cr

* lol

* Add support for custom JSON encoding (#440)

* Add support for custom JSON encoding

* Add tests

* Fix lint issues

* Validate encoding matches stdlib

Co-authored-by: Patrick Fiedler <[email protected]>

* Not sure how it happened, but I forgot a couple of metric counts (#442)

* Some resources are failing in CD because they are not unique enough (#443)

Honestly, surprised it took this long to spot this

* AWS TF resources with Valid Resource Names (#447)

* AWS TF resources with Valid Resource Names

* Fix TF Formatting.

* Check if DynamoDB Exists (#449)

* Check if DynamoDB Exists

* Fix Lint.

* Postgres db identifier bugfix (#446)

This is not backwards incompatible because
1. We have lifecycle ignore changes on the db identifier
2. Upper case letters in the db identifier is not allowed, so there would be no identifiers of running clusters who would be changed

* Handling routing without domain via all/path,  */path and /path (#450)

* Dynamodb module (#444)

* Dynamodb module

no local or secondary index for now

* linting/formating

* addressing cr

* addressing cr

* addressing cr

* testfixing

* addressing nitn's comments

* Ignore capitalization of auto (#452)

* Ignore capitalization of auto

* terraform fmt

* adding more tests

* Make local k8s service use universal helm chart (#434)

I think it should work as straightforward as with the other clouds
also there where some features missing form local k8s service which should now be present.

Deleting

* Fix k8s-service failing when public_uri is not specified (#454)

Co-authored-by: Patrick Fiedler <[email protected]>

Co-authored-by: Juan Diego Palomino <[email protected]>
Co-authored-by: Nilesh Sarupriya <[email protected]>
Co-authored-by: Nitin Aggarwal <[email protected]>
Co-authored-by: Anthony Campolo <[email protected]>
Co-authored-by: Patrick Fiedler <[email protected]>
Co-authored-by: Patrick Fiedler <[email protected]>
@ghost ghost added the action-required label Nov 9, 2021
@ghost ghost added the stale Stale issue label Jan 8, 2022
@ghost
Copy link

ghost commented Jan 8, 2022

This issue has been automatically marked as stale because it has not had any activity for 60 days. It will be closed if no further activity occurs within 15 days of this comment.

@ghost ghost closed this as completed Jan 16, 2022
@ghost
Copy link

ghost commented Jan 16, 2022

This issue will now be closed because it hasn't had any activity for 7 days after stale. sossickd feel free to comment again on the next 7 days to reopen or open a new issue after that time if you still have a question/issue or suggestion.

@ghost ghost locked as resolved and limited conversation to collaborators Feb 15, 2022
This issue was closed.
Sign up for free to subscribe to this conversation on GitHub. Already have an account? Sign in.
Projects
None yet
Development

No branches or pull requests

7 participants