Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Pod Metric Testing is Flakey due to the Lack of the default namespace #887

Open
jonathan-innis opened this issue Dec 19, 2023 · 3 comments
Labels
kind/bug Categorizes issue or PR as related to a bug. kind/testing Issues that involve adding test coverage lifecycle/stale Denotes an issue or PR has remained open with no activity and has become stale.

Comments

@jonathan-innis
Copy link
Member

Description

Observed Behavior:

Karpenter testing is flaking often with failures in the pod metric testing like

[FAILED] Expected success, but got an error:
      <*errors.StatusError | 0xc000138aa0>: 
      namespaces "default" not found
      {
          ErrStatus: {
              TypeMeta: {Kind: "", APIVersion: ""},
              ListMeta: {
                  SelfLink: "",
                  ResourceVersion: "",
                  Continue: "",
                  RemainingItemCount: nil,
              },
              Status: "Failure",
              Message: "namespaces \"default\" not found",
              Reason: "NotFound",
              Details: {Name: "default", Group: "", Kind: "namespaces", UID: "", Causes: nil, RetryAfterSeconds: 0},
              Code: 404,
          },
      }
  In [It] at: /home/runner/work/karpenter/karpenter/pkg/controllers/metrics/pod/suite_test.go:58 @ 12/18/23 17:16:47.[69](https://github.com/kubernetes-sigs/karpenter/actions/runs/7251475446/job/19753908191?pr=885#step:5:70)6

An example job where this occurred can be seen here: https://github.com/kubernetes-sigs/karpenter/actions/runs/7251475446/job/19753908191?pr=885

This is being tracked upstream to see what their recommendation is e.g. whether we should do a downstream check or whether this is something that can be fixed upstream: kubernetes-sigs/controller-runtime#2626

Expected Behavior:

Tests shouldn't be flakey.

Reproduction Steps (Please include YAML):

Versions:

  • Chart Version:
  • Kubernetes Version (kubectl version):
  • Please vote on this issue by adding a 👍 reaction to the original issue to help the community and maintainers prioritize this request
  • Please do not leave "+1" or "me too" comments, they generate extra noise for issue followers and do not help prioritize the request
  • If you are interested in working on this issue or have submitted a pull request, please leave a comment
@jonathan-innis jonathan-innis added kind/bug Categorizes issue or PR as related to a bug. needs-triage Indicates an issue or PR lacks a `triage/foo` label and requires one. kind/testing Issues that involve adding test coverage and removed needs-triage Indicates an issue or PR lacks a `triage/foo` label and requires one. labels Dec 19, 2023
@jonathan-innis
Copy link
Member Author

This change was merged in kubernetes-sigs/controller-runtime#2668 in upstream controller-runtime. As soon as they release their next version, the hope is that we can bump and we should stop seeing flakes.

@jonathan-innis
Copy link
Member Author

This shouldn't be seen anymore once we get to v0.18.0 of controller-runtime.

@k8s-triage-robot
Copy link

The Kubernetes project currently lacks enough contributors to adequately respond to all issues.

This bot triages un-triaged issues according to the following rules:

  • After 90d of inactivity, lifecycle/stale is applied
  • After 30d of inactivity since lifecycle/stale was applied, lifecycle/rotten is applied
  • After 30d of inactivity since lifecycle/rotten was applied, the issue is closed

You can:

  • Mark this issue as fresh with /remove-lifecycle stale
  • Close this issue with /close
  • Offer to help out with Issue Triage

Please send feedback to sig-contributor-experience at kubernetes/community.

/lifecycle stale

@k8s-ci-robot k8s-ci-robot added the lifecycle/stale Denotes an issue or PR has remained open with no activity and has become stale. label May 13, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
kind/bug Categorizes issue or PR as related to a bug. kind/testing Issues that involve adding test coverage lifecycle/stale Denotes an issue or PR has remained open with no activity and has become stale.
Projects
None yet
Development

No branches or pull requests

3 participants