Test-infra: setup release 1.9 branch

Created on 17 Nov 2017  Â·  27Comments  Â·  Source: kubernetes/test-infra

The branch was cut tonight, I'm working on getting all the right configs updated and figured it would be useful to collect all the PRs involved against this issue for next time.

/area jobs
/area testgrid

arejobs areprovidegcp aretestgrid prioritimportant-soon sicluster-lifecycle simulticluster sinetwork sinode siscalability sitesting

Most helpful comment

@spiffxp Yes

All 27 comments

Derailed for the moment by looking into https://github.com/kubernetes/kubernetes/issues/55917, I suspect that one will not have a great simple fix, it looks like multiple PRs changing the versioning logic went in recently.

Moving the checklist here and breaking up the PR, we decided to do it piece-meal and put in the build job(s) first:

To setup release-1.9:

Testgrid:

  • [x] create sig-release-1.9-all dashboard
  • [x] create sig-release-1.9-blocking dashboard
  • [x] [WIP] add entries for all release-1.9 jobs to the above dashboards
  • [x] add 1.9 jobs to any other appropriate dashboards
  • [x] [WIP] sig dashboards for sig relatedrelease-1.9 jobs
  • [x] move appropriate jobs into the blocking dashboard (~ match release-1.8)

Jenkins:

  • [x] add ci-kubernetes-build-1.9, bump stableN for build jobs
  • [x] add kubernetes-test-go-release-1.9
  • [x] add kubernetes-verify-release-1.9
  • [x] update kubernetes-kubemark-5-gce-last-release trigger job to ci-kubernetes-build-1.9
  • [ ] take a second pass to check for any remaining jobs

Prow:

  • [x] ensure presubmits will run against release-1.9 branch the same as existing master
  • [x] add 1.9 versions of bazel ci/periodic jobs
  • [x] add 1.9 versions of kubeadm ci/periodic jobs
  • [x] add 1.9 versions of device-plugin / gpu jobs
  • [ ] add 1.9 versions of federation jobs (punting to sig-multicluster, see comments below)
  • [x] setup 1.9 versions kubectl-skew jobs

Other TODO:

/area jobs

@shashidharatd @marun what's the deal with kubernetes-federation-build-1.8 etc.?

Complete list of merged PRs involved (will keep up to date):

Semi related PRs:

@shashidharatd @marun what's the deal with kubernetes-federation-build-1.8 etc.?

now federation is moved to k/f repo and the ci build for head is here https://k8s-testgrid.appspot.com/sig-multicluster-federation-gce#build

To create a federation-build release against 1.9 we need to use the jobs which are running against the master branch. Also there are some specific requirements(like increased quota) for federation e2e jobs.

Definetely the federation build is a non-blocker for k8s 1.9 release. We still need to conclude the federation release timelines, whether it should be inline with k8s releases or should have independent release which works for the corresponding k8s release. We will discuss this in our SIG meeting on tuesday.

/cc @kubernetes/sig-multicluster-pr-reviews

I think this is done minus the federation jobs.
Edit and kops :^) (Fixed)

A few nits while attempting to reconcile with diff. In an ideal world, release-1.y-blocking would just be a copy of release-master-blocking with jobs pointed at different branches. Trying to understand why the differences exist. I'm hoping there are some common explanations which we could add in comment form someplace, eg:

  • was added as a release-blocker during v1.8
  • can only be run against release-1.y branches
  • etc.

@kubernetes/sig-gcp-test-failures The gci-{gce,gke} names in sig-release-{master,1.8} are now just {gce,gke} in sig-release-1.9-blocking, can we strip the gci- prefix from the other names? There are both gce and gci-gce jobs in sig-release-1.7-blocking

~@kubernetes/sig-testing-test-failures These jobs in sig-release-1.9-blocking have no equivalent in sig-release-{master,1.8,1.7}-blocking. Should they be removed from release-1.9 or added to any of the others?~ these were added during the v1.9 dev cycle, #5654 ensure they're in master, and we're not backporting

  • ~bazel-build-1.9~
  • ~bazel-test-1.9~

@kubernetes/sig-cluster-lifecycle-test-failures These jobs in sig-release-1.9-blocking have no equivalent in sig-release-{master,1.8}-blocking. Should they be removed from release-1.9 or added to any of the others?

  • gce-kubeadm-upgrade-1.8-1.9

@kubernetes/sig-multicluster-test-failures These jobs in sig-release-1.8-blocking have no equivalent in sig-release-{master,1.9,1.7}-blocking. Should they be removed from release-1.8 or added to any of the others?

  • federation-build-1.8
  • federation-test-1.8

~@kubernetes/sig-network-test-failures These jobs in sig-release-{master,1.9}-blocking have no equivalent in sig-release-{1.8,1.7}-blocking. Should they be added to any of the others?~ disregard, @spiffxp was blind

  • ~gce-ingress-1.9~
  • ~gke-ingress-1.9~

@kubernetes/sig-scalability-test-failures These jobs in sig-release-master-blocking have no equivalent in sig-release-{1.9,1.8,1.7}-blocking. Should they be added to any of the others?

  • gce-scale-correctness
  • gce-scale-performance

~@kubernetes/sig-auth-test-failures These jobs in sig-release-master-blocking have no equivalent in sig-release-{1.9,1.8,1.7}-blocking. Should they be added to any of the others?~ The goal is to fold these tests into the standard suite, and remove the gci-gce-audit job (ref: kubernetes/kubernetes#55953)... and the job is removed via #5690

  • ~gci-gce-audit~

~@kubernetes/sig-scheduling-test-failures These jobs in sig-release-master-blocking have no equivalent in sig-release-{1.9,1.8,1.7}-blocking. Should they be added to any of the others?~ I'm going to claim this was added during the v1.9 dev cycle, and we've added it to the 1.9 board via #5661

  • ~gke-device-plugin-gpu-p100~

~@kubernetes/sig-node-test-failures These jobs in sig-release-master-blocking have no equivalent in sig-release-1.9-blocking. Should they be added to any of the others?~ #5649 added kubelet-1.9

  • ~kubelet~

misc:

  • ~gci-gce-100 (test_group_name is ci-kubernetes-e2e-gci-gce-scalability, there is a gce-scalability job in 1.8 and 1.7)~ this has been added
  • ~soak-gce-gci~ this was dropped from master via #5529

For the ones I know:

  • I have blown away all the 1.7-gke-'cvm' jobs
  • @BenTheElder has added the bazel jobs, we'll need to add that to 1.9-blocking dash
  • we don't create separate gce-scale jobs for release (they are expensive!) but we probably should have a gci-gce-100 for 1.9, I'll create it.
  • I'll add a kubelet job
  • Do we still want a soak job? I'll create one for sure.

+1 to the comments on why something exists, I've at least tried to make sure all the changes we make for 1.9 show up https://github.com/kubernetes/test-infra/issues/5558#issuecomment-345392515

@kubernetes/sig-node-test-failures These jobs in sig-release-master-blocking have no equivalent in sig-release-1.9-blocking. Should they be added to any of the others?

kubelet

Yes, we have kubelet-1.8, and kubelet 1.7 jobs.

@BenTheElder has added the bazel jobs, we'll need to add that to 1.9-blocking dash

These jobs block on presubmit, and are no more flaky than unit / non-bazel build. They're in the 1.9 blocking dash but we should have them in master blocking too.

has added the bazel jobs, we'll need to add that to 1.9-blocking dash

Yeah, they're there, I guess my question to @BenTheElder is, is there any reason they're blocking 1.9 but not blocking master?

lol ok question answered above, as a member of the release team I'll lgtm a pr that adds them based on history of passing against PR's

I'm adding them to master. We have no reason to not block on these.

Federation jobs for 1.9 have been punted to federation to sort out how they want them but "the federation build is a non-blocker for k8s 1.9 release" (since they moved out of k/k?) https://github.com/kubernetes/test-infra/issues/5558#issuecomment-345624723

For sig questions about blocking tests in comment above (https://github.com/kubernetes/test-infra/issues/5558#issuecomment-346385580) I am asking the following sig members to respond with how they would like these issues handled :-)

  • cc @enisoc @abgworrall for possible GCP job name bikeshed
  • cc @luxas for kubeadm / sig-cluster-lifecycle jobs potentially missing from master and previous releases
  • @shashidharatd discussed federation / sig-multicluster above though this isn't solved yet it shouldn't be a problem for now
  • cc @MrHohn for sig-network jobs missing from previous releases
  • not sure who to flag down for sig-auth, but the sigs have all been pinged at their test failures handle anyhow..
  • cc @mindprince for sig-scheduling / GPU jobs only present in master blocking but not any of the release branches (including 1.9)

Please let us know what makes sense for these questions and I will make the necessary changes.

These jobs in sig-release-master-blocking have no equivalent in sig-release-{1.9,1.8,1.7}-blocking. Should they be added to any of the others?
gke-device-plugin-gpu-p100

Yes. Let's add them to release-1.9-blocking. When I added that job, release-1.9-blocking didn't exist, so I only added it to master-blocking.

Yes. Let's add them to release-1.9-blocking. When I added that job, release-1.9-blocking didn't exist, so I only added it to master-blocking.

Thanks, we'll do that :+1:

kubernetes/sig-network-test-failures These jobs in sig-release-{master,1.9}-blocking have no equivalent in sig-release-{1.8,1.7}-blocking. Should they be added to any of the others?

  • gce-ingress-1.9
  • gke-ingress-1.9

@BenTheElder I think we do have equivalent jobs, just that the naming format changed?

@MrHohn you're right, that was my bad, not sure how I missed those

Thanks @MrHohn, with all of the jobs I don't have full track of every job 🙃

@spiffxp Audit tests will be added to the default suite in gce/gke env by https://github.com/kubernetes/kubernetes/pull/55953

@crassirostris does this mean we should remove the audit-specific job once that PR merges?

@spiffxp Yes

There are a few unanswered questions above, but the branch has been set up and running for some time (with a few fixes since then), I think we should close this.

Was this page helpful?
0 / 5 - 0 ratings

Related issues

cblecker picture cblecker  Â·  4Comments

fejta picture fejta  Â·  4Comments

stevekuznetsov picture stevekuznetsov  Â·  3Comments

lavalamp picture lavalamp  Â·  3Comments

BenTheElder picture BenTheElder  Â·  3Comments