#48 kempty-n12.ci.centos.org seeds pods with wrong umask
Closed: Fixed by pingou. Opened by jlebon.

Pods started on node kempty-n12.ci.centos.org will get a umask of 0000, which is unexpected and leads to errors in some of our CI jobs. Admittedly, our tests should be able to handle this, but a default umask of 0000 is very unusual and is bound to cause subtle issues for other workloads as well.

I haven't been able to reproduce this on any other node, which makes me believe there's some configuration drift somehow affecting only n12. Is there anything special about that node?

FWIW, manifest I'm testing with:

apiVersion: v1
metadata:
    name: coreos-assembler-sleep
kind: Pod
spec:
  containers:
   - name: coreos-assembler-sleep
     image: quay.io/coreos-assembler/coreos-assembler:latest
     imagePullPolicy: Always
     workingDir: /srv/
     command: ['/usr/bin/dumb-init']
     args: ['sleep', 'infinity']
     resources:
       requests:
         cpu: "4"

Metadata Update from @dkirwan:
- Issue tagged with: centos-ci-infra

We have the ability to just reinstall a node on demand, I'm thinking I'll just mark this node unschedulable, drain it of pods, and reinstall it. Will look into this next week.

Metadata Update from @dkirwan:
- Issue priority set to: Waiting on Assignee (was: Needs Review)
- Issue tagged with: low-gain, low-trouble

Metadata Update from @dkirwan:
- Issue assigned to dkirwan

Metadata Update from @dkirwan:
- Issue priority set to: None (was: Waiting on Assignee)

kempty-n12.ci.centos.org has been reinstalled.

Metadata Update from @dkirwan:
- Issue status updated to: Closed (was: Open)

Issue status updated to: Open (was: Closed)

Metadata Update from @dkirwan:
- Issue status updated to: Closed (was: Open)

Issue status updated to: Open (was: Closed)

Issue status updated to: Closed (was: Open)
Issue close_status updated to: Fixed

Log in to comment on this ticket.

Metadata