[Date Prev][Date Next] [Thread Prev][Thread Next] [Date Index] [Thread Index]

Bug#1014506: slurm-wlm: flaky autopkgtest: sbatch fails without



Source: slurm-wlm
Version: 21.08.8.2-1
Severity: serious
X-Debbugs-CC: debian-ci@lists.debian.org
User: debian-ci@lists.debian.org
Usertags: flaky

Dear maintainer(s),

I looked at the results of the autopkgtest of you package on armhf because it was showing up as a regression for the upload of perl. I noticed that the test regularly fails and I saw failures on other architectures too.

Because the unstable-to-testing migration software now blocks on
regressions in testing, flaky tests, i.e. tests that flip between
passing and failing without changes to the list of installed packages,
are causing people unrelated to your package to spend time on these
tests.

Don't hesitate to reach out if you need help and some more information
from our infrastructure.

Paul

https://ci.debian.net/packages/s/slurm-wlm/

https://ci.debian.net/data/autopkgtest/testing/armhf/s/slurm-wlm/23115549/log.gz

autopkgtest [01:29:11]: test sbatch: [-----------------------
● slurmctld.service - Slurm controller daemon
Loaded: loaded (/lib/systemd/system/slurmctld.service; enabled; vendor preset: enabled)
     Active: active (running) since Mon 2022-06-27 01:29:11 UTC; 10s ago
       Docs: man:slurmctld(8)
   Main PID: 3099 (slurmctld)
      Tasks: 10
     Memory: 1.4M
        CPU: 4.907s
     CGroup: /system.slice/slurmctld.service
             ├─3099 /usr/sbin/slurmctld -D -s
             └─3103 "slurmctld: slurmscriptd"

Jun 27 01:29:11 ci-178-32c04b32 systemd[1]: Started Slurm controller daemon.
Jun 27 01:29:11 ci-178-32c04b32 slurmctld[3099]: slurmctld: error: High latency for 1000 calls to gettimeofday(): 288 microseconds Jun 27 01:29:11 ci-178-32c04b32 slurmctld[3099]: slurmctld: Recovered state of 1 nodes Jun 27 01:29:11 ci-178-32c04b32 slurmctld[3099]: slurmctld: Recovered JobId=1 Assoc=0 Jun 27 01:29:11 ci-178-32c04b32 slurmctld[3099]: slurmctld: Recovered information about 1 jobs Jun 27 01:29:11 ci-178-32c04b32 slurmctld[3099]: slurmctld: Recovered state of 0 reservations Jun 27 01:29:11 ci-178-32c04b32 slurmctld[3099]: slurmctld: read_slurm_conf: backup_controller not specified Jun 27 01:29:11 ci-178-32c04b32 slurmctld[3099]: slurmctld: Running as primary controller Jun 27 01:29:11 ci-178-32c04b32 slurmctld[3099]: slurmctld: No parameter for mcs plugin, default values set Jun 27 01:29:11 ci-178-32c04b32 slurmctld[3099]: slurmctld: mcs: MCSParameters = (null). ondemand set.
● slurmd.service - Slurm node daemon
Loaded: loaded (/lib/systemd/system/slurmd.service; enabled; vendor preset: enabled)
     Active: active (running) since Mon 2022-06-27 01:29:11 UTC; 10s ago
       Docs: man:slurmd(8)
   Main PID: 3106 (slurmd)
      Tasks: 1
     Memory: 1.1M
        CPU: 106ms
     CGroup: /system.slice/slurmd.service
             └─3106 /usr/sbin/slurmd -D -s

Jun 27 01:29:11 ci-178-32c04b32 systemd[1]: Started Slurm node daemon.
Jun 27 01:29:11 ci-178-32c04b32 slurmd[3106]: slurmd: slurmd version 21.08.8-2 started Jun 27 01:29:11 ci-178-32c04b32 slurmd[3106]: slurmd: slurmd started on Mon, 27 Jun 2022 01:29:11 +0000 Jun 27 01:29:11 ci-178-32c04b32 slurmd[3106]: slurmd: CPUs=1 Boards=1 Sockets=1 Cores=1 Threads=1 Memory=513603 TmpDisk=2806 Uptime=1228661 CPUSpecList=(null) FeaturesAvail=(null) FeaturesActive=(null)
PARTITION AVAIL  TIMELIMIT  NODES  STATE NODELIST
test*        up   infinite      1   idle localhost
NODELIST   NODES PARTITION STATE
localhost      1     test* idle
Submitted batch job 2
autopkgtest [01:29:27]: test sbatch: -----------------------]

Attachment: OpenPGP_signature
Description: OpenPGP digital signature


Reply to: