Source: slurm-wlm Version: 21.08.8.2-1 Severity: serious X-Debbugs-CC: debian-ci@lists.debian.org User: debian-ci@lists.debian.org Usertags: flaky Dear maintainer(s),I looked at the results of the autopkgtest of you package on armhf because it was showing up as a regression for the upload of perl. I noticed that the test regularly fails and I saw failures on other architectures too.
Because the unstable-to-testing migration software now blocks on regressions in testing, flaky tests, i.e. tests that flip between passing and failing without changes to the list of installed packages, are causing people unrelated to your package to spend time on these tests. Don't hesitate to reach out if you need help and some more information from our infrastructure. Paul https://ci.debian.net/packages/s/slurm-wlm/ https://ci.debian.net/data/autopkgtest/testing/armhf/s/slurm-wlm/23115549/log.gz autopkgtest [01:29:11]: test sbatch: [----------------------- ● slurmctld.service - Slurm controller daemonLoaded: loaded (/lib/systemd/system/slurmctld.service; enabled; vendor preset: enabled)
Active: active (running) since Mon 2022-06-27 01:29:11 UTC; 10s ago Docs: man:slurmctld(8) Main PID: 3099 (slurmctld) Tasks: 10 Memory: 1.4M CPU: 4.907s CGroup: /system.slice/slurmctld.service ├─3099 /usr/sbin/slurmctld -D -s └─3103 "slurmctld: slurmscriptd" Jun 27 01:29:11 ci-178-32c04b32 systemd[1]: Started Slurm controller daemon.Jun 27 01:29:11 ci-178-32c04b32 slurmctld[3099]: slurmctld: error: High latency for 1000 calls to gettimeofday(): 288 microseconds Jun 27 01:29:11 ci-178-32c04b32 slurmctld[3099]: slurmctld: Recovered state of 1 nodes Jun 27 01:29:11 ci-178-32c04b32 slurmctld[3099]: slurmctld: Recovered JobId=1 Assoc=0 Jun 27 01:29:11 ci-178-32c04b32 slurmctld[3099]: slurmctld: Recovered information about 1 jobs Jun 27 01:29:11 ci-178-32c04b32 slurmctld[3099]: slurmctld: Recovered state of 0 reservations Jun 27 01:29:11 ci-178-32c04b32 slurmctld[3099]: slurmctld: read_slurm_conf: backup_controller not specified Jun 27 01:29:11 ci-178-32c04b32 slurmctld[3099]: slurmctld: Running as primary controller Jun 27 01:29:11 ci-178-32c04b32 slurmctld[3099]: slurmctld: No parameter for mcs plugin, default values set Jun 27 01:29:11 ci-178-32c04b32 slurmctld[3099]: slurmctld: mcs: MCSParameters = (null). ondemand set.
● slurmd.service - Slurm node daemonLoaded: loaded (/lib/systemd/system/slurmd.service; enabled; vendor preset: enabled)
Active: active (running) since Mon 2022-06-27 01:29:11 UTC; 10s ago Docs: man:slurmd(8) Main PID: 3106 (slurmd) Tasks: 1 Memory: 1.1M CPU: 106ms CGroup: /system.slice/slurmd.service └─3106 /usr/sbin/slurmd -D -s Jun 27 01:29:11 ci-178-32c04b32 systemd[1]: Started Slurm node daemon.Jun 27 01:29:11 ci-178-32c04b32 slurmd[3106]: slurmd: slurmd version 21.08.8-2 started Jun 27 01:29:11 ci-178-32c04b32 slurmd[3106]: slurmd: slurmd started on Mon, 27 Jun 2022 01:29:11 +0000 Jun 27 01:29:11 ci-178-32c04b32 slurmd[3106]: slurmd: CPUs=1 Boards=1 Sockets=1 Cores=1 Threads=1 Memory=513603 TmpDisk=2806 Uptime=1228661 CPUSpecList=(null) FeaturesAvail=(null) FeaturesActive=(null)
PARTITION AVAIL TIMELIMIT NODES STATE NODELIST test* up infinite 1 idle localhost NODELIST NODES PARTITION STATE localhost 1 test* idle Submitted batch job 2 autopkgtest [01:29:27]: test sbatch: -----------------------]
Attachment:
OpenPGP_signature
Description: OpenPGP digital signature