Description
Cluster: Baobab
Login node crashed with a kernel panic. We restarted it.
–
HPC Team
Status : Resolved 
start: 2026-01-02T18:53:00Z
end: 2026-01-05T07:50:00Z
Cluster: Baobab
Login node crashed with a kernel panic. We restarted it.
–
HPC Team
start: 2026-01-02T18:53:00Z
end: 2026-01-05T07:50:00Z
Cluster: Bamboo
Login node crashed with a kernel panic. We restarted it.
–
HPC Team
start: 2025-12-31T07:17:00Z
end: 2026-01-05T07:55:00Z
Cluster: Bamboo
Storage server was down for part of scratch storage.
–
HPC Team
start: 2025-12-30T20:22:00Z
end: 2026-01-05T09:19:00Z
Cluster: Bamboo
Login1 has crashed, server has been rebooted
start: 2026-01-12T10:06:00Z
end: 2026-01-11T10:15:00Z
Cluster: Yggdrasil
We are experiencing issue with SLURM causing delays in job management. We are actively working to resolve this incident and limit its impact.
The issue has been resolved and Slurm is back online.
start: 2026-01-15T10:05:00Z
end: 2026-01-15T10:30:00Z
Cluster: Yggdrasil
We are currently facing a power outage affecting the Yggdrasil cluster, which may result in multiple nodes being unreachable. We will provide an update once power has been fully restored.
start: 2026-01-19T10:00:00Z
end: 2026-01-19T10:30:00Z
Cluster: all
Since the latest slurm update, some interactive jobs (using srun or salloc) are killed prematurely. For the user it appears as if the job reached its timelimit, but in the admin logs it is indicated that the job was killed due to inactivity timeout. We opened a case at schedMD.
start: 2025-12-19T10:00:00Z
end: