Some users are experiencing slow filesystem issues on fram.
We are currently trying to resolve the issue, but it is a periodical problem that makes debugging difficult.
Sorry for the inconvenience this is causing.
Best Regards
Infra team
Some users are experiencing slow filesystem issues on fram.
We are currently trying to resolve the issue, but it is a periodical problem that makes debugging difficult.
Sorry for the inconvenience this is causing.
Best Regards
Infra team
[UPDATE, 2022-05-13 08:00] The maintenance stop is over. There may still be some file system issues. Please report via regular support channels
[UPDATE, 2022-05-11 08:00] The maintenance stop has now started.
The will be a maintenance of the storage system on the Fram supercomputer. The cluster will be unavailable from the 11th of May at 08:00 until the 12th of May at 20:00.
We need to replace an IO module and HDD in Fram storage today.
This should not affect your work and the system will be running as usual during the procedure. If you get any file system issues during this service please send us a support request.
Best Regards
Infrastructure Team
We need to conduct some work on the filesystem controllers for NIRD – TOS. Unfortunately this results in a short unavailability (downtime) period.
All services connected to- and/or utilizing TOS (Tromsø) part of NIRD will be affected. Exported NFS services mounted on FRAM will unfortunately NOT be available either.
The maintenance is set for Thursday 07.04.22 from 09:00-11:00 AM
We are sorry for any inconveniences that may occur. Opslog is updated as soon as the system is back in production.
UPDATE 07-04-2022 – 11:25 … we are still working on the issue and starting to bring the file system up, we hope to back in production soon
UPDATE 07-04-2022 – 12:25 … we are struggling and fighting with the file system, doing our best, we are very sorry for the troubles the issue is causing you
UPDATE 07-04-2022 – 15:35 … the file system is back up and running
[Update, 2022-02-24 22:30]: The maintenance is over and Fram is in production again. Thank you for your patience!
[Update, 2022-02-24 20:30]: The maintenance is taking a little longer than planned. We plan to get back into production at 22:00.
[Update, 2022-02-23 12:00]: The maintenance stop has now begun.
Fram supercomputer will be unavailable due to maintenance on the cooling system from February 23rd 12:00 until 24th 20:00
If time allows it we will also upgrade whole or parts/components of the storage system, including file system clients (compute nodes)
We are currently experiencing issues with the filesystem on Saga. The problem affects the /cluster/software area only. That is why modules are not able to load. /cluster/{home,work,project} is not affected. We are trying to fix it as quickly as possible.
Dear NIRD users,
we will have to do some minor maintenance on the file system controllers in Trondheim again. The operation has been scheduled for today 15.30 …
We don’t expect any major outage but a slight file system performance issue might occur.
Thanks for your understanding
Dear NIRD users,
we will have to do some minor maintenance on the file system controllers in Trondheim. The operation has been scheduled for today 15 p.m. …
We don’t expect any major outage but a slight file system performance issue might occur.
Thanks for your understanding
Good morning,
we are going to perform some routine maintenance on one of the file system controllers of FRAM. This should have no significant implications for production, users might experience slightly degraded Lustre (file system) performance.
The operation is scheduled for today – 11 a.m. …
Update 8.07: There were also performance issues with the login nodes. This and the controller maintenance is now finished.
Dear Stallo Users,
UPDATE – 27.11/16:20 – we have opened the machine for you guys but there might be some instabilities on global file system as we have also lost one object storage server. The issue is being investigated and we are waiting for some spare parts.
We have some major problems with Lustre file system at the moment. One of the main storage coolers is down. We are kicking out all users now and hope to get the machine back to an operational state ASAP.
Thank you for your patience.
HPC staff (UiT)