Update 10:40: Access to cyclone is reopened now. Delay was caused by missing kernel modules and old Lustre packages.
Cyclone will be sharply rebooted at 09:00 to apply new filesystem settings.
Update 10:40: Access to cyclone is reopened now. Delay was caused by missing kernel modules and old Lustre packages.
Cyclone will be sharply rebooted at 09:00 to apply new filesystem settings.
Update 2018-12-03 12:36:
Dear Hexagon User,
We must reboot Hexagon due to repeated errors on the interconnect.
Will update this case when Hexagon is up and functional again.
Update 12_11 21:30:
Migration is over, we manage to take up Lustre filesystem with new MDS server. /shared and /work filesystem is mounted on cyclone.hpc.uib.no and grunch.hpc.uib.no. Hexagon is up and running again. Samba and NFS exports are also running on Leo.hpc.uib.no.
Update 12_11 15:00 :
Migration is still ongoing, we will keep you posted.
Update 02_11 09:30 :
Due to the delayed delivery of physical parts, we have to postpone our downtime to 12th November. Corresponding node reservation on the hexagon is also postponed to 12th November.
Thank you for your consideration!
Dear HPC User,
The metadata server for the /shared file system has to be replaced/upgraded and therefore it must be unmounted from all the clients.
This will result in scheduled downtime for Hexagon, Grunch and Cyclone machines. We start at 08:00 AM on the 5th of November and expect to be ready by the end of the working day.
Thank you for your consideration!
Update 14-06-2018 11:15: Access is now limited to UiB and IMR campus only.
Starting with 14 of June, access to Hexagon will be limited to campus network.
Please find more details at docs.hpc.uib.no/wiki/Getting_started.
Update 23.05.2018 15:19 File system issues were solved and mounted back to both Hexagon and Grunch. Access to is reopened.
There is a scheduled downtime for Hexagon and /shared file system for Wednesday, 23rd of May. Scheduled downtime will start at 09:00 and we expect to have the systems back by 16:00, same day.
Our apologies for any inconvenience this downtime can give you.
Looks like after the reboot of the machine on Friday not all data nodes hosting /work picked up proper settings and /work fs is temporary slow. We are working to resolve this issue ASAP.
Update 10:00 This issue is resolved now.
There are issues with the cooling in the machine room. We've contacted building maintenance team and are hoping to bring machine up in few hours.
Update: 14:00 Cooling problems are over. Machine is back online.
Interactive job submission is working again. (qsub -I)
From this night there is a problem with one of our storage systems serving /work, we are looking into the problem.
Update 02.12.2014 13:00: Issue has been remediated, /work should be OK now.
module swap cray-mpich cray-mpich2
module swap cray-libsci cray-libsci/12.2.0
module load craype-barcelona
xtpe-interlagos
- replaced by
craype-interlagos