Upd:Linux Cluster: UV still unavailable, MPP returned to operation
LRZ aktuell
publish at lrz.de
Mi Feb 6 12:53:08 CET 2013
Changes to this message: MPP returned to user operation
Update (Feb 6, 13:00): The MPP cluster has been returned to user
operation. Please see below what you need to do as a consequence of the
configuration changes. Dear users of the Linux Cluster systems at LRZ,
unfortunately the Ultraviolet systems are presently unavailable for
user operation, because a defective NUMAlink component needs
replacement.
We'll keep you updated about the status, and of course are working
toward a speedy return to user operation. Apologies for the disruption
of services.
-----------------------------------------------------------------------
Recent configuration changes:
* The default MPI environment on the MPP cluster was changed from
Parastation MPI to Intel MPI. However, the mpi.parastation module
will remain available for legacy use until the end of 2013. On the
sgi ICE and UV systems, the sgi MPI (mpi.mpt) will remain default.
We strongly recommend that you recompile your applications with
Intel MPI. If you insist on still running Parastation MPI binaries,
you need to add the following lines to your SLURM job scripts:
module unload mpi.intel
module load mpi.parastation/5.0/intel
* The 8-way Myrinet Cluster has been retired from parallel
processing, and the nodes were added to the serial processing pool.
This implies that the partition "myri_std" in the SLURM cell "myri"
is now unavailable.
* For the serial queues, SLURM fair share scheduling was introduced.
For the parallel queues, a combination of fair share scheduling and
favoring large jobs is now active. This is to prevent a single user
from monopolizing cluster segments for long times if there are many
jobs in the queue.
* New storage systems have been introduced for the WORK (==PROJECT)
and SCRATCH file systems. Please note that LRZ has only migrated
WORK data to the new file system; data in SCRATCH have not been
migrated. However the old SCRATCH file system will remain available
as a separate mount in read-only mode on the login nodes until the
end of March, 2013. Migration of data can then be done via commands
like
cd $SCRATCH_LEGACY
cp -a <my_scratch_subdirectory> $SCRATCH
The environment variable $SCRATCH_LEGACY will remain defined and
point to the legacy scratch area until end of March, 2013.
This information is also available on our web server
http://www.lrz-muenchen.de/services/compute/aktuell/ali4501/
Reinhold Bader
Mehr Informationen über die Mailingliste aktuell