Daily Bulletin Archive

March 11, 2014

The core switch-router at NWSC requires an upgrade to the current supported software version. The update is scheduled for 6 a.m. MT, March 11, and is expected to last 15 minutes.

During the update, users will be unable to log into Yellowstone or to access other services from external hosts. Active SSH and VNC sessions into the Yellowstone environment will also be terminated. However, the update will not affect running jobs on Yellowstone, Geyser, or Caldera.

HPSS will also take an earlier maintenance downtime from 5:30 am - 6:30 am to minimize potential disruption.

Users who experience connection problems early Tuesday morning should try again a little later.

February 26, 2014

An update to the SAM accounting system introduced a charging error for Yellowstone jobs, and jobs that ran yesterday were overcharged. We expect jobs will proceed normally through the scheduler within the hour. CISL staff are in the process of fixing the charging issue and correcting the overcharges. We apologize for the inconvenience.

February 21, 2014

Users who use PGI compilers may experience their jobs being killed with this error:

ERROR: 0031-917 Invalid MP_COMPILER value

As a workaround to avoid this, set the MP_COMPILER environment variable to GNU, as in these examples:

export MP_COMPILER=gnu (for bash/sh/ksh)
setenv MP_COMPILER gnu  (for csh/tcsh)

Jobs submitted with 32 tasks per node using batch option -R "span[ptile=32]" are killed with this error message:

ERROR: 0031-758 AFFINITY: [ys0116] Oversubscribe: 32 tasks in total,  each task requires 1 resource, but there are only 16 available resource. Affinity can not be applied

Correction

To avoid this problem, submit your job with environment variable MP_TASK_AFFINITY set to cpu as shown here:

export MP_TASK_AFFINITY=cpu (for bash/sh/ksh)
setenv MP_TASK_AFFINITY cpu   (for csh/tcsh)

We are working closely with IBM to get a permanent solution for PGI compiler problem soon. Our sincere apologies for the inconvenience.

February 18, 2014

The CISL Help Desk and Consulting Services will be closed in observance of President's Day Monday, February 17. Please report problems or direct requests to cislhelp@ucar.edu. Incoming help requests will be reviewed by the CISL network operations center staff,  but responses to some issues may be deferred until Tuesday.

February 14, 2014

Yellowstone users can ask for assistance from CISL staff experts if they need help visualizing their data. This service is available to researchers who want help using specialized visualization applications on the Geyser or Caldera clusters within the Yellowstone environment. See Advanced visualization support to learn how to submit a request, what criteria must be met, and other details.

February 14, 2014

An XSEDE class hosted by the Texas Advanced Computing Center from 9 a.m. to 5 p.m. CST on Friday, February 21, focuses on optimizing code for the Intel Xeon Phi coprocessor. Participants can attend in person or via webcast and should be able to program using MPI and OpenMP. See the XSEDE calendar for details and to register for the class.

February 12, 2014

Starting at 8 a.m. MT on February 11, CISL will take a full-day outage of Yellowstone, Geyser, Caldera and Pronghorn to update several system software components, including LSF, IBM Parallel Environment (PE), and the Mellanox FCA software. The upgrades should not require any action on the part of users, but the changes require a reboot of the system.

A system-wide reservation should prevent the launching of any jobs that would run into the maintenance window; any jobs running after 8 a.m. MT, February 11, will be killed by system administrators.

During the outage, GLADE and HPSS will remain in service; users can access their data on these resources via the data access nodes at data-access.ucar.edu.

February 10, 2014

Researchers seeking high-performance computing and data storage systems to support extremely demanding, high-profile climate simulations can now submit proposals for Climate Simulation Laboratory (CSL) awards. The deadline for submitting project proposals is March 24, 2014. If you are currently using the CSL and wish to continue after May 30, 2014, please submit a renewal proposal and a progress report by the deadline.

The CSL represents CISL’s premier opportunity for researchers seeking high-performance computing and data storage systems to support large-scale climate simulations. Such simulations require high resolution, span many centuries of simulated time, encompass large numbers of ensembles, integrate new physics or models, or address national and international scientific priorities. Proposals must request a minimum of 10 million core-hours to be used on NCAR’s petascale Yellowstone system between June 1, 2014, and May 30, 2015.

With this call for proposals, the CSL is evolving into a focused effort to support NSF awardees in addressing scientific questions that require large-scale, long-running simulations of the Earth's climate system. Support from an NSF award (or awards) is required.

Projects selected for this round of CSL allocations will have access to Yellowstone, a 1.5-petaflops IBM iDataPlex cluster, and its associated storage, data analysis, and visualization resources. For more information about Yellowstone, see the CISL web site (www2.cisl.ucar.edu). Additional information about the CSL, the competitive review process, and how to submit a proposal can be found at www2.cisl.ucar.edu/csl.

If you have questions about this opportunity, please contact David Hart, User Services Manager, 303-497-1234, or email alloc@ucar.edu. Please share this information with any other colleagues who might be interested.

February 7, 2014

Students interested in attending the fifth International Summer School on HPC Challenges in Computational Sciences, June 1-6, 2014, in Budapest, Hungary, should apply by March 9, 2014. Graduate students and postdoctoral scholars from institutions in Europe, Canada, Japan, and the United States are eligible. Meals, housing, and travel expenses are paid for the selected participants.

Preference is given to applicants with parallel programming experience and a research plan that will benefit from the use of high performance computing systems. The summer school is sponsored by the European Union Seventh Framework Program’s Partnership for Advanced Computing in Europe Implementation Phase project (PRACE-3IP), the National Science Foundation’s Extreme Science and Engineering Discovery Environment (XSEDE) project, RIKEN Advanced Insti­tute for Computational Science (RIKEN AICS), and Compute/Calcul Canada. See http://www.prace-ri.eu/International-Summer-School-2014 for details.

February 3, 2014

CISL has recently deployed a pilot service to explore the features provided by the Globus Plus service, and we would like to invite existing CISL users to explore and evaluate the capabilities offered by this service.

Globus Plus adds data sharing features to the Globus data transfer service that is currently available to Yellowstone users and recommended by CISL for large-scale data transfers into and out of GLADE. (See http://www2.cisl.ucar.edu/docs/transfer/globus.)

Globus Plus extends the Globus paradigm by giving users the tools to share data (i.e., create shareable Globus endpoints), as well as define and manage access rights for external users and groups.

CISL expects that Globus Plus may prove useful for sharing large-scale files on a temporary basis, hosting data sets associated with publications, or providing a back-end for longer-term data-sharing services. Users of the pilot service should NOT begin long-term activities that require the sharing features.

From early users, we are seeking your feedback on the service, how it meets (or doesn’t meet) your data-sharing requirements, and its potential as a long-term offering from CISL. The pilot project will also give us an opportunity to develop policies associated with this service.

If you are interested in putting Globus Plus through its paces, please contact us via cislhelp@ucar.edu, and we will set you up with access and pointers to some preliminary documentation.

Pages