Best in Class
The most efficient experiments on GPGrid that used more than 100,000 hours since October 1, 2016 were LArIAT (96.11%) and Minos (95.5%).
The most efficient big non-production user on GPGrid who used more than 100,000 hours since October 1, 2016 was Konstantinos Vellidis with 98.7% efficiency.
The experiment with the most opportunistic hours on OSG between October 1, 2016 and November 30, 2016 was mu2e with 3,607,577 hours.
This newsletter is brought to you by:
- Shreyas Bhat
- Joe Boyd
- Vito Di Benedetto
- Lisa Giacchetti
- Ken Herner
- Burt Holzman
- Mike Kirby
- Tanya Levshina
- Anna Mazzacane
- Marc Mengel
- Kevin Retzke
We welcome articles you might want to submit. Please email email@example.com.
How to get a whole bunch of jobs going while everyone else is sipping eggnog
While everyone enjoys a break from work this time of year, one thing that won’t be taking a break is grid computing. GPGrid will run at full capacity at all times, as will many of the usual offsite computing clusters. We encourage users to continue to submit jobs so that they can run over the holidays.
High-energy physics experiments have an ever-growing need for computing, but all the experiments don't need all the cycles all the time. The need is driven by machine performance, experiment and conference schedules, and even new physics ideas. Computing facilities are purchased with the intention to meet peak workload rather than the average, which impacts the overall computing cost for the facility and the experiment.
The HEP computing model is constantly evolving, and one change that is currently taking place is increased use of High Performance Computing (HPC) resources. Some of these HPC resources include supercomputing sites such as NERSC, as well as the EXtreme Science and Engineering Discovery Environment (XSEDE). XSEDE is actually a collection of several HPC resources, including the Stampede cluster at the Texas Advanced Computing Center.
FIFE provides collaborative scientific-data processing solutions for Frontier Experiments.
FIFE takes the collective experience from current and past experiments to provide options for designing offline computing for experiments. It is not a mandate from Scientific Computing Division about how an experiment should or should not design their offline computing. FIFE is modular so experiments can take what they need, and new tools from outside communities can be incorporated as they develop.
Please click HERE for detailed documentation on the FIFE services and tools.