11 - Benchmarking CPU resources with HEPIX [ID:20090]
50 von 189 angezeigt

Okay, perfect. Okay, so thanks a lot for the invitation, first of all. I will report about

the work we are doing in the HEPIX benchmarking working group to benchmark WLHG resources.

So essentially, okay, this is the outline of my talk, so first I'll discuss why we need

Sorry Andrea, we see slide number seven currently.

Oh, sorry. I, yeah.

Sorry, I was on the wrong screen. Okay, sorry. So, okay, so I discussed the introduction,

so this is the outline of my talk, essentially I discuss why we need benchmarking, what was

the old approach and its limitation and what is the new approach, implementation, the status

and some outlook. So, in the previous talk, we already introduced, so Lucas already introduced

a bit CERN and the computing environment. So essentially, from the computing resources

point of view, the work that the LHC experiments at CERN are doing, essentially they execute

their jobs on the WLHG, which is the worldwide LHC computing grid. This is a distributed

resource infrastructure, which includes many countries and many sites, so hundreds of sites.

And so the job is distributed on different types of CPUs and there is a need to understand

which CPUs and how many they use. So you see here in this plot, there are essentially

one million CPU cores, but essentially each core is different from another, because some

are able to do more work than others and also, of course, the price is different for different

types of CPUs. So essentially, why we need to benchmark CPU resources for various reasons.

The first is accounting, in the sense that experiments request the amount of resources

they need for their computing, giving a number. So X CPU resources, the funding agencies and

the sites must provision this X and then the resource review boards must compare this X

which was used to the one which was requested. Related, but slightly different, is procurement

in the sense that then the site actually have to buy the resources to provide this X. And

these benchmarks in any case, they can also be used for other things like scheduling and

software optimization. Now, I should say, when I heard that I had to give this talk, I mean,

I'm going to discuss what we are doing, which is benchmarking of computing resources, which

is maybe not benchmarking of the software itself. Of course, it's very related, but

let's say the numbers we get are essentially to decide how much work a given software can

do on some different types of resources. So on the next slide, this is essentially the

benchmark that has been used so far in WLCG is called HEPSpec 06. So you can see here

a typical plot which is shown at the review board. You can see the evolution of how much

CPU power in terms of this HEPSpec 06 has been delivered over time to the four different

experiments. And then you can also see on the other plot essentially how much this compares

to actually what was used, how much was given with respect to what was pledged, what was

promised. Just as an approximate rule of thumb, then HEPSpec 06 corresponds to one core, essentially.

Now, HEPSpec 06 is derived from SPEC CPU 2006. SPEC CPU 2006 is a benchmark developed by

the Standard Performance Evaluation Corporation, and it's an industry standard since the late

80s. Since this is actually an open source software workshop, we should mention also

that SPEC CPU 06 is not open source software, so it's licensed software.

One of the main points about this benchmark is that it's based on real application. Now,

CPU benchmarks in general, let's say there are three kinds. There can be synthetic kernel

benchmarks, kernel benchmarks for real applications. This is based on real applications. However,

it's not based on applications from HEP. It's based on applications essentially from other

scientific domains. In particular, in this suite, there are seven benchmarks. You see

the details there, four of which are floating point benchmarks and three are integer benchmarks.

These are essentially the C++ benchmarks because our main language is C++.

On slide seven, I say why actually this was chosen back in 2009, so more than 10 years

ago. Essentially, two things. One is that it showed good correlation to the throughputs

of HEP workloads, essentially to how many events per second you can process on a given

resource. Essentially, there was a good correlation between the benchmark given by SPEC 06 and

Teil einer Videoserie :

Zugänglich über

Offener Zugang

Dauer

00:21:42 Min

Aufnahmedatum

2020-07-24

Hochgeladen am

2020-07-24 19:16:24

Sprache

en-US

Speaker

Lukas Heinrich, CERN

Content

Benchmarking WLCG resources using HEP experiment workloads

The Workshop

The Workshop on Open-Source Software Lifecycles (WOSSL) was held in the context of the European  Science Cluster of Astronomy & Particle Physics ESFRI infrastructures (ESCAPE), bringing together people, data and services to contribute to the European Open Science Cloud. The workshop was held online from 23rd-28th July 2020, organized@FAU.

Copyright: CC-BY 4.0

Einbetten
Wordpress FAU Plugin
iFrame
Teilen