2020-08-10 15:27:46 +02:00
|
|
|
{
|
2020-09-16 12:22:55 +02:00
|
|
|
stdenv
|
|
|
|
, nixpkgs
|
|
|
|
, pkgs
|
2020-08-10 15:27:46 +02:00
|
|
|
, genApp
|
|
|
|
, genConfigs
|
2020-09-02 17:07:09 +02:00
|
|
|
, runWrappers
|
2020-08-10 15:27:46 +02:00
|
|
|
}:
|
|
|
|
|
2020-09-02 17:07:09 +02:00
|
|
|
with stdenv.lib;
|
|
|
|
|
2020-08-10 15:27:46 +02:00
|
|
|
let
|
2020-09-21 14:34:08 +02:00
|
|
|
bsc = pkgs.bsc;
|
|
|
|
|
2020-09-02 17:07:09 +02:00
|
|
|
# Set variable configuration for the experiment
|
|
|
|
varConfig = {
|
2020-09-21 14:34:08 +02:00
|
|
|
cc = [ bsc.icc ];
|
2020-10-05 12:33:44 +02:00
|
|
|
mpi = [ bsc.impi ];
|
2020-08-25 18:39:31 +02:00
|
|
|
blocksize = [ 1024 ];
|
2020-08-12 14:00:04 +02:00
|
|
|
};
|
|
|
|
|
2020-09-02 17:07:09 +02:00
|
|
|
# Common configuration
|
|
|
|
common = {
|
|
|
|
# Compile time nbody config
|
2020-09-21 14:34:08 +02:00
|
|
|
gitBranch = "garlic/tampi+send+oss+task";
|
2020-09-02 17:07:09 +02:00
|
|
|
|
|
|
|
# nbody runtime options
|
2020-10-05 10:47:16 +02:00
|
|
|
particles = 1024*4;
|
|
|
|
timesteps = 10;
|
2020-09-02 17:07:09 +02:00
|
|
|
|
|
|
|
# Resources
|
2020-10-05 10:47:16 +02:00
|
|
|
ntasksPerNode = "2";
|
|
|
|
nodes = "2";
|
2020-09-02 17:07:09 +02:00
|
|
|
|
|
|
|
# Stage configuration
|
2020-10-05 16:40:22 +02:00
|
|
|
enableRunexp = true;
|
2020-09-02 17:07:09 +02:00
|
|
|
enableSbatch = true;
|
|
|
|
enableControl = true;
|
|
|
|
enableExtrae = false;
|
|
|
|
enablePerf = false;
|
2020-09-21 14:34:08 +02:00
|
|
|
enableCtf = false;
|
2020-09-02 17:07:09 +02:00
|
|
|
|
|
|
|
# MN4 path
|
|
|
|
nixPrefix = "/gpfs/projects/bsc15/nix";
|
2020-08-10 15:27:46 +02:00
|
|
|
};
|
|
|
|
|
|
|
|
# Compute the cartesian product of all configurations
|
2020-09-02 17:07:09 +02:00
|
|
|
configs = map (conf: conf // common) (genConfigs varConfig);
|
2020-08-12 14:00:04 +02:00
|
|
|
|
2020-09-02 17:07:09 +02:00
|
|
|
stageProgram = stage:
|
|
|
|
if stage ? programPath
|
|
|
|
then "${stage}${stage.programPath}" else "${stage}";
|
|
|
|
|
|
|
|
w = runWrappers;
|
|
|
|
|
2020-09-30 09:32:25 +02:00
|
|
|
sbatch = {stage, conf, ...}: with conf; w.sbatch (
|
|
|
|
# Allow a user to define a custom reservation for the job in MareNostrum4,
|
|
|
|
# by setting the garlic.sbatch.reservation attribute in the
|
|
|
|
# ~/.config/nixpkgs/config.nix file. If the attribute is not set, no
|
|
|
|
# reservation is used. The user reservation may be overwritten by the
|
|
|
|
# experiment, if the reservation is set like with nodes or ntasksPerNode.
|
|
|
|
optionalAttrs (pkgs.config ? garlic.sbatch.reservation) {
|
|
|
|
inherit (pkgs.config.garlic.sbatch) reservation;
|
|
|
|
} // {
|
|
|
|
program = stageProgram stage;
|
|
|
|
exclusive = true;
|
|
|
|
time = "02:00:00";
|
|
|
|
qos = "debug";
|
|
|
|
jobName = "nbody-tampi";
|
|
|
|
inherit nixPrefix nodes ntasksPerNode;
|
|
|
|
}
|
|
|
|
);
|
2020-08-12 14:00:04 +02:00
|
|
|
|
2020-09-02 17:07:09 +02:00
|
|
|
control = {stage, conf, ...}: with conf; w.control {
|
|
|
|
program = stageProgram stage;
|
|
|
|
};
|
|
|
|
|
|
|
|
srun = {stage, conf, ...}: with conf; w.srun {
|
|
|
|
program = stageProgram stage;
|
2020-10-05 10:47:16 +02:00
|
|
|
srunOptions = "--cpu-bind=verbose,socket";
|
2020-09-02 17:07:09 +02:00
|
|
|
inherit nixPrefix;
|
2020-08-17 18:50:18 +02:00
|
|
|
};
|
|
|
|
|
2020-09-02 17:07:09 +02:00
|
|
|
statspy = {stage, conf, ...}: with conf; w.statspy {
|
|
|
|
program = stageProgram stage;
|
|
|
|
};
|
|
|
|
|
|
|
|
perf = {stage, conf, ...}: with conf; w.perf {
|
|
|
|
program = stageProgram stage;
|
|
|
|
perfArgs = "sched record -a";
|
|
|
|
};
|
|
|
|
|
2020-10-05 16:40:22 +02:00
|
|
|
isolate = {stage, conf, ...}: with conf; w.isolate {
|
2020-09-02 17:07:09 +02:00
|
|
|
program = stageProgram stage;
|
2020-10-05 12:33:44 +02:00
|
|
|
clusterName = "mn4";
|
2020-10-05 16:40:22 +02:00
|
|
|
inherit nixPrefix;
|
2020-09-02 17:07:09 +02:00
|
|
|
};
|
|
|
|
|
|
|
|
extrae = {stage, conf, ...}: w.extrae {
|
|
|
|
program = stageProgram stage;
|
|
|
|
traceLib = "mpi"; # mpi -> libtracempi.so
|
|
|
|
configFile = ./extrae.xml;
|
|
|
|
};
|
|
|
|
|
2020-09-21 14:34:08 +02:00
|
|
|
ctf = {stage, conf, ...}: w.argv {
|
|
|
|
program = stageProgram stage;
|
|
|
|
env = ''
|
|
|
|
export NANOS6=ctf
|
|
|
|
export NANOS6_CTF2PRV=0
|
|
|
|
'';
|
|
|
|
};
|
|
|
|
|
2020-09-02 17:07:09 +02:00
|
|
|
argv = {stage, conf, ...}: w.argv {
|
|
|
|
program = stageProgram stage;
|
|
|
|
env = ''
|
|
|
|
set -e
|
|
|
|
export I_MPI_THREAD_SPLIT=1
|
|
|
|
'';
|
|
|
|
argv = ''( -t ${toString conf.timesteps}
|
|
|
|
-p ${toString conf.particles} )'';
|
|
|
|
};
|
|
|
|
|
2020-09-21 14:34:08 +02:00
|
|
|
bscOverlay = import ../../../overlay.nix;
|
2020-09-16 12:22:55 +02:00
|
|
|
|
|
|
|
genPkgs = newOverlay: nixpkgs {
|
|
|
|
overlays = [
|
|
|
|
bscOverlay
|
|
|
|
newOverlay
|
|
|
|
];
|
2020-09-02 17:07:09 +02:00
|
|
|
};
|
|
|
|
|
2020-09-16 12:22:55 +02:00
|
|
|
# We may be able to use overlays by invoking the fix function directly, but we
|
|
|
|
# have to get the definition of the bsc packages and the garlic ones as
|
|
|
|
# overlays.
|
|
|
|
|
|
|
|
nbodyFn = {stage, conf, ...}: with conf;
|
|
|
|
let
|
|
|
|
# We set the mpi implementation to the one specified in the conf, so all
|
|
|
|
# packages in bsc will use that one.
|
|
|
|
customPkgs = genPkgs (self: super: {
|
|
|
|
bsc = super.bsc // { mpi = conf.mpi; };
|
|
|
|
});
|
|
|
|
in
|
|
|
|
customPkgs.bsc.garlic.nbody.override {
|
|
|
|
inherit cc blocksize mpi gitBranch;
|
|
|
|
};
|
|
|
|
|
2020-10-05 12:33:44 +02:00
|
|
|
launch = w.launch.override {
|
|
|
|
nixPrefix = common.nixPrefix;
|
|
|
|
};
|
|
|
|
|
2020-09-02 17:07:09 +02:00
|
|
|
stages = with common; []
|
2020-10-05 16:40:22 +02:00
|
|
|
# Launch the experiment remotely
|
|
|
|
#++ optional enableRunexp runexp
|
|
|
|
|
2020-09-02 17:07:09 +02:00
|
|
|
# Use sbatch to request resources first
|
|
|
|
++ optional enableSbatch sbatch
|
|
|
|
|
|
|
|
# Repeats the next stages N times
|
2020-10-05 16:40:22 +02:00
|
|
|
++ optionals enableControl [ isolate control ]
|
2020-09-02 17:07:09 +02:00
|
|
|
|
|
|
|
# Executes srun to launch the program in the requested nodes, and
|
|
|
|
# immediately after enters the nix environment again, as slurmstepd launches
|
|
|
|
# the next stages from outside the namespace.
|
2020-10-05 16:40:22 +02:00
|
|
|
++ [ srun isolate ]
|
2020-09-02 17:07:09 +02:00
|
|
|
|
|
|
|
# Intrumentation with extrae
|
|
|
|
++ optional enableExtrae extrae
|
|
|
|
|
|
|
|
# Optionally profile the next stages with perf
|
|
|
|
++ optional enablePerf perf
|
|
|
|
|
2020-09-21 14:34:08 +02:00
|
|
|
# Optionally profile nanos6 with the new ctf
|
|
|
|
++ optional enableCtf ctf
|
|
|
|
|
2020-09-02 17:07:09 +02:00
|
|
|
# Execute the nbody app with the argv and env vars
|
|
|
|
++ [ argv nbodyFn ];
|
|
|
|
|
|
|
|
# List of actual programs to be executed
|
|
|
|
jobs = map (conf: w.stagen { inherit conf stages; }) configs;
|
2020-08-12 14:00:04 +02:00
|
|
|
|
2020-10-05 16:40:22 +02:00
|
|
|
launcher = launch jobs;
|
|
|
|
|
|
|
|
runexp = stage: w.runexp {
|
|
|
|
program = stageProgram stage;
|
|
|
|
nixPrefix = common.nixPrefix;
|
|
|
|
};
|
|
|
|
|
|
|
|
isolatedRun = stage: isolate {
|
|
|
|
inherit stage;
|
|
|
|
conf = common;
|
|
|
|
};
|
|
|
|
|
|
|
|
final = runexp (isolatedRun launcher);
|
|
|
|
|
|
|
|
|
2020-08-10 15:27:46 +02:00
|
|
|
in
|
2020-09-02 17:07:09 +02:00
|
|
|
# We simply run each program one after another
|
2020-10-05 16:40:22 +02:00
|
|
|
#launch jobs
|
|
|
|
final
|