Commit a74f5d06 authored by incardon's avatar incardon

Fixing run

parent 32acc8b9
......@@ -14,11 +14,8 @@ parallel (
stage ('run_nyu')
{
parallel (
"proc2" : {sh "cd openfpm_vcluster && ./run.sh $WORKSPACE $NODE_NAME 2"}
"proc3" : {sh "cd openfpm_vcluster && ./run.sh $WORKSPACE $NODE_NAME 3"}
)
sh "cd openfpm_vcluster && ./run.sh $WORKSPACE $NODE_NAME 2"
sh "cd openfpm_vcluster && ./run.sh $WORKSPACE $NODE_NAME 3"
sh "cd openfpm_vcluster && ./run.sh $WORKSPACE $NODE_NAME 4"
sh "cd openfpm_vcluster && ./run.sh $WORKSPACE $NODE_NAME 5"
}
......@@ -40,11 +37,8 @@ parallel (
stage ('run_sb15')
{
parallel (
"proc2" : {sh "cd openfpm_vcluster && ./run.sh $WORKSPACE $NODE_NAME 2"}
"proc3" : {sh "cd openfpm_vcluster && ./run.sh $WORKSPACE $NODE_NAME 3"}
)
sh "cd openfpm_vcluster && ./run.sh $WORKSPACE $NODE_NAME 2"
sh "cd openfpm_vcluster && ./run.sh $WORKSPACE $NODE_NAME 3"
sh "cd openfpm_vcluster && ./run.sh $WORKSPACE $NODE_NAME 4"
sh "cd openfpm_vcluster && ./run.sh $WORKSPACE $NODE_NAME 5"
sh "cd openfpm_vcluster && ./run.sh $WORKSPACE $NODE_NAME 6"
......
......@@ -13,20 +13,9 @@ if [ "$2" == "wetcluster" ]; then
export MODULEPATH="/sw/apps/modules/modulefiles:$MODULEPATH"
## Run on the cluster
bsub -o output_run2.%J -K -n 2 -R "span[hosts=1]" "module load openmpi/1.8.1 ; module load gcc/4.9.2; mpirun -np 2 ./src/vcluster"
bsub -o output_run2.%J -K -n 2 -R "span[hosts=1]" "module load openmpi/1.8.1 ; module load gcc/4.9.2; mpirun -np $3 ./src/vcluster"
if [ $? -ne 0 ]; then exit 1 ; fi
bsub -o output_run4.%J -K -n 4 -R "span[hosts=1]" "module load openmpi/1.8.1 ; module load gcc/4.9.2; mpirun -np 4 ./src/vcluster"
if [ $? -ne 0 ]; then exit 1 ; fi
bsub -o output_run8.%J -K -n 8 -R "span[hosts=1]" "module load openmpi/1.8.1 ; module load gcc/4.9.2; mpirun -np 8 ./src/vcluster"
if [ $? -ne 0 ]; then exit 1 ; fi
bsub -o output_run12.%J -K -n 12 -R "span[hosts=1]" "module load openmpi/1.8.1 ; module load gcc/4.9.2; mpirun -np 12 ./src/vcluster"
if [ $? -ne 0 ]; then exit 1 ; fi
# bsub -o output_run32.%J -K -n 32 "module load openmpi/1.8.1 ; module load gcc/4.9.2; mpirun -np 32 ./src/vcluster"
# if [ $? -ne 0 ]; then exit 1 ; fi
# bsub -o output_run32.%J -K -n 64 "module load openmpi/1.8.1 ; module load gcc/4.9.2; mpirun -np 64 ./src/vcluster"
# if [ $? -ne 0 ]; then exit 1 ; fi
# bsub -o output_run32.%J -K -n 128 "module load openmpi/1.8.1 ; module load gcc/4.9.2; mpirun -np 128 ./src/vcluster"
# if [ $? -ne 0 ]; then exit 1 ; fi
elif [ "$2" == "taurus" ]; then
echo "Running on taurus"
......@@ -38,31 +27,27 @@ elif [ "$2" == "taurus" ]; then
### to exclude --exclude=taurusi[6300-6400],taurusi[5400-5500]
salloc --nodes=1 --ntasks-per-node=24 --time=00:05:00 --mem-per-cpu=1800 --partition=haswell bash -c "ulimit -s unlimited && mpirun -np 24 src/vcluster --report_level=no"
salloc --nodes=1 --ntasks-per-node=$3 --time=00:05:00 --mem-per-cpu=1800 --partition=haswell bash -c "ulimit -s unlimited && mpirun -np $3 src/vcluster --report_level=no"
if [ $? -ne 0 ]; then exit 1 ; fi
sleep 5
salloc --nodes=2 --ntasks-per-node=24 --time=00:05:00 --mem-per-cpu=1800 --partition=haswell bash -c "ulimit -s unlimited && mpirun -np 48 src/vcluster --report_level=no"
if [ $? -ne 0 ]; then exit 1 ; fi
sleep 5
salloc --nodes=4 --ntasks-per-node=24 --time=00:05:00 --mem-per-cpu=1800 --partition=haswell bash -c "ulimit -s unlimited && mpirun -np 96 src/vcluster --report_level=no"
if [ $? -ne 0 ]; then exit 1 ; fi
sleep 5
salloc --nodes=8 --ntasks-per-node=24 --time=00:05:00 --mem-per-cpu=1800 --partition=haswell bash -c "ulimit -s unlimited && mpirun -np 192 src/vcluster --report_level=no"
if [ $? -ne 0 ]; then exit 1 ; fi
sleep 5
salloc --nodes=10 --ntasks-per-node=24 --time=00:5:00 --mem-per-cpu=1800 --partition=haswell bash -c "ulimit -s unlimited && mpirun -np 240 src/vcluster --report_level=no"
if [ $? -ne 0 ]; then exit 1 ; fi
# salloc --nodes=2 --ntasks-per-node=24 --time=00:05:00 --mem-per-cpu=1800 --partition=haswell bash -c "ulimit -s unlimited && mpirun -np 48 src/vcluster --report_level=no"
# if [ $? -ne 0 ]; then exit 1 ; fi
# sleep 5
# salloc --nodes=4 --ntasks-per-node=24 --time=00:05:00 --mem-per-cpu=1800 --partition=haswell bash -c "ulimit -s unlimited && mpirun -np 96 src/vcluster --report_level=no"
# if [ $? -ne 0 ]; then exit 1 ; fi
# sleep 5
# salloc --nodes=8 --ntasks-per-node=24 --time=00:05:00 --mem-per-cpu=1800 --partition=haswell bash -c "ulimit -s unlimited && mpirun -np 192 src/vcluster --report_level=no"
# if [ $? -ne 0 ]; then exit 1 ; fi
# sleep 5
# salloc --nodes=10 --ntasks-per-node=24 --time=00:5:00 --mem-per-cpu=1800 --partition=haswell bash -c "ulimit -s unlimited && mpirun -np 240 src/vcluster --report_level=no"
# if [ $? -ne 0 ]; then exit 1 ; fi
else
source $HOME/.bashrc
echo "$PATH"
mpirun -np 2 ./src/vcluster
if [ $? -ne 0 ]; then exit 1 ; fi
mpirun -np 3 ./src/vcluster
if [ $? -ne 0 ]; then exit 1 ; fi
mpirun -np 4 ./src/vcluster
mpirun -np $3 ./src/vcluster
if [ $? -ne 0 ]; then exit 1 ; fi
fi
......
Markdown is supported
0% or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment