Skip to content
GitLab
Projects
Groups
Snippets
Help
Loading...
Help
Help
Support
Community forum
Keyboard shortcuts
?
Submit feedback
Sign in
Toggle navigation
O
openfpm_vcluster
Project overview
Project overview
Details
Activity
Releases
Repository
Repository
Files
Commits
Branches
Tags
Contributors
Graph
Compare
Locked Files
Issues
0
Issues
0
List
Boards
Labels
Service Desk
Milestones
Merge Requests
0
Merge Requests
0
Requirements
Requirements
List
CI / CD
CI / CD
Pipelines
Jobs
Schedules
Security & Compliance
Security & Compliance
Dependency List
License Compliance
Operations
Operations
Environments
Analytics
Analytics
CI / CD
Code Review
Insights
Issue
Repository
Value Stream
Wiki
Wiki
Members
Members
Collapse sidebar
Close sidebar
Activity
Graph
Create a new issue
Jobs
Commits
Issue Boards
Open sidebar
openfpm
openfpm_vcluster
Commits
a74f5d06
Commit
a74f5d06
authored
Oct 24, 2016
by
incardon
Browse files
Options
Browse Files
Download
Email Patches
Plain Diff
Fixing run
parent
32acc8b9
Changes
2
Hide whitespace changes
Inline
Side-by-side
Showing
2 changed files
with
19 additions
and
40 deletions
+19
-40
Jenkinsfile
Jenkinsfile
+4
-10
run.sh
run.sh
+15
-30
No files found.
Jenkinsfile
View file @
a74f5d06
...
...
@@ -14,11 +14,8 @@ parallel (
stage
(
'run_nyu'
)
{
parallel
(
"proc2"
:
{
sh
"cd openfpm_vcluster && ./run.sh $WORKSPACE $NODE_NAME 2"
}
"proc3"
:
{
sh
"cd openfpm_vcluster && ./run.sh $WORKSPACE $NODE_NAME 3"
}
)
sh
"cd openfpm_vcluster && ./run.sh $WORKSPACE $NODE_NAME 2"
sh
"cd openfpm_vcluster && ./run.sh $WORKSPACE $NODE_NAME 3"
sh
"cd openfpm_vcluster && ./run.sh $WORKSPACE $NODE_NAME 4"
sh
"cd openfpm_vcluster && ./run.sh $WORKSPACE $NODE_NAME 5"
}
...
...
@@ -40,11 +37,8 @@ parallel (
stage
(
'run_sb15'
)
{
parallel
(
"proc2"
:
{
sh
"cd openfpm_vcluster && ./run.sh $WORKSPACE $NODE_NAME 2"
}
"proc3"
:
{
sh
"cd openfpm_vcluster && ./run.sh $WORKSPACE $NODE_NAME 3"
}
)
sh
"cd openfpm_vcluster && ./run.sh $WORKSPACE $NODE_NAME 2"
sh
"cd openfpm_vcluster && ./run.sh $WORKSPACE $NODE_NAME 3"
sh
"cd openfpm_vcluster && ./run.sh $WORKSPACE $NODE_NAME 4"
sh
"cd openfpm_vcluster && ./run.sh $WORKSPACE $NODE_NAME 5"
sh
"cd openfpm_vcluster && ./run.sh $WORKSPACE $NODE_NAME 6"
...
...
run.sh
View file @
a74f5d06
...
...
@@ -13,20 +13,9 @@ if [ "$2" == "wetcluster" ]; then
export
MODULEPATH
=
"/sw/apps/modules/modulefiles:
$MODULEPATH
"
## Run on the cluster
bsub
-o
output_run2.%J
-K
-n
2
-R
"span[hosts=1]"
"module load openmpi/1.8.1 ; module load gcc/4.9.2; mpirun -np
2
./src/vcluster"
bsub
-o
output_run2.%J
-K
-n
2
-R
"span[hosts=1]"
"module load openmpi/1.8.1 ; module load gcc/4.9.2; mpirun -np
$3
./src/vcluster"
if
[
$?
-ne
0
]
;
then
exit
1
;
fi
bsub
-o
output_run4.%J
-K
-n
4
-R
"span[hosts=1]"
"module load openmpi/1.8.1 ; module load gcc/4.9.2; mpirun -np 4 ./src/vcluster"
if
[
$?
-ne
0
]
;
then
exit
1
;
fi
bsub
-o
output_run8.%J
-K
-n
8
-R
"span[hosts=1]"
"module load openmpi/1.8.1 ; module load gcc/4.9.2; mpirun -np 8 ./src/vcluster"
if
[
$?
-ne
0
]
;
then
exit
1
;
fi
bsub
-o
output_run12.%J
-K
-n
12
-R
"span[hosts=1]"
"module load openmpi/1.8.1 ; module load gcc/4.9.2; mpirun -np 12 ./src/vcluster"
if
[
$?
-ne
0
]
;
then
exit
1
;
fi
# bsub -o output_run32.%J -K -n 32 "module load openmpi/1.8.1 ; module load gcc/4.9.2; mpirun -np 32 ./src/vcluster"
# if [ $? -ne 0 ]; then exit 1 ; fi
# bsub -o output_run32.%J -K -n 64 "module load openmpi/1.8.1 ; module load gcc/4.9.2; mpirun -np 64 ./src/vcluster"
# if [ $? -ne 0 ]; then exit 1 ; fi
# bsub -o output_run32.%J -K -n 128 "module load openmpi/1.8.1 ; module load gcc/4.9.2; mpirun -np 128 ./src/vcluster"
# if [ $? -ne 0 ]; then exit 1 ; fi
elif
[
"
$2
"
==
"taurus"
]
;
then
echo
"Running on taurus"
...
...
@@ -38,31 +27,27 @@ elif [ "$2" == "taurus" ]; then
### to exclude --exclude=taurusi[6300-6400],taurusi[5400-5500]
salloc
--nodes
=
1
--ntasks-per-node
=
24
--time
=
00:05:00
--mem-per-cpu
=
1800
--partition
=
haswell bash
-c
"ulimit -s unlimited && mpirun -np 24
src/vcluster --report_level=no"
salloc
--nodes
=
1
--ntasks-per-node
=
$3
--time
=
00:05:00
--mem-per-cpu
=
1800
--partition
=
haswell bash
-c
"ulimit -s unlimited && mpirun -np
$3
src/vcluster --report_level=no"
if
[
$?
-ne
0
]
;
then
exit
1
;
fi
sleep
5
salloc
--nodes
=
2
--ntasks-per-node
=
24
--time
=
00:05:00
--mem-per-cpu
=
1800
--partition
=
haswell bash
-c
"ulimit -s unlimited && mpirun -np 48 src/vcluster --report_level=no"
if
[
$?
-ne
0
]
;
then
exit
1
;
fi
sleep
5
salloc
--nodes
=
4
--ntasks-per-node
=
24
--time
=
00:05:00
--mem-per-cpu
=
1800
--partition
=
haswell bash
-c
"ulimit -s unlimited && mpirun -np 96 src/vcluster --report_level=no"
if
[
$?
-ne
0
]
;
then
exit
1
;
fi
sleep
5
salloc
--nodes
=
8
--ntasks-per-node
=
24
--time
=
00:05:00
--mem-per-cpu
=
1800
--partition
=
haswell bash
-c
"ulimit -s unlimited && mpirun -np 192 src/vcluster --report_level=no"
if
[
$?
-ne
0
]
;
then
exit
1
;
fi
sleep
5
salloc
--nodes
=
10
--ntasks-per-node
=
24
--time
=
00:5:00
--mem-per-cpu
=
1800
--partition
=
haswell bash
-c
"ulimit -s unlimited && mpirun -np 240 src/vcluster --report_level=no"
if
[
$?
-ne
0
]
;
then
exit
1
;
fi
#
salloc --nodes=2 --ntasks-per-node=24 --time=00:05:00 --mem-per-cpu=1800 --partition=haswell bash -c "ulimit -s unlimited && mpirun -np 48 src/vcluster --report_level=no"
#
if [ $? -ne 0 ]; then exit 1 ; fi
#
sleep 5
#
salloc --nodes=4 --ntasks-per-node=24 --time=00:05:00 --mem-per-cpu=1800 --partition=haswell bash -c "ulimit -s unlimited && mpirun -np 96 src/vcluster --report_level=no"
#
if [ $? -ne 0 ]; then exit 1 ; fi
#
sleep 5
#
salloc --nodes=8 --ntasks-per-node=24 --time=00:05:00 --mem-per-cpu=1800 --partition=haswell bash -c "ulimit -s unlimited && mpirun -np 192 src/vcluster --report_level=no"
#
if [ $? -ne 0 ]; then exit 1 ; fi
#
sleep 5
#
salloc --nodes=10 --ntasks-per-node=24 --time=00:5:00 --mem-per-cpu=1800 --partition=haswell bash -c "ulimit -s unlimited && mpirun -np 240 src/vcluster --report_level=no"
#
if [ $? -ne 0 ]; then exit 1 ; fi
else
source
$HOME
/.bashrc
echo
"
$PATH
"
mpirun
-np
2 ./src/vcluster
if
[
$?
-ne
0
]
;
then
exit
1
;
fi
mpirun
-np
3 ./src/vcluster
if
[
$?
-ne
0
]
;
then
exit
1
;
fi
mpirun
-np
4 ./src/vcluster
mpirun
-np
$3
./src/vcluster
if
[
$?
-ne
0
]
;
then
exit
1
;
fi
fi
...
...
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment