Started by user arnoldg Running as SYSTEM Building in workspace /var/lib/jenkins/jobs/mpiGraph/workspace [SSH] script: PATH="/sw/tools/bin:/sw/tools/bin:$PATH" NODELIST="""" ACCOUNT="bbmb-hydro" PARTITION="sandybridge" NODES="16" # tar xvf /var/tmp/mpiGraph.tar # galen's copy of the benchmark cd mpiGraph export PATH=~/mpiGraph:$PATH # if the NODELIST parameter was filled in, proceed with a targeted test if [[ ! -z ${NODELIST} ]] then srun --mem=16g --nodes=${NODES} --ntasks-per-node=1 --time=00:10:00 --job-name=mpiGraph \ --partition=${PARTITION} --nodelist=${NODELIST} \ --account=${ACCOUNT} ${GPUS} \ mpiGraph 1048576 1000 10 > mpiGraph.out if [ "${NODES}" == "16" ] then bash clip-plot.sh 16 fi exit fi # otherwise, if NODELIST is "", run the default test on 1st avail nodes srun --nodes=${NODES} \ --mem=16g --ntasks-per-node=1 --time=00:10:00 --job-name=mpiGraph \ --account=${ACCOUNT} \ --partition=${PARTITION} \ mpiGraph 1048576 1000 10 > mpiGraph.out if [ "${NODES}" == "16" ] then bash clip-plot.sh 16 fi [SSH] executing... NODECOUNT=$1 + NODECOUNT=16 # clip out the Send table and plot it egrep --after-context=$NODECOUNT "^Send\s+hy" mpiGraph.out \ | sed -e s/.to//g > send.txt + egrep --after-context=16 '^Send\s+hy' mpiGraph.out + sed -e s/.to//g python3 send.py + python3 send.py hydro15:0 hydro16:1 hydro17:2 ... hydro29:13 hydro30:14 hydro31:15 Send ... hydro15:0 0.000 8445.598 8542.273 ... 3996.850 5967.628 8563.961 hydro16:1 8588.018 0.000 8340.577 ... 3004.817 5930.123 5969.165 hydro17:2 8458.453 8605.242 0.000 ... 2953.077 3004.901 3994.577 hydro18:3 6566.959 7173.455 7245.029 ... 2942.261 2404.457 3003.288 hydro19:4 8411.161 8464.080 8563.994 ... 1983.856 2003.694 2994.887 hydro20:5 8315.311 8404.943 8375.025 ... 1976.718 1718.974 2985.979 hydro22:6 8225.513 8300.488 8407.679 ... 1979.790 1717.635 1987.664 hydro23:7 8293.147 8205.379 8364.407 ... 2946.739 1718.592 1976.614 hydro24:8 8537.363 8200.005 8223.714 ... 2953.306 2002.485 1971.815 hydro25:9 1722.770 1721.196 1722.455 ... 8370.497 8310.841 8287.548 hydro26:10 2005.335 1720.496 1718.646 ... 8389.156 8413.308 8270.272 hydro27:11 2405.345 2004.538 1719.585 ... 7161.000 7398.799 7450.837 hydro28:12 3002.774 2410.643 2006.536 ... 8346.327 8436.095 8312.796 hydro29:13 4006.101 3012.836 2411.817 ... 0.000 8293.291 8511.281 hydro30:14 5958.831 4003.027 3003.832 ... 8610.551 0.000 8343.998 hydro31:15 8297.819 5965.736 3992.823 ... 8537.267 8590.586 0.000 [16 rows x 16 columns] # clip out the Recv table and plot it egrep --after-context=$NODECOUNT "^Recv\s+hy" mpiGraph.out \ | sed -e s/.from//g > recv.txt + egrep --after-context=16 '^Recv\s+hy' mpiGraph.out + sed -e s/.from//g python3 recv.py + python3 recv.py hydro15:0 hydro16:1 hydro17:2 ... hydro29:13 hydro30:14 hydro31:15 Recv ... hydro15:0 0.000 8590.268 8539.212 ... 4013.985 5965.691 8303.527 hydro16:1 8463.819 0.000 8615.164 ... 3016.418 4012.170 5979.208 hydro17:2 8568.031 8437.499 0.000 ... 2415.398 3011.338 4000.431 hydro18:3 7454.352 8299.364 7155.602 ... 2007.950 2408.980 3005.002 hydro19:4 8586.848 8516.505 8516.488 ... 1724.079 2011.002 2413.944 hydro20:5 8536.511 8565.634 8518.056 ... 1722.011 1727.176 2019.250 hydro22:6 8562.051 8532.243 8588.289 ... 1725.314 1724.054 1724.004 hydro23:7 8548.510 8545.188 8519.229 ... 2012.767 1725.772 1724.117 hydro24:8 8549.198 8560.656 8554.213 ... 2413.174 2011.991 1725.496 hydro25:9 2004.365 2005.260 2012.629 ... 8613.818 8599.036 8572.635 hydro26:10 2009.410 1726.925 1719.924 ... 8580.326 8628.208 8572.554 hydro27:11 2408.632 2005.826 1719.207 ... 8343.862 8096.559 8150.015 hydro28:12 3017.299 2410.671 2009.843 ... 8581.544 8563.974 8559.905 hydro29:13 4006.374 3009.333 3000.386 ... 0.000 8618.282 8613.776 hydro30:14 5986.530 5978.050 3010.442 ... 8309.125 0.000 8592.298 hydro31:15 8591.207 5981.438 4002.751 ... 8618.887 8416.753 0.000 [16 rows x 16 columns] [SSH] completed [SSH] exit-status: 0 [workspace] $ /bin/sh -xe /tmp/jenkins2975711940737975292.sh + scp 'HYDRO_REMOTE:mpiGraph/*.txt' /var/lib/jenkins/jobs/mpiGraph/workspace + scp 'HYDRO_REMOTE:mpiGraph/*.png' /var/lib/jenkins/jobs/mpiGraph/workspace Finished: SUCCESS