Started by user arnoldg Running as SYSTEM Building in workspace /var/lib/jenkins/jobs/mpiGraph/workspace [SSH] script: PATH="/sw/tools/bin:/sw/tools/bin:$PATH" NODELIST="""" ACCOUNT="bbmb-hydro" PARTITION="sandybridge" NODES="16" # tar xvf /var/tmp/mpiGraph.tar # galen's copy of the benchmark cd mpiGraph export PATH=~/mpiGraph:$PATH # if the NODELIST parameter was filled in, proceed with a targeted test if [[ ! -z ${NODELIST} ]] then srun --mem=16g --nodes=${NODES} --ntasks-per-node=1 --time=00:10:00 --job-name=mpiGraph \ --partition=${PARTITION} --nodelist=${NODELIST} \ --account=${ACCOUNT} ${GPUS} \ mpiGraph 1048576 1000 10 > mpiGraph.out if [ "${NODES}" == "16" ] then bash clip-plot.sh 16 fi exit fi # otherwise, if NODELIST is "", run the default test on 1st avail nodes srun --nodes=${NODES} \ --mem=16g --ntasks-per-node=1 --time=00:10:00 --job-name=mpiGraph \ --account=${ACCOUNT} \ --partition=${PARTITION} \ mpiGraph 1048576 1000 10 > mpiGraph.out if [ "${NODES}" == "16" ] then bash clip-plot.sh 16 fi [SSH] executing... NODECOUNT=$1 + NODECOUNT=16 # clip out the Send table and plot it egrep --after-context=$NODECOUNT "^Send\s+hy" mpiGraph.out \ | sed -e s/.to//g > send.txt + egrep --after-context=16 '^Send\s+hy' mpiGraph.out + sed -e s/.to//g python3 send.py + python3 send.py hydro15:0 hydro16:1 hydro17:2 ... hydro29:13 hydro30:14 hydro31:15 Send ... hydro15:0 0.000 8492.943 8550.089 ... 3995.568 5972.168 8568.378 hydro16:1 8563.595 0.000 8402.924 ... 3004.263 5922.388 5974.525 hydro17:2 8462.514 8602.585 0.000 ... 2950.011 3005.911 3995.572 hydro18:3 8366.948 8459.843 8577.795 ... 2948.602 2404.967 3003.419 hydro19:4 7775.652 7830.982 8241.859 ... 1978.263 2002.801 2990.642 hydro20:5 8306.475 8400.933 8371.697 ... 1974.592 1717.998 2981.969 hydro22:6 8224.366 8292.805 8402.107 ... 1985.553 1717.297 1986.606 hydro23:7 8284.837 8194.705 8360.968 ... 2937.920 1719.182 1977.998 hydro24:8 8539.058 8205.040 8217.734 ... 2949.259 2003.191 1971.576 hydro25:9 1722.957 1722.666 1722.002 ... 8375.612 8293.746 8256.412 hydro26:10 2004.968 1721.553 1718.891 ... 8411.067 8418.013 8316.341 hydro27:11 2405.476 2009.315 1723.037 ... 8518.307 8398.009 8404.712 hydro28:12 3000.081 2405.920 2007.082 ... 8395.127 8440.804 8358.955 hydro29:13 3996.446 3004.371 2408.570 ... 0.000 8339.419 8455.672 hydro30:14 5962.031 4001.335 3002.006 ... 8598.197 0.000 8422.048 hydro31:15 8302.936 5964.647 3998.637 ... 8590.340 8600.242 0.000 [16 rows x 16 columns] # clip out the Recv table and plot it egrep --after-context=$NODECOUNT "^Recv\s+hy" mpiGraph.out \ | sed -e s/.from//g > recv.txt + egrep --after-context=16 '^Recv\s+hy' mpiGraph.out + sed -e s/.from//g python3 recv.py + python3 recv.py hydro15:0 hydro16:1 hydro17:2 ... hydro29:13 hydro30:14 hydro31:15 Recv ... hydro15:0 0.000 8576.936 8548.791 ... 4001.145 5967.796 8316.367 hydro16:1 8513.272 0.000 8605.122 ... 3007.595 4011.551 5983.323 hydro17:2 8582.617 8412.622 0.000 ... 2411.684 3009.560 4005.119 hydro18:3 8531.558 8570.270 8411.158 ... 2010.963 2412.915 3008.686 hydro19:4 8568.221 8500.916 8454.043 ... 1719.803 2005.002 2410.588 hydro20:5 8543.081 8562.137 8525.889 ... 1723.024 1726.895 2015.797 hydro22:6 8565.664 8536.855 8590.681 ... 1726.758 1726.222 1725.127 hydro23:7 8529.006 8521.478 8538.104 ... 2013.103 1728.103 1725.230 hydro24:8 8556.257 8568.621 8571.775 ... 2414.554 2012.272 1728.748 hydro25:9 2003.489 2005.861 2011.621 ... 8615.139 8594.933 8578.429 hydro26:10 2009.964 1722.750 1719.542 ... 8576.885 8627.782 8589.494 hydro27:11 2411.910 2007.827 1722.011 ... 8584.842 8581.956 8621.457 hydro28:12 3017.653 2407.996 2008.777 ... 8600.714 8571.816 8558.809 hydro29:13 4003.639 3008.347 2997.398 ... 0.000 8599.328 8608.497 hydro30:14 5991.276 5974.379 3011.992 ... 8351.663 0.000 8617.726 hydro31:15 8565.603 5990.739 4000.803 ... 8543.414 8437.211 0.000 [16 rows x 16 columns] [SSH] completed [SSH] exit-status: 0 [workspace] $ /bin/sh -xe /tmp/jenkins2865828138345476601.sh + scp 'HYDRO_REMOTE:mpiGraph/*.txt' /var/lib/jenkins/jobs/mpiGraph/workspace + scp 'HYDRO_REMOTE:mpiGraph/*.png' /var/lib/jenkins/jobs/mpiGraph/workspace Finished: SUCCESS