vetter/shoc

Cosmetic differences in output format

Opened this issue · 1 comments

There are some minor differences in the format of the units that ought to be easy to fix:

 Running benchmark BusSpeedDownload
     result for bspeed_download:                 25.0373 GB/sec
 Running benchmark BusSpeedReadback
     result for bspeed_readback:                 26.3023 GB/sec
 Running benchmark DeviceMemory
     result for gmem_readbw:                   3098.5600 GB/s
     result for gmem_readbw_strided:            607.3330 GB/s
     result for gmem_writebw:                  2252.4600 GB/s
     result for gmem_writebw_strided:           161.8770 GB/s
     result for lmem_readbw:                  14329.3000 GB/s
     result for lmem_writebw:                 16107.4000 GB/s
     result for tex_readbw:                    1586.6800 GB/sec

where "GB/s" and "GB/sec" are both used, and similarly

 Running benchmark KernelCompile
     result for ocl_kernel:                       0.0876 sec
 Running benchmark QTC
     result for qtc:                              1.5015 s
     result for qtc_kernel:                       1.1845 s

I'd be inclined to use "sec" (and "msec") everywhere.

Another variation is in the order of output between the single-GPU case

 Running benchmark Spmv
     result for spmv_csr_scalar_sp:              98.8228 Gflop/s
     result for spmv_csr_scalar_sp_pcie:          3.1424 Gflop/s
     result for spmv_csr_scalar_dp:              90.6637 Gflop/s
     result for spmv_csr_scalar_dp_pcie:          2.0450 Gflop/s
     result for spmv_csr_scalar_pad_sp:          88.7928 Gflop/s
     result for spmv_csr_scalar_pad_sp_pcie:      3.3503 Gflop/s
     result for spmv_csr_scalar_pad_dp:          81.4095 Gflop/s
     result for spmv_csr_scalar_pad_dp_pcie:      2.8388 Gflop/s
     result for spmv_csr_vector_sp:             135.6520 Gflop/s
     result for spmv_csr_vector_sp_pcie:          3.1697 Gflop/s
     result for spmv_csr_vector_dp:             118.4620 Gflop/s
     result for spmv_csr_vector_dp_pcie:          2.0329 Gflop/s
     result for spmv_csr_vector_pad_sp:         146.2390 Gflop/s
     result for spmv_csr_vector_pad_sp_pcie:      3.3713 Gflop/s
     result for spmv_csr_vector_pad_dp:         133.8620 Gflop/s
     result for spmv_csr_vector_pad_dp_pcie:      2.8474 Gflop/s
     result for spmv_ellpackr_sp:               117.4200 Gflop/s
     result for spmv_ellpackr_dp:                93.4976 Gflop/s

and the multi-GPU case:

 Running benchmark Spmv
     result for spmv_csr_scalar_sp:             101.0680 Gflop/s
     result for spmv_csr_vector_sp:             140.4280 Gflop/s
     result for spmv_ellpackr_sp:               119.4360 Gflop/s
     result for spmv_csr_scalar_dp:              91.7853 Gflop/s
     result for spmv_csr_vector_dp:             121.0850 Gflop/s
     result for spmv_ellpackr_dp:                94.9549 Gflop/s

I can see that the former splits the list by scalar/vector and the latter splits it by single/double.
Maybe that's more intuitive for reading, but it also makes it harder to line up the results.
I'd be inclined to order the latter as

 Running benchmark Spmv
     result for spmv_csr_scalar_sp:             101.0680 Gflop/s
     result for spmv_csr_scalar_dp:              91.7853 Gflop/s
     result for spmv_csr_vector_sp:             140.4280 Gflop/s
     result for spmv_csr_vector_dp:             121.0850 Gflop/s
     result for spmv_ellpackr_sp:               119.4360 Gflop/s
     result for spmv_ellpackr_dp:                94.9549 Gflop/s