Provided by: likwid_5.2.2+dfsg1-2_amd64 bug

NAME

       likwid-perfctr  -  configure  and  read  out hardware performance counters on x86, ARM and POWER CPUs and
       Nvidia GPUs

SYNOPSIS

       likwid-perfctr  [-vhHmaiefO]  [-c  core_list]  [-C  core_list_for_pinning]   [-g   performance_group   or
       performance_event_string]  [-t  timeline_frequency]  [-S monitoring_time] [-T group_switch_frequency] [-V
       verbosity] [-M access_mode]  [-o  output_file]  [-s  skip_mask]  [-E  search_str]  [-G  gpu_list(*)]  [-W
       gpu_performance_group or gpu_performance_event_string(*)] [--stats]

DESCRIPTION

       likwid-perfctr  is  a lightweight command line application to configure and read out hardware performance
       monitoring data on supported x86, ARM and POWER processors and Nvidia GPUs.  It  can  measure  either  as
       wrapper  without  changing  the  measured application or with marker API functions inside the code, which
       will turn on and off the counters.  There are preconfigured performance groups with useful event sets and
       derived metrics. Additionally, arbitrary events can be measured with custom event sets.  The  marker  API
       can  measure  multiple  named  regions  and  the  results  are  accumulated  over  multiple region calls.
       (*)OptiononlyavailableifbuiltwithNvidiaGPUsupport

OPTIONS

       -v, --version
              prints version information to standard output, then exits.

       -h, --help
              prints a help message to standard output, then exits.

       -H     prints group help message (use together with -g switch).

       -V <level>, --verbose <level>
              verbose output during execution for debugging. 0 for only errors, 1 for  informational  output,  2
              for detailed output and 3 for developer output

       -m, --marker
              run in marker API mode

       -a     print available performance groups for current processor, then exit.

       -e     print  available  counters  and  performance events of current processor and (if available) Nvidia
              GPUs.

       -o, --output <filename>
              store all output to a file instead of stdout. For the  filename  the  following  placeholders  are
              supported:  %j  for PBS_JOBID, %r for MPI RANK (only Intel MPI at the moment), %h host name and %p
              for process pid.  The placeholders must be separated by underscore as, e.g.,  -o  test_%h_%p.  You
              must  specify  a  suffix  to  the filename. For txt the output is printed as is to the file. Other
              suffixes trigger a filter on the output.  Available filters are csv (comma separated values),  xml
              and json at the moment.

       -O     print  output  in  CSV  format  (conform  to RFC 4180, see https://tools.ietf.org/html/rfc4180 for
              details).

       -i, --info
              print cpuid information about processor and about  Intel  Performance  Monitoring  features,  then
              exit.

       -c <cpu expression>
              specify  a  numerical list of processors. The list may contain multiple items, separated by comma,
              and ranges. For example 0,3,9-11.

       -C <cpu expression>
              specify a numerical list of processors. The list may contain multiple items, separated  by  comma,
              and  ranges.  For  example  0,3,9-11.  This  variant  will also pin the threads to the cores. Also
              logical numberings can be used.

       -g, --group <performance group> or <performance event set string>
              specify which performance group to measure. This can be one of the tags output with the  -a  flag.
              Also  a  custom event set can be specified by a comma separated list of events. Each event has the
              format eventId:register with the the register being one of a  architecture  supported  performance
              counter registers.

       -t <frequency of measurements>
              timeline  mode  for  time resolved measurements. The time unit must be given on command line, e.g.
              4s, 500ms or 900us.

       -S <waittime between measurements>
              End-to-end measurement using likwid-perfctr but sleep instead of  executing  an  application.  The
              time unit must be given on command line, e.g. 4s, 500ms or 900us.

       -T <time between group switches>
              Frequency  to  switch groups if multiple are given on commandline, default is 2s. Value is ignored
              for a single event set and default frequency of 30s is used to catch overflows. The time unit must
              be given on command line, e.g. 4s, 500ms or 900us.

       -s, --skip <mask>
              Specify skip mask as HEX number. For each set bit the corresponding thread is skipped.

       -f, --force
              Force writing of registers even if they are in use.

       -E <search_str>
              Print only events and corresponding counters matching <search_str>

       -G, --gpus <gpu_list>
              specify a numerical list of GPU IDs. The list may contain multiple items, separated by comma,  and
              ranges. For example 0,3,9-11.

       -W, --gpugroup <gpu performance group> or <gpu performance event set string>
              specify  which  performance group to measure on the specified Nvidia GPUs.  This can be one of the
              tags output with the -a flag in the GPU section.  Also a custom event set can be  specified  by  a
              comma  separated list of events. Each event has the format eventId:GPUx (x=0,1,2,...). You can add
              as many events to the string until you hit an error.

       --stats
              Always print statistics table

EXAMPLE

       Because likwid-perfctr measures on processors and not single applications it is necessary to ensure  that
       processes  and  threads are pinned to dedicated resources. You can either pin the application yourself or
       use the builtin pin functionality.

       1.  As wrapper with performance group:

       likwid-perfctr -C 0-2 -g TLB ./cacheBench -n 2 -l 1048576 -i 100 -t Stream

       The parent process is pinned to processor 0, Thread 0 to processor 1 and Thread 1 to processor 2.

       2.  As wrapper with custom event set on AMD:

       likwid-perfctr -C 0-4 -g INSTRUCTIONS_RETIRED_SSE:PMC0,CPU_CLOCKS_UNHALTED:PMC3 ./cacheBench

       It is specified that the event INSTRUCTIONS_RETIRED_SSE  is  measured  on  counter  PMC0  and  the  event
       CPU_CLOCKS_UNHALTED  on  counter PMC3.  It is possible calculate the run time of all threads based on the
       CPU_CLOCKS_UNHALTED event. If you want this you have to include this event in your custom event string as
       shown above.

       3.  As wrapper with custom event set on Intel:

       likwid-perfctr -C 0 -g INSTR_RETIRED_ANY:FIXC0,CPU_CLK_UNHALTED_CORE:FIXC1 ./stream-icc

       On Intel processors fixed events are measured on dedicated  counters.  These  are  INSTR_RETIRED_ANY  and
       CPU_CLK_UNHALTED_CORE.  If you configure these fixed counters, likwid-perfctr will calculate the run time
       and CPI metrics for your run.

       4.  Using  the marker API to measure only parts of your code (this can be used both with groups or custom
           event sets):

       likwid-perfctr -m -C 0-4 -g INSTRUCTIONS_RETIRED_SSE:PMC0,CPU_CLOCKS_UNHALTED:PMC3 ./cacheBench

       You have to link you code against liblikwid.so and use the marker API calls.  Examples can  be  found  in
       examples folder /usr/share/likwid/examples.  The following code snippet shows the necessary calls:

       #include <likwid-marker.h>

       /* only one thread calls init */
       LIKWID_MARKER_INIT;

       /* Can be called by each thread the should
        * perform measurements. It is only needed
        * if the pinning feature of LIKWID failed
        * and the threads need to be pinned explicitly.
        *
        * If you place it in the same parallel
        * region as LIKWID_MARKER_START, perform a
        * barrier between the statements to avoid
        * timing problems.
        */
       LIKWID_MARKER_THREADINIT;

       /* If you run the code region only once, register
        * the region tag previously to reduce the overhead
        * of START and STOP calls. Call it once for each
        * thread in parallel environment.
        * Note: No whitespace characters are allowed in the region tags
        * This call is optional but RECOMMENDED, START will do the same operations.
        */
       LIKWID_MARKER_REGISTER("name");

       /* Start measurement
        * Note: No whitespace characters are allowed in the region tags
        */
       LIKWID_MARKER_START("name");
       /*
        * Your code to be measured is here
        * You can also nest named regions
        * No whitespaces are allowed in the region names!
        */
       LIKWID_MARKER_STOP("name");

       /* If you want to measure multiple groups/event sets
        * Switches through groups in round-robin fashion
        */
       LIKWID_MARKER_SWITCH;

       /* If you want to get the data of a region inside your application
        * nevents in an (int*) and used as input length of the events array. After the
        * call, nevents contains the actual amount of events
        * events is an array of doubles (double*), time is a pointer to double to
        * retrieve the measured runtime of the region and count is a pointer to int
        * and is filled with the call count of the region.
        */
       LIKWID_MARKER_GET("name", nevents, events, time, count);

       /* If you want to reset the counts for a region
        */
       LIKWID_MARKER_RESET("name");

       /* Finally */
       LIKWID_MARKER_CLOSE;

       5.  Using likwid in timeline mode:

       likwid-perfctr -c 0-3 -g FLOPS_DP -t 300ms ./cacheBench > out.txt

       This  will  read  out  the counters every 300ms on physical hardware threads 0-3 and write the results to
       out.txt.  The application is not pinned to the CPUs. The output syntax of the timeline mode is for custom
       event sets:

       <groupID>  <numberOfEvents>   <numberOfThreads>   <Timestamp>   <Event1_Thread1>   <Event2_Thread1>   ...
       <Event1_Thread2> ... <EventN_ThreadM>

       For   performance   groups   with  metrics:  <groupID>  <numberOfMetrics>  <numberOfThreads>  <Timestamp>
       <Metric1_Thread1> <Metric2_Thread1> ... <Metric1_Thread2> ...<MetricN_ThreadM>

       For timeline mode there is a frontend application likwid-perfscope(1), which  enables  live  plotting  of
       selected  events.  Please  be  aware that with high frequencies (<100ms), the values differ from the real
       results but the behavior of them is valid.

       6.  Using likwid in stethoscope mode:

       likwid-perfctr -c 0-3 -g FLOPS_DP -S 2s

       This will start the counters and read them out after 2s on physical hardware threads 0-3  and  write  the
       results to stdout.

       7.  Using likwid with counter options:

       likwid-perfctr -c S0:1@S1:1 -g LLC_LOOKUPS_DATA_READ:CBOX0C0:STATE=0x9 ./cacheBench

       This  will  program  the  counter  CBOX0C0  (the  counter  0 of the LLC cache box 0) to measure the event
       LLC_LOOKUPS_DATA_READ and filter the increments by the state of a cacheline.  STATE=0x9  for  this  event
       means  all  <invalid>  and  <modified>  cachelines.  Which options are allowed for which box is listed in
       LIKWID's html documentation. The values  for  the  options  can  be  found  in  the  vendors  performance
       monitoring  documentations.  Likwid measures the first CPU of socket 0 and the first CPU of socket 1. See
       likwid-pin(1) for details regarding the cpu expressions.  For more code  examples  have  a  look  at  the
       likwid WIKI pages and LIKWID's html documentation.

       7.   Using  likwid  with GPU events and NvMarkerAPI. The CUDA library and CUPTI library must be reachable
            (path in LD_LIBRARY_PATH).

       likwid-perfctr -G 0,1 -W FLOPS_DP -m ./cudaApp

       This runs the application in NvMarkerAPI mode on GPUs 0 and 1 and measures  the  single-precision  flops.
       The NvMarkerAPI is similar to the CPU MarkerAPI (compile -DLIKWID_NVMON):

       #include <likwid-marker.h>

       /* Initialize the library and add configured eventset */ LIKWID_NVMARKER_INIT;

       /* If you run the code region only once, register
        * the region tag previously to reduce the overhead
        * of START and STOP calls. Call it before calling START() for
        * the region the first time.
        *
        * Place it around your CUDA kernel call.
        *
        * Note: No whitespace characters are allowed in the region tags
        * This call is optional but RECOMMENDED, START will do the same operations.
        */ LIKWID_NVMARKER_REGISTER("name");

       /* Start measurement on Nvidia GPUs
        * Note: No whitespace characters are allowed in the region tags
        */ LIKWID_NVMARKER_START("name"); /*
        * Your code to be measured is here
        * You can also nest named regions
        */

       /* Stop measurment on Nvidia GPUs
        * No whitespaces are allowed in the region names!
        */ LIKWID_NVMARKER_STOP("name");

       /* If you want to measure multiple groups/event sets
        * Switches through groups in round-robin fashion.
        */ LIKWID_NVMARKER_SWITCH;

       /* If you want to get the data of a region inside your application
        * nevents in an (int*) and used as input length of the events array. After the
        * call, nevents contains the actual amount of events. Same for ngpus.
        * events is an array of doubles (double*), time is a pointer to double to
        * retrieve the measured runtime of the region and count is a pointer to int
        * and is filled with the call count of the region.
        */ LIKWID_NVMARKER_GET("name", ngpus, nevents, events, time, count);

       /* If you want to reset the counts for a region
        */ LIKWID_NVMARKER_RESET("name");

       /* Finally */ LIKWID_NVMARKER_CLOSE;

AUTHOR

       Written by Thomas Gruber <thomas.roehl@googlemail.com>.

BUGS

       Report Bugs on <https://github.com/RRZE-HPC/likwid/issues>.

SEE ALSO

       likwid-topology(1), likwid-perfscope(1), likwid-pin(1), likwid-bench(1)

likwid-5                                           26.07.2022                                  LIKWID-PERFCTR(1)