.. _debugging_profiling: Debugging and Profiling Tools ------------------------------ In software development sooner or later one experiences that written software does not always behave as intended, or with the performance as expected. This is were debugging and profiling tools come to hand that help understand what software is actually doing, or identify bottlenecks in the program execution. However, not only software developers, also users of software written by others may benefit from debugging and profiling tools. Overview of debugging and profiling tools ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ In the following various debugging and profiling tools are described that may be of help when developing or working with |FOURC|. - Get MPI collective timer statistics with a :ref:`time monitor ` from ``Trilinos`` package ``Teuchos`` - Code profiling of |FOURC| with :ref:`callgrind ` - Debugging/MPI debugging with :ref:`VS Code ` Useful options for Debugging with gdb (or your IDE) ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ Build debug version """"""""""""""""""" Create a directory, where you want to build your debug version and build a |FOURC| with debug flag using the correct preset. This should contain:: { "name": "debug", "displayName": "Debug build for a general workstation", "binaryDir": "<4C-debug-execdir>", "generator": "Ninja", "inherits": [ "lnm_workstation" ], "cacheVariables": { "CMAKE_BUILD_TYPE": "DEBUG" } } Pretty printing of the Standard Library """""""""""""""""""""""""""""""""""""""""" Add the following to your `~/.gdbinit` :: python import sys sys.path.insert(0, '/usr/share/gcc/python') from libstdcxx.v6.printers import register_libstdcxx_printers register_libstdcxx_printers (None) end enable pretty-printer Useful settings for MPI Debugging """""""""""""""""""""""""""""""""""""""""" **"Standard" parallel errors** In this mode, all processes are paused once one process hits a debug event:: set detach-on-fork off set schedule-multiple on **Tracking down race conditions** If you have to track down race conditions, you need manual control over each process. You can start attach gdb to each process after it has already started. Start |FOURC| via :: mpirun -np 2 ./4C ../path/to/input.dat ../path/to/output --interactive The process id of each mpi process is being displayed. Once all gdb instances are connected, you can press any key to start the execution. **gdb + valgrind to track down memory leaks / invalid reads/writes** :: valgrind --tool=memcheck -q --vgdb-error=0 ./4C path/to/input.dat path/to/output Then, you can connect with a gdb debugger: :: gdb ./4C (gdb) target remote | bgdb --pid= Or visually using VS Code: Add a configuration in `launch.json`:: { "name": "(gdb) Attach to valgrind", "type": "cppdbg", "request": "launch", "program": "<4C-debug-execdir>", "targetArchitecture": "x64", "customLaunchSetupCommands": [ { "description": "Enable pretty-printing for gdb", "text": "-enable-pretty-printing", "ignoreFailures": true }, { "description": "Attach to valgrind", "text": "target remote | vgdb --pid=", "ignoreFailures": false } ], "stopAtEntry": false, "cwd": "/path/to/run/", "environment": [], "externalConsole": false, "MIMode": "gdb" } If you need to run it in combination with mpirun, start it with :: mpirun -np 2 valgrind --tool=memcheck -q --vgdb-error=0 ./4C path/to/input.dat path/to/output and connect to each process individually. .. _profiling_callgrind: Code profiling with ``callgrind`` ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ "Callgrind is a profiling tool that records the call history among functions in a program's run as a call-graph. By default, the collected data consists of the number of instructions executed, their relationship to source lines, the caller/callee relationship between functions, and the numbers of such calls." (from `callgrind `_) Configure and build for profiling ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ **Note:** For general information about configuring and building of |FOURC| refer to :ref:`Configure and Build <4Cinstallation>` and the ``README.md``. You probably want to configure with `CMAKE_BUILD_TYPE` set to `RELWITHDEBINFO`. This results in a release version of the |FOURC| build with additional per-line annotations. That way, when examining the results one can see the exact lines of code where computation time is spent. .. note:: * Beware that code gets inlined with the profiling build of |FOURC| and hot spots might appear within the inlined section. * The debug version of |FOURC| also contains per-line annotations but without the effect of inlining and can thus also be used to profile |FOURC|. However, the debug version is compiled without compiler optimizations and thus does not give a representative view of hot spots. * For a quick profiling without per-line annotations also the release version can be used. This already gives a nice overview of computationally expensive methods. Run simulation with `valgrind` ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ Run a |FOURC| simulation with ``valgrind`` in parallel using the command:: mpirun -np valgrind --tool=callgrind /<4C-execdir>/4C In addition to the usual |FOURC| output, ``valgrind`` writes output for each mpi rank in the files ``callgrind.out.``. .. note:: - For profiling a simulation in serial execute:: valgrind --tool=callgrind /<4C-execdir>/4C - It is also possible to examine the post processing of result files, simply wrap the corresponding command:: mpirun -np valgrind --tool=callgrind - Wrapping the |FOURC| simulation using ``valgrind`` increases the runtime by a factor of about 100. Therefore, to reduce the total wall time think about running only a few time steps of your |FOURC| simulation. Depending on the problem type it might be reasonable to do this after a restart in order to examine characteristic parts. Follow the steps as described below:: mpirun -np /<4C-execdir>/4C mpirun -np valgrind --tool=callgrind /<4C-execdir>/4C restart= Examine results with ``kcachegrind`` ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ Using `kcachegrind` (refer to `kcachegrind `_ for documentation and download) the output can be visualized via:: kcachegrind callgrind.out.* It is also possible to only open the output of a specific mpi rank with processor id via:: kcachegrind callgrind.out. **Note:** Be sure to check out the |FOURC| version the code is compiled with in your local git repo to make use of the per-line annotations. **Example:** In the figure below a screenshot of `kcachegrind` is given where the profiling output of a Smoothed Particle Hydrodynamics (SPH) simulation is visualized. .. figure:: /_assets/kcachegrind.png :alt: Picture of kcachegrind :width: 100% .. _teuchos-time-monitor: Teuchos Time Monitor ~~~~~~~~~~~~~~~~~~~~ The ``TimeMonitor`` from ``Trilinos`` package ``Teuchos`` provides MPI collective timer statistics. Refer to the ``Teuchos::TimeMonitor`` Class Reference https://trilinos.org/docs/dev/packages/teuchos/doc/html/classTeuchos_1_1TimeMonitor.html for a detailed documentation. Add a timer for a method in |FOURC| ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ In order to get parallel timing statistics of a method in |FOURC| include the following header :: #include in the ``.cpp``-file that contains the implementation of the method and add the macro ``TEUCHOS_FUNC_TIME_MONITOR`` with the name of the method in the implementation of the method: .. code-block:: cpp void ::(...) { TEUCHOS_FUNC_TIME_MONITOR("::"); /* implementation */ } Running a simulation on 3 processors for example yields the following ``TimeMonitor`` output in the terminal: :: ============================================================================================================ TimeMonitor results over 3 processors Timer Name MinOverProcs MeanOverProcs MaxOverProcs MeanOverCallCounts ------------------------------------------------------------------------------------------------------------ :: 0.1282 (1000) 0.2134 (1000) 0.2562 (1000) 0.0002132 (1001) ============================================================================================================ The output gives the minimum, maximum, and mean ``execution time (number of counts)`` for all processors and also the mean execution time over all counts. **Note:** The ``TimeMonitor`` output of a |FOURC| simulation in general already contains a variety of methods that are monitored, meaning there is a line with timings for each method in the output. How to interpret the output of the ``TimeMonitor`` ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ Examining the output of the ``TimeMonitor`` is probably one of the easiest steps in profiling the behaviour of a program. The information given in the output of the `TimeMonitor` may server for - getting an idea of the execution time of a method - knowing how often a method is called during the runtime of the program - investigating the parallel load balancing of a method (compare ``MinOverProcs`` with ``MaxOverProcs``) and thereby helps identifying bottlenecks in the overall program.