summaryrefslogtreecommitdiffstats
path: root/tde-i18n-en_GB/docs/tdesdk/tdecachegrind/index.docbook
blob: 443be3d40cde508b17345f0038bc55d13b602999 (plain)
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
418
419
420
421
422
423
424
425
426
427
428
429
430
431
432
433
434
435
436
437
438
439
440
441
442
443
444
445
446
447
448
449
450
451
452
453
454
455
456
457
458
459
460
461
462
463
464
465
466
467
468
469
470
471
472
473
474
475
476
477
478
479
480
481
482
483
484
485
486
487
488
489
490
491
492
493
494
495
496
497
498
499
500
501
502
503
504
505
506
507
508
509
510
511
512
513
514
515
516
517
518
519
520
521
522
523
524
525
526
527
528
529
530
531
532
533
534
535
536
537
538
539
540
541
542
543
544
545
546
547
548
549
550
551
552
553
554
555
556
557
558
559
560
561
562
563
564
565
566
567
568
569
570
571
572
573
574
575
576
577
578
579
580
581
582
583
584
585
586
587
588
589
590
591
592
593
594
595
596
597
598
599
600
601
602
603
604
605
606
607
608
609
610
611
612
613
614
615
616
617
618
619
620
621
622
623
624
625
626
627
628
629
630
631
632
633
634
635
636
637
638
639
640
641
642
643
644
645
646
647
648
649
650
651
652
653
654
655
656
657
658
659
660
661
662
663
664
665
666
667
668
669
670
671
672
673
674
675
676
677
678
679
680
681
682
683
684
685
686
687
688
689
690
691
692
693
694
695
696
697
698
699
700
701
702
703
704
705
706
707
708
709
710
711
712
713
714
715
716
717
718
719
720
721
722
723
724
725
726
727
728
729
730
731
732
733
734
735
736
737
738
739
740
741
742
743
744
745
746
747
748
749
750
751
752
753
754
755
756
757
758
759
760
761
762
763
764
765
766
767
768
769
770
771
772
773
774
775
776
777
778
779
780
781
782
783
784
785
786
787
788
789
790
791
792
793
794
795
796
797
798
799
800
801
802
803
804
805
806
807
808
809
810
811
812
813
814
815
816
817
818
819
820
821
822
823
824
825
826
827
828
829
830
831
832
833
834
835
836
837
838
839
840
841
842
843
844
845
846
847
848
849
850
851
852
853
854
855
856
857
858
859
860
861
862
863
864
865
866
867
868
869
870
871
872
873
874
875
876
877
878
879
880
881
882
883
884
885
886
887
888
889
890
891
892
893
894
895
896
897
898
899
900
901
902
903
904
905
906
907
908
909
910
911
912
913
914
915
916
917
918
919
920
921
922
923
924
925
926
927
928
929
930
931
932
933
934
935
936
937
938
939
940
941
942
943
944
945
946
947
948
949
950
951
952
953
954
955
956
957
958
959
960
961
962
963
964
965
966
967
968
969
970
971
972
973
974
975
976
977
978
979
980
981
982
983
984
985
986
987
988
989
990
991
992
993
994
995
996
997
998
999
1000
1001
1002
1003
1004
1005
1006
1007
1008
1009
1010
1011
1012
1013
1014
1015
1016
1017
1018
1019
1020
1021
1022
1023
1024
1025
1026
1027
1028
1029
1030
1031
1032
1033
1034
<?xml version="1.0" ?>
<!DOCTYPE book PUBLIC "-//KDE//DTD DocBook XML V4.2-Based Variant V1.1//EN" "dtd/kdex.dtd" [
  <!ENTITY tdecachegrind '<application
>KCachegrind</application
>'>
  <!ENTITY cachegrind "<application
>Cachegrind</application
>">
  <!ENTITY calltree "<application
>Calltree</application
>">
  <!ENTITY callgrind "<application
>Callgrind</application
>">
  <!ENTITY valgrind "<application
>Valgrind</application
>">
  <!ENTITY oprofile "<application
>OProfile</application
>">
  <!ENTITY kappname "&tdecachegrind;">
  <!ENTITY package "tdesdk">
  <!ENTITY % addindex "IGNORE">
  <!ENTITY % British-English "INCLUDE">
]>

<!-- ................................................................ -->

<book lang="&language;">

<bookinfo>
<title
>The &tdecachegrind; Handbook</title>

<authorgroup>
<author
><firstname
>Josef</firstname
> <surname
>Weidendorfer</surname
> <affiliation
> <address
><email
>Josef.Weidendorfer@gmx.de</email
></address>
</affiliation>
</author>

<othercredit role="translator"
><firstname
>Andrew</firstname
><surname
>Coles</surname
><affiliation
><address
><email
>andrew_coles@yahoo.co.uk</email
></address
></affiliation
><contrib
>Conversion to British English</contrib
></othercredit
> 

</authorgroup>

<copyright>
<year
>2002-2004</year>
<holder
>Josef Weidendorfer</holder>
</copyright>
<legalnotice
>&FDLNotice;</legalnotice>

<date
>2004-07-27</date>
<releaseinfo
>0.4.6</releaseinfo>

<abstract>
<para
>&tdecachegrind; is a profile data visualisation tool, written using the &kde; environment. </para>
</abstract>

<keywordset>
<keyword
>KDE</keyword>
<keyword
>tdesdk</keyword>
<keyword
>Cachegrind</keyword>
<keyword
>Callgrind</keyword>
<keyword
>Valgrind</keyword>
<keyword
>Profiling</keyword>
</keywordset>

</bookinfo>


<chapter id="introduction">
<title
>Introduction</title>

<para
>&kappname; is a browser for data produced by profiling tools. This chapter explains what profiling is for, how it is done, and gives some examples of profiling tools available. </para>

<sect1 id="introduction-profiling">
<title
>Profiling</title>

<para
>When developing a program, one of the last steps often involves performance optimisations. As it makes no sense to optimise functions rarely used, because that would be a waste of time, one needs to know in which part of a program most of the time is spent. </para>

<para
>For sequential code, collecting statistical data of the programs runtime characteristic like time numbers spent in functions and code lines usually is enough. This is called Profiling. The program is run under control of a profiling tool, which gives the summary of an execution run at the end. In contrast, for parallel code, performance problems typically are caused when one processor is waiting for data from another. As this waiting time usually can not easily attributed, here it is better to generate timestamped event traces. KCachegrind can not visualise this kind of data. </para>

<para
>After analysing the produced profile data, it should be easy to see the hot spots and bottlenecks of the code: for example, assumptions about call counts can be checked, and identified code regions can be optimised. Afterwards, the success of the optimisation should be verified with another profile run. </para>
</sect1>

<sect1 id="introduction-methods">
<title
>Profiling Methods</title>

<para
>To exactly measure the time passed or record the events happening during the execution of a code region (e.g. a function), additional measurement code needs to be inserted before and after the given region. This code reads the time, or a global event count, and calculates differences. Thus, the original code has to be changed before execution. This is called instrumentation. Instrumentation can be done by the programmer itself, the compiler, or by the runtime system. As interesting regions usually are nested, the overhead of measurement always influences the measurement itself. Thus, instrumentation should be done selectively and results have to be interpreted with care. Of course, this makes performance analysis by exact measurement a very complex process.</para>

<para
>Exact measurement is possible because of hardware counters (including counters incrementing on a time tick) provided in modern processors, which are incremented whenever an event is happening. As we want to attribute events to code regions, without the counters, we would have to handle every event by incrementing a counter for the current code region ourself. Doing this in software is, of course, not possible; but, on the assumption that the event distribution over source code is similar when looking only at every n-th event instead of every event, a measurement method whose overhead is tunable has been developed: it is called Sampling. Time Based Sampling (TBS) uses a timer to regularly look at the program counter to create a histogram over the program code. Event Based Sampling (EBS) exploits the hardware counters of modern processors, and uses a mode where an interrupt handler is called on counter underflow to generate a histogram of the corresponding event distribution: in the handler, the event counter is always reinitialised to the 'n' of the sampling method. The advantage of sampling is that the code does not have to be changed, but it is still a compromise: the above assumption will be more correct if n is small, but the smaller the n, the higher the overhead of the interrupt handler.</para>

<para
>Another measurement method is to simulate things happening in the computer system when executing a given code, i.e. execution driven simulation. The simulation is always derived from a more or less accurate machine model; however, with very detailed machine models, giving very close approximations to reality, the simulation time can be unacceptably high in practise. The advantage of simulation is that arbitrarily complex measurement/simulation code can be inserted in a given code without perturbing results. Doing this directly before execution (called runtime instrumentation), using the original binary, is very comfortable for the user: no re-compilation is necessary. Simulation becomes usable when simulating only parts of a machine with a simple model; another advantage is that the results produced by simple models are often easier to understand: often, the problem with real hardware is that results include overlapping effects from different parts of the machine.</para>
</sect1>

<sect1 id="introduction-tools">
<title
>Profiling Tools</title>

<para
>Most known is the GCC profiling tool <application
>gprof</application
>: One needs to compile the program with option <option
>-pg</option
>; running the program generates a file <filename
>gmon.out</filename
>, which can be transformed into human-readable form with <command
>gprof</command
>. One disadvantage is the needed re-compilation step to prepare the executable, which has to be statically linked. The method used here is compiler-generated instrumention - which measures call arcs happening among functions and corresponding call counts - in conjunction with TBS - which gives a histogram of time distribution over the code. Using both pieces of information, it is possible to heuristically calculate inclusive time of functions, i.e. time spent in a function together with all functions called from it. </para>

<para
>For exact measurement of events happening, libraries exist with functions able to read out hardware performance counters. Most known here is the PerfCtr patch for Linux, and the architecture independent libraries PAPI and PCL. Still, exact measurement needs instrumentation of code, as stated above. Either one uses the libraries itself or uses automatic instrumentation systems like ADAPTOR (for FORTRAN source instrumentation) or DynaProf (code injection via DynInst).</para>

<para
>&oprofile; is a system-wide profiling tool for Linux using Sampling.</para>

<para
>In many aspects, a comfortable way of Profiling is using Cachegrind or Callgrind, which are simulators using the runtime instrumentation framework &valgrind;. Because there is no need to access hardware counters (often difficult with today's Linux installations), and binaries to be profiled can be left unmodified, it is a good alternative to other profiling tools. The disadvantage of simulation - slowdown - can be reduced by doing the simulation on only the interesting program parts, and perhaps only on a few iterations of a loop. Without measurement/simulation instrumentation, Valgrind's usage only has a slowdown factor in the range of 3 to 5. Also, when only the call graph and call counts are of interest, the cache simulator can be switched off. </para>

<para
>Cache simulation is the first step in approximating real times; as ,on modern systems, runtime is very sensitive to the exploitation of so called caches (small and fast buffers which accelerate repeated accesses to the same main memory cells.) &cachegrind; does cache simulation by catching memory accesses. The data produced includes the number of instruction/data memory accesses and 1st/2nd level cache misses, and relates it to source lines and functions of the run program. By combining these miss counts, using miss latencies from typical processors, an estimation of spent time can be given. </para>

<para
>Callgrind is an extension of &cachegrind; that builds up the call graph of a program on-the-fly, &ie; how the functions call each other and how many events happen while running a function. Also, the profile data to be collected can separated by threads and call chain contexts. It can provide profiling data on an instruction level to allow for annotation of disassembled code. </para>
</sect1>

<sect1 id="introduction-visualization">
<title
>Visualisation</title>

<para
>Profiling tools typically produce a large amount of data. The wish to easily browse down and up the call graph, together with fast switching of the sorting mode of functions and display of different event types, motivates a GUI application to accomplish this task. </para>

<para
>&kappname; is an visualisation for profile data fulfilling these wishes. Despite being programmed first with browsing the data from &cachegrind; and &calltree; in mind, there are converters available to be able to display profile data produced by other tools. In the appendix, a description of the Cachegrind/Callgrind file format is given. </para>

<para
>Besides a list of functions sorted according exclusive or inclusive cost metrics, and optionally grouped by source file, shared library or C++ class, &kappname; features various visualisation views for a selected function, namely <itemizedlist>
<listitem
><para
>a call-graph view, which shows a section of the call graph around the selected function,</para>
</listitem>
<listitem
><para
>a tree-map view, which allows nested-call relations to be visualised, together with inclusive cost metric for fast visual detection of problematic functions,</para>
</listitem>
<listitem
><para
>source code and disassembler annotation views, allowing to see details of cost related to source lines and assembler instructions.</para>
</listitem>
</itemizedlist>

</para>
</sect1>
</chapter>

<chapter id="using-tdecachegrind">
<title
>Using &tdecachegrind;</title>

<sect1 id="using-profile">
<title
>Generate Data to Visualise</title>

<para
>First, one wants to generate performance data by measuring aspects of the runtime characteristics of an application, using a profiling tool. &tdecachegrind; itself does not include any profiling tool, but is good in being used together with &callgrind;, and by using a converter, also can be used to visualise data produced with &oprofile;. Although the scope of this manual is not to document profiling with these tools, the next section provides short quickstart tutorials to get you started. </para>

<sect2>
<title
>&callgrind;</title>

<para
>&callgrind; is available from <ulink url="http://tdecachegrind.sf.net"
> http://tdecachegrind.sf.net</ulink
>. Note that it previously was called &calltree;, but that name was misleading. </para>

<para
>Most common use is to prefix the command line to start your application with <application
>callgrind</application
>, like in <blockquote
><para
><command
>callgrind myprogram myargs</command
></para
></blockquote
> At program termination, a file <filename
>callgrind.out.pid</filename
> will be generated which can be loaded into &tdecachegrind;. </para>

<para
>More advanced use is to dump out profile data whenever a given function of your application is called. E.g. for <command
>konqueror</command
>, to see profile data only for rendering a web page, you could decide to dump the data whenever you select the menu item View/Reload. This corresponds to a call to <symbol
>KonqMainWindow::slotReload</symbol
>. Use <blockquote
><para
><command
> callgrind --dump-before=KonqMainWindow::slotReload konqueror </command
></para
></blockquote
> This will produce multiple profile data files with an additional sequential number at the end of the filename. A file without such an number at the end (only ending in the process PID) will also be produced; by loading this file into &tdecachegrind;, all others are loaded too, and can be seen in the Parts Overview and Parts list. </para>

</sect2>

<sect2>
<title
>&oprofile;</title>

<para
>&oprofile; is available from <ulink url="http://oprofile.sf.net"
> http://oprofile.sf.net</ulink
>. Follow the installation instructions on the web site; but, before you do, check if your distribution does not already provide it as package (like SuSE). </para>

<para
>System-wide profiling is only permitted to the root user, as all actions on the system can be observed; therefore, the following has to be done as root. First, configure the profiling process, using the GUI <command
>oprof_start</command
> or the command-line tool opcontrol. Standard configuration should be timer mode (TBS, see introduction). To start the measurement, run <command
>opcontrol -s</command
>. Then run the application you are interested in and, afterwards, do a <command
>opcontrol -d</command
>. This will write out the measurement results into files under directory <filename
>/var/lib/oprofile/samples/</filename
>. To be able to visualise the data in &tdecachegrind;, do in an empty directory: <blockquote
><para
><command
> opreport -gdf | op2callgrind </command
></para
></blockquote
> This will produce a lot of files, one for every program which was running on the system. Each one can be loaded into &tdecachegrind; on its own. </para>

</sect2>
</sect1>

<sect1 id="using-basics">
<title
>User Interface Basics</title>

<para
>When starting &tdecachegrind; with a profile data file as argument, or after loading one with File/Open, you will see a sidebar containing the function list at the left; and, on the right the main part, an area with visualisations for a selected function. This visualisation area can be arbitrarily configured to show multiple visualisations at once. </para>

<para
>At first start, this area will be divided into a top and a bottom part, each with different visualisations selectable by tabs. To move visualisation views, use the context menu of the tabs, and adjust the splitters between visualisations. To quickly switch between different visualisation layouts, use View/Layouts/Duplicate, change the layout and switch between layouts with View/Layout/Next (or, even better, use the corresponding keyboard shortcuts). </para>

<para
>The active event type is important for visualisation: for &callgrind;, this is, for example, Cache Misses or Cycle Estimation; for &oprofile;, this is "Timer" in the simplest case. You can change the event type via a combobox in the toolbar or in the Event Type view. A first overview of the runtime characteristics should be given when you select function <symbol
>main</symbol
> in the left list, and look at the call graph visualisation; there, you see calls happening in your program. Note that the call graph view only shows functions with high event count. By double-clicking a function in the graph, it will change to show the called functions around the selected one. </para>

<para
>To explore the GUI further, in addition to this manual, also have a look at the documentation section on the web site <ulink url="http://tdecachegrind.sf.net"
> http://tdecachegrind.sf.net</ulink
>. Also, every widget in &tdecachegrind; has <quote
>What's this</quote
> help. </para>
</sect1>

</chapter>


<chapter id="tdecachegrind-concepts">
<title
>Basic Concepts</title>

<para
>This chapter explains some concepts of the &tdecachegrind;, and introduces terms used in the interface. </para>

<sect1 id="concepts-model">
<title
>The Data Model for Profile Data</title>

<sect2>
<title
>Cost Entities</title>

<para
>Cost counts of event types (like L2 Misses) are attributed to cost entities, which are items with relationship to source code or data structures of a given program. Cost entities not only can be simple code or data positions, but also position tuples. For example, a call has a source and a target, or a data address can have a data type and an code position where its allocation happened. </para>

<para
>The cost entities known to KCachegrind are given in the following. Simple Positions: <itemizedlist
> <listitem
><para
> Instruction. An assembler instruction at a specified address. </para
></listitem
> <listitem
><para
> Source Line of a Function. All instructions that the compiler (via debug information) maps to a given source line specified by source file name and line number, and which are executed in the context of some function. The latter is needed because a source line inside of an inlined function can appear in the context of multiple functions. Instructions without any mapping to an actual source line are mapped to line number 0 in file "???". </para
></listitem
> <listitem
><para
> Function. All source lines of a given function make up the function itself. A function is specified by its name and its location in some binary object if available. The latter is needed because binary objects of a single program each can hold functions with the same name (these can be accessed e.g. with dlopen/dlsym; the runtime linker resolves functions in a given search order of binary objects used). If a profiling tool can not detect the symbol name of a function, e.g. because debug information is not available, either the address of the first executed instruction typically is used, or "???". </para
></listitem
> <listitem
><para
> Binary Object. All functions whose code is inside the range of a given binary object, either the main executable or a shared library. </para
></listitem
> <listitem
><para
> Source File. All functions whose first instruction is mapped to a line of the given source file. </para
></listitem
> <listitem
><para
> Class. Symbol names of functions typically are hierarchically ordered in name spaces, e.g. C++ namespaces, or classes of object oriented languages; thus, a class can hold functions of the class or embedded classes itself. </para
></listitem
> <listitem
><para
> Profile Part. Some time section of a profile run, with a given thread ID, process ID, and command line executed. </para
></listitem
> </itemizedlist
> As can be seen from the list, a set of cost entities often defines another cost entity; thus, there is a inclusion hierarchy of cost entities which should be obvious from the description above. </para>

<para
>Positions tuples: <itemizedlist
> <listitem
><para
> Call from instruction address to target function. </para
></listitem
> <listitem
><para
> Call from source line to target function. </para
></listitem
> <listitem
><para
> Call from source function to target function. </para
></listitem
> <listitem
><para
> (Un)conditional Jump from source to target instruction. </para
></listitem
> <listitem
><para
> (Un)conditional Jump from source to target line. </para
></listitem
> </itemizedlist
> Jumps between functions are not allowed, as this makes no sense in a call graph; thus, constructs like exception handling and long jumps in C have to be translated to popping the call stack as needed. </para>

</sect2>



<sect2>
<title
>Event Types</title>

<para
>Arbitrary event types can be specified in the profile data by giving them a name. Their cost related to a cost entity is a 64-bit integer. </para>
<para
>Event types whose costs are specified in a profile data file are called real events. Additionally, one can specify formulae for event types calculated from real events, which are called inherited events. </para>
</sect2>

</sect1>

<sect1 id="concepts-state">
<title
>Visualisation State</title>

<para
>The Visualisation state of a KCachegrind window includes: <itemizedlist
> <listitem
><para
> the primary and secondary event type chosen for display, </para
></listitem
> <listitem
><para
> the function grouping (used in the Function Profile list and entity colouring), </para
></listitem
> <listitem
><para
> the profile parts whose costs are to be included in visualisation, </para
></listitem
> <listitem
><para
> an active cost entity (e.g. a function selected from the function profile dockable), </para
></listitem
> <listitem
><para
> a selected cost entity. </para
></listitem
> </itemizedlist
> This state influences visualisations. </para>
<para
>Visualisations always are shown for one, the active, cost entity. When a given visualisation is not appropriate for a cost entity, it is disabled (e.g. when selecting an ELF object in the group list by double-clicking, source annotation for an ELF object make no sense). </para>
<para
>For example, for an active function, the callee list shows all the functions called from the active one: one can select one of these functions without making it active; also, if the call-graph is shown nearside, it will automatically select the same function. </para>

</sect1>

<sect1 id="concepts-guiparts">
<title
>Parts of the GUI</title>

<sect2>
<title
>Sidedocks</title>
<para
>Sidedocks (Dockables) are side windows which can be placed at any border of an KCachegrind window. They always contain a list of cost entities sorted in some manner. <itemizedlist>
<listitem
><para
>Function Profile. The Function Profile is a list of functions showing inclusive and exclusive cost, call count, name and position of functions. </para
></listitem>
<listitem
><para
>Parts Overview </para
></listitem>
<listitem
><para
>Call Stack </para
></listitem>
</itemizedlist>
</para>
</sect2>

<sect2>
<title
>Visualisation Area</title>
<para
>The visualisation area, typically the right part of a KCachegrind main window, is made up of one (default) or more Tab Views, either lined up horizontally or vertically. Each tab view holds different visualisation views of only one cost entity at a time. The name of this entity is given at the top of the tab view. If there are multiple tab views, only one is active. The entity name in the active tab view is shown in bold and determines the active cost entity of the KCachegrind window. </para>
</sect2>

<sect2>
<title
>Areas of a Tab View</title>
<para
>Each tab view can hold up to four view areas, namely Top, Right, Left, and Bottom. Each area can hold multiple stacked visualisation views. The visible view of an area is selected by a tab bar. The tab bars of the top and right area are at the top; the tab bars of the left and bottom area are at the bottom. You can specify which kind of visualisation should go into which area by using the context menus of the tabs. </para>
</sect2>

<sect2>
<title
>Synchronised Visualisation via Selected Entity in a Tab View</title>
<para
>Besides an active entity, each tab view has an selected entity. As most visualisation types show multiple entities with the active one somehow centred, you can change the selected item by navigating inside a visualisation (by clicking with the mouse or using the keyboard). Typically, selected items are shown in an highlighted state. By changing the selected entity in one of the visualisations of a tab view, all other visualisations in the tab view accordingly highlight the new selected entity. </para>
</sect2>

<sect2>
<title
>Synchronisation between Tab Views</title>
<para
>If there are multiple tab views, a selection change in one tab view leads to an activation change in the next (to right/to bottom) tab view. This kind of linkage should, for example, allow for fast browsing in call graphs. </para>
</sect2>

<sect2>
<title
>Layouts</title>
<para
>The layout of all the tab views of a window can be saved (see menu item View/Layout). After duplicating the current layout (Ctrl+Plus or menu) and changing some sizes or moving a visualisation view to another area of an tab view, you can quickly switch between the old and the new layout via Ctrl+Left/Right. The set of layouts will be stored between KCachegrind sessions of the same profiled command. You can make the current set of layouts as the default one for new KCachegrind sessions, or restore to the default layout set. </para>
</sect2>
</sect1>

<sect1 id="concepts-sidedocks">
<title
>Sidedocks</title>

<sect2>
<title
>Flat Profile</title>
<para
>The flat profile contains a group list and a function list. The group list contains all groups where cost is spent in, depending on the chosen group type. The group list is hidden when grouping is switched off. </para>
<para
>The function list contains the functions of the selected group (or all functions if grouping is switched off), ordered by some column, e.g. inclusive or self costs spent therein. There is a maximal number of functions shown in the list, which is configurable in Settings/Configure KCachegrind. </para>
</sect2>

<sect2>
<title
>Parts Overview</title>
<para
>In a profile run, multiple profile data files can be produced, which can be loaded together into KCachegrind. The Parts Overview dockable shows these, horizontally ordered according creation time; the rectangle sizes are proportional to the cost spent in the parts. You can select one or several parts to constrain the costs shown in the other KCachegrind views to these parts only. </para>
<para
>The parts are further subdivided: there is a partitioning and an inclusive cost split mode: <itemizedlist>
<listitem
><para
>Partitioning: You see the partitioning into groups for a profile data part, according to the group type selected. For example, if ELF object groups are selected, you see coloured rectangles for each used ELF object (shared library or executable), sized according to the cost spent therein. </para
></listitem>
<listitem
><para
>Inclusive Cost Split: A rectangle showing the inclusive cost of the current active function in the part is shown. This, again, is split up to show inclusive costs of its callees. </para
></listitem>
</itemizedlist>
</para>
</sect2>

<sect2>
<title
>Call Stack</title>
<para
>This is a purely fictional 'most probable' call stack. It is built up by starting with the current active function and adds the callers/callees with highest cost at the top and to bottom. </para>
<para
>The 'Cost' and 'Calls' columns show the cost used for all calls from the function in the line above. </para>
</sect2>
</sect1>

<sect1 id="concepts-visualizations">
<title
>Visualisations</title>

<sect2>
<title
>Event Types</title>
<para
>This list shows all cost types available and the corresponding self and inclusive cost of the current active function for that event type. </para>
<para
>By choosing an event type from the list, you change the type of costs shown all over KCachegrind to be the selected one. </para>
</sect2>

<sect2>
<title
>Call Lists</title>
<para
>These lists show calls to/from the current active function. With ''all' callers/callees functions are meant which can be reached in caller/callee direction, even when other functions are in between. </para>
<para
>Call list views include: <itemizedlist>
<listitem
><para
>Direct Callers </para
></listitem>
<listitem
><para
>Direct Callees </para
></listitem>
<listitem
><para
>All Callers </para
></listitem>
<listitem
><para
>All Callees </para
></listitem>
</itemizedlist>
</para>
</sect2>

<sect2>
<title
>Maps</title>
<para
>A treemap visualisation of the primary event type, up or down the call hierarchy. Each coloured rectangle represents a function; its size tries to be proportional to the cost spent therein while the active function is running (however, there are drawing constrains). </para>
<para
>For the Caller Map, the graph shows the nested hierarchy of all callers of the current activated function; for the Callee Map, it shows the nested hierarchy of all callees of the current activated function. </para>
<para
>Appearance options can be found in the in the context menu. To get exact size proportions, choose 'Hide incorrect borders'. As this mode can be very time consuming, you may want to limit the maximum drawn nesting level before. 'Best' determinates the split direction for children from the aspect ratio of the parent. 'Always Best' decides on remaining space for each sibling. 'Ignore Proportions' takes space for function name drawing before drawing children. Note that size proportions can get heavily wrong. </para>
<para
>Keyboard navigation is available with the left/right arrow keys for traversing siblings, and up/down arrow keys to go a nesting level up/down. 'Return' activates the current item. </para>
</sect2>

<sect2>
<title
>Call Graph</title>
<para
>This view shows the call graph around the active function. The shown cost is only the cost which is spent while the active function was actually running; i.e. the cost shown for main() - if it's visible - should be the same as the cost of the active function, as that is the part of inclusive cost of main() spent while the active function was running. </para>
<para
>For cycles, blue call arrows indicate that this is an artificial call added for correct drawing which actually never happened. </para>
<para
>If the graph is larger than the widget area, a bird's eye view is shown in one edge. There are similar visualisation options as for the Call Treemap; the selected function is highlighted. </para>
</sect2>

<sect2>
<title
>Annotations</title>
<para
>The annotated source/assembler lists show the source lines/disassembled instructions of the current active function together with (self) cost spent while executing the code of a source line/instruction. If there was a call, lines with details on the call are inserted into the source: the (inclusive) cost spent inside of the call, the number of calls happening, and the call destination. </para>
<para
>Select such a call information line to activate the call destination. </para>
</sect2>
</sect1>

</chapter>


<chapter id="commands">
<title
>Command Reference</title>

<sect1 id="tdecachegrind-mainwindow">
<title
>The main &tdecachegrind; window</title>
<para
></para>

<sect2>
<title
>The <guimenu
>File</guimenu
> Menu</title>
<para>
<variablelist>

<varlistentry>
<term
><menuchoice
><shortcut
> <keycombo
>&Ctrl;<keycap
>N</keycap
></keycombo
> </shortcut
> <guimenu
>File</guimenu
> <guimenuitem
>New</guimenuitem
> </menuchoice
></term>
<listitem
><para
><action
> Opens an empty toplevel window into which you can load profile data. </action
> This action is not really needed, as File/Open will give you a new toplevel window when the current one shows already some data. </para
></listitem>
</varlistentry>

<varlistentry>
<term
><menuchoice
><shortcut
> <keycombo
>&Ctrl;<keycap
>O</keycap
></keycombo
> </shortcut
> <guimenu
>File</guimenu
> <guimenuitem
>Open</guimenuitem
> </menuchoice
></term>
<listitem
><para
><action
> Pops up the File Open Dialogue to choose a profile data file to be loaded. </action
> If there is some data already shown in the current toplevel window, this will open a new window; if you want to open additional profile data in the current window, use File/Add. </para>
<para
>The name of profile data files usually ends in ..-, where and are optional and are used for multiple profile data files belonging to one application run. By loading a file ending only in ., eventually existing data files for this run, but with additional endings, are loaded too. </para>
<para
>Example: If there exist profile data files cachegrind.out.123 and cachegrind.out.123.1, by loading the first, the second will be automatically loaded too. </para
></listitem>
</varlistentry>

<varlistentry>
<term
><menuchoice
><guimenu
>File</guimenu
> <guimenuitem
>Add</guimenuitem
> </menuchoice
></term>
<listitem
><para
><action
> Adds a profile data file to the current window. </action
> Using this, you can force multiple data files to be loaded into the same toplevel window even if they are not from the same run as given by the profile data file naming convention. This can, for example, be used for nearside comparison. </para
></listitem>
</varlistentry>

<varlistentry>
<term
><menuchoice
><guimenu
>File</guimenu
> <guimenuitem
>Reload</guimenuitem
> </menuchoice
></term>
<listitem
><para
><action
> Reload the profile data. </action
> This is most interesting after another profile data file was generated for an already loaded application run. </para
></listitem>
</varlistentry>

<varlistentry>
<term
><menuchoice
><shortcut
> <keycombo
>&Ctrl;<keycap
>Q</keycap
></keycombo
> </shortcut
> <guimenu
>File</guimenu
> <guimenuitem
>Quit</guimenuitem
> </menuchoice
></term>
<listitem
><para
><action
>Quits</action
> &kappname;</para
></listitem>
</varlistentry>
</variablelist>
</para>

</sect2>

<sect2>
<title
>The <guimenu
>View</guimenu
> Menu</title>
<para>
<variablelist>

<varlistentry>
<term
><menuchoice
><guimenu
>View</guimenu
> <guimenuitem
>Primary Event Type</guimenuitem
> </menuchoice
></term>
<listitem
><para
><action
>(To-do)</action
></para
></listitem>
</varlistentry>

<varlistentry>
<term
><menuchoice
><guimenu
>View</guimenu
> <guimenuitem
>Secondary Event Type</guimenuitem
> </menuchoice
></term>
<listitem
><para
><action
>(To-do)</action
></para
></listitem>
</varlistentry>

<varlistentry>
<term
><menuchoice
><guimenu
>View</guimenu
> <guimenuitem
>Grouping</guimenuitem
> </menuchoice
></term>
<listitem
><para
><action
>(To-do)</action
></para
></listitem>
</varlistentry>

<varlistentry>
<term
><menuchoice
><guimenu
>View</guimenu
> <guimenuitem
>Layout</guimenuitem
> </menuchoice
></term>
<listitem
><para
><action
>(To-do)</action
></para
></listitem>
</varlistentry>

<varlistentry>
<term
><menuchoice
><guimenu
>View</guimenu
> <guimenuitem
>Split</guimenuitem
> </menuchoice
></term>
<listitem
><para
><action
>(To-do)</action
></para
></listitem>
</varlistentry>

</variablelist>
</para>

</sect2>


</sect1>
</chapter>

<chapter id="faq">
<title
>Questions and Answers</title>
&reporting.bugs; &updating.documentation; <qandaset id="faqlist">


<qandaentry>
<question>
<para
>What is &tdecachegrind; for? I have no idea. </para>
</question>
<answer>
<para
>&tdecachegrind; is a helpful at a later stage in software development, called Profiling. If you don't develop applications, you don't need &tdecachegrind;. </para>
</answer>
</qandaentry>

<qandaentry>
<question>
<para
>What is the difference between 'Incl.' and 'Self' ? </para>
</question>
<answer>
<para
>These are cost attributes for functions regarding some event type. As functions can call each other, it makes sense to distinguish the cost of the function itself ('Self Cost') and the cost including all called functions ('Inclusive Cost'). 'Self' is sometimes also referred to as 'Exclusive' costs. </para>
<para
>So, for example, for main(), you will always have a inclusive cost of almost 100%, whereas the self cost is negligible when the real work is done in another function. </para>
</answer>
</qandaentry>

<qandaentry>
<question>
<para
>The toolbar/menubar of my KCachegrind looks so spartanic. Is this normal?</para>
</question>
<answer>
<para
>Obviously KCachegrind is wrongly installed on your system. It is recommended to compile it with the installation prefix to be your system wide KDE base directory like <command
>configure --prefix=/opt/kde3; make install</command
>. If you choose another directory like $HOME/kde, you should set the environment variable TDEDIR to this directory before running KCachegrind. </para>
</answer>
</qandaentry>

<qandaentry>
<question>
<para
>If I double-click on a function down in the Call Graph View, it shows for the function main the same cost as the selected function. Isn't this supposed to be constant 100% ? </para>
</question>
<answer>
<para
>You have activated a function below main() with cost less than main(). For any function, only that part of the full cost of the function is shown, that is spent while the activated function is running; that is, the cost shown for any function can never be higher than the cost of the activated function. </para>
</answer>
</qandaentry>


</qandaset>
</chapter>

<chapter id="glossary">
<title
>Glossary</title>

<para
>The following is a mixed list of terms. <itemizedlist>
<listitem
><para
>Profiling: The process of collecting statistical information about runtime characteristics of program runs. </para
></listitem>
<listitem
><para
>Tracing: The process of supervising a program run and storing events happening sorted by a timestap in a output file, the Trace. </para
></listitem>
<listitem
><para
>Trace: A sequence of timestamped events that occurred while tracing a program run. Its size is typically linear to the execution time of the program run. </para
></listitem>
<listitem
><para
>Profile Data File: A file containing data measured in a profile experiment (or part of) or produced by postprocessing a trace. Its size is typically linear to the code size of the program. </para
></listitem>
<listitem
><para
>Profile Data Part (incorrectly used also: Trace Part): Data from a profile data file. </para
></listitem>
<listitem
><para
>Profile Experiment: A program run supervised by a profiling tool, producing possibly multiple profile data files from parts and/or threads of the run. </para
></listitem>
<listitem
><para
>Profile Project: A configuration for profile experiments used for one program which has to be profiled, perhaps in multiple versions. Comparisons of profile data typically only makes sense between profile data produced in experiments of one profile project. </para
></listitem>
<listitem
><para
>Cost Entity: An abstract item related to source code to which event counts can be attributed. Dimensions for cost entities are code location (e.g. source line, function), data location (e.g. accessed data type, data object), execution location (e.g. thread, process) and tuples or triples of the aforementioned positions (e.g. calls, object access from statement, evicted data from cache). </para
></listitem>
<listitem
><para
>Event Type: The kind of event of which costs can be attributed to a cost entity. There exist real event types and inherited event types. </para
></listitem>
<listitem
><para
>Real Event Type: A event type that can be measured by a tool. This needs the existence of a sensor for the given event type. </para
></listitem>
<listitem
><para
>Inherited Event Type: A virtual event type only visible in the visualisation which is defined by a formula to be calculated from real event types. </para
></listitem>
<listitem
><para
>Event Costs: Sum of events of some event type occurring while the execution is related to some cost entity. The cost is attributed to the entity. </para
></listitem>
</itemizedlist>
</para>
</chapter>

<chapter id="credits">


<title
>Credits and License</title>

<para
>&kappname; </para>
<para
>Thanks to Julian Seward for his excellent &valgrind;, and Nicholas Nethercote for the &cachegrind; addition. Without these programs, <application
>KCachegrind</application
> would not exist. Some ideas for this &GUI; were from them, too. </para>
<para
>And thanks for all the bug reports/suggestions from different users. </para>
&underFDL; </chapter>

<appendix id="installation">
<title
>Installation</title>

<sect1 id="getting-tdecachegrind">
<title
>How to obtain &tdecachegrind;</title>

<para
>&tdecachegrind; is part of the &package; package of &kde;. For less supported interim releases, &callgrind; and further documentation, see the homepage at <ulink url="http://tdecachegrind.sf.net"
> http://tdecachegrind.sf.net</ulink
>. Look there for further installation and compile instructions. </para>
</sect1>

<sect1 id="requirements">
<title
>Requirements</title>

<para
>In order to successfully use &tdecachegrind;, you need &kde; 3.x. For generating profile data, &cachegrind; or &calltree;/&callgrind; is recommend. </para>
</sect1>

<sect1 id="compilation">
<title
>Compilation and Installation</title>
&install.compile.documentation; </sect1>

<sect1 id="configuration">
<title
>Configuration</title>

<para
>All configuration options are either in the configuration dialogue or in the context popup menus of the visualisations. </para>

</sect1>

</appendix>

&documentation.index;
</book>
<!--
Local Variables:
mode: sgml
sgml-minimize-attributes:nil
sgml-general-insert-case:lower
sgml-indent-step:0
sgml-indent-data:nil
End:
-->