/[PAMELA software]/PamCut/doc/PamCutDevGuide.tex
ViewVC logotype

Annotation of /PamCut/doc/PamCutDevGuide.tex

Parent Directory Parent Directory | Revision Log Revision Log


Revision 1.5 - (hide annotations) (download) (as text)
Tue Oct 27 10:24:36 2009 UTC (15 years, 1 month ago) by pam-fi
Branch: MAIN
Changes since 1.4: +32 -13 lines
File MIME type: application/x-tex
Added to repository.

1 pam-fi 1.1 \documentclass{article}
2    
3     \title{PamCut Developer's Guide}
4     \author{Nicola Mori}
5    
6     \begin{document}
7     \maketitle
8     \tableofcontents
9     \section{The philosophy}
10    
11     {\bf PamCut} is an abstract class, defining the interface for a cut object used
12     in the analysis of PAMELA data. The main idea beyond the development is that a
13     cut object must be capable of saying if an event is good or not, ie., if it
14     satisfies some selection criteria. This criteria have to be implemented from
15     scratch every time a specific cut is needed, by defining a concrete class which
16     inherits form {\bf PamCut} and provides an implementation for the pure virtual
17     method {\bf Check}.
18     A special derived class is {\bf PamCutCollection}. This is a sort of a
19 pam-fi 1.5 container for basic cuts, and is considered as a single cut. It has an {\bf
20     AddCut} method which will add a cut to the collection. Its {\bf Check}
21 pam-fi 1.1 implementation simply invokes all the {\bf Check}s of the single cuts, and it
22 pam-fi 1.5 is successful if all the single cuts are succesful. {\bf PamCut} also provides
23     the interface for two methods for post-processing: {\bf OnGood} and {\bf
24     OnBad}. In derived classes these can contain specific tasks to be performed
25     whenever an event satisfy the {\bf Check} condition or not. The method {\bf
26     ApplyCut} takes care of invoking {\bf Check} and subsequently calls {\bf
27     OnGood} or {\bf OnBad} according to the result of {\bf Check}. Summarizing,
28     {\bf Check}ing an event simply means to ask the object if the event satisfy
29     the selection criteria; applying a cut means to check and then perform
30     post-selection tasks. The cut can be applied to a bunch of events by means of
31     the {\bf Process} method.
32 pam-fi 1.1
33     \subsection{More on collections}
34     A collection is an object which inherits from {\bf PamCutCollection}, which in
35     turn inherits from {\bf PamCut}. So a collection is a cut itself, meaning that
36     its interface contains a {\bf Check} method, an {\bf ApplyCut} method and so
37     on. Logically, this is in agreement with the fact that a bunch of cuts can be
38     thought of as a single cut whose result is the logical AND of all the basic
39     cuts. More specifically, the implementation chosen for {\bf PamCutCollection}
40     methods consists in simply performing a cyclic call of the corresponding
41     methods in basic cuts. So {\bf PamCutCollection::Check} will subsequently call
42     all the {\bf Check} methods of the basic cuts it contains, and {\bf
43     PamCutCollection::ApplyCut} will call the basic {\bf ApplyCut}s. This last
44     feature deserves some more words. When the collection calls a basic {\bf
45     ApplyCut}, the basic cut will also perform its specific post-selection tasks as
46     defined in its implementations of {\bf OnGood} and {\bf OnBad}. If all the
47     basic cuts result to be satisfied, then {\bf PamCutCollection::ApplyCut} will
48     call {\bf PamCutCollection::OnGood}, allowing for the execution of a
49     post-processing task which occurs only if all the basic selection criteria are
50     satisfied. Indeed, as said above a collection is a cut, so it behaves exactly
51     like a cut: when you apply it, you will also do the apprpriate post-processing.
52     We have then two levels of post-processing: the first is triggered by success
53     of a basic cut, the second by the success of the whole sequence of basic cuts.
54     This modular behaviour achieved with collections allow for a definition of a
55     hierarchy of cuts: a collection can contain other collections, which in turn
56     can contain other collections or basic cuts, in a tree-like hierarchy. Each
57     level has its own post-selection tasks, allowing for a fine control of the
58     post-processing procedure.
59    
60     To perform some tests it could be useful to have a collection which applies
61     the whole bunch cuts to all the events, regardless if some cuts are not
62     satisfied for a specific event ({\bf PamCutCollection} stop the evaluation for
63     the current event as soon as a cut is not satisfied). This is achieved with the
64     {\bf BlindCutCollection} class, which blindly checks all the cuts for all the
65     events. This will lead to a call to {\bf OnGood} or {\bf OnBad} for all the
66     cuts for each event; the collection will the call its own {\bf OnGood} or {\bf
67     OnBad} if all the cuts have been satisfied or not, much like in
68     {\bf PamCutCollection}. See the Doxygen html documentation for more info about
69     {\bf BlindCutCollection} and other specific cut implementations.
70    
71 pam-fi 1.5 \subsection{Collections and memory management}
72    
73     Regarding to cuts ownership, a collection will by default own its cuts. This
74     means that when you add a cut to a collection calling {\bf AddCut} with a
75     pointer to the cut as argument, the collection will store this pointer and
76     will take charge of deleting it. This is done by the collection's destructor.
77     This way, a user can simply add a cut to a collection and forget about
78     deleting it. However, sometimes this feature is not desirable, so it can be
79     turned off by passing a proper value to the collection's constructor. In this
80     case, the user is in charge of managing the memory allocated for the cuts.
81    
82    
83 pam-fi 1.1 \section{Actions and the SmartCollection}
84     When performing an analysis, each time an event is selected as good some
85     actions are likely to be performed, like filling histograms or writing a
86     report. To automate these tasks, the class {\bf CollectionAction} has been
87     designed. A {\bf CollectionAction} object has a {\bf Setup} method which
88     may contain the initialization procedures, like reading parameters from a file.
89     Also the constructor can contain some initializations. The finalization, like
90     writing histograms on a file, is in the {\bf Finalize} method. The specific
91     actions for good and bad events are to be defined in the {\bf OnGood} and {\bf
92     OnBad} methods, much like in collections. {\bf CollecionAction} is an abstract
93     class, which does nothing but defining the interface. Its concrete
94     implementations will be called actions.\\
95     Actions are automatically handled by the {\bf SmartCollection} class. It
96     inherits from {PamCutCollection}, and contains a vector of {\bf
97 pam-fi 1.5 CollectionAction} objects. These actions can be added calling {\bf
98 pam-fi 1.1 SmartCollection::AddAction}; for each of them, the collection will take care of
99     calling {\bf Setup} at the beginning of the analysis, {\bf OnGood} and {\bf
100     OnBad} for every event (depending on the selection result), and {\bf Finalize}
101     at the end of the analysis. In all other aspects, it behaves exactly as {\bf
102 pam-fi 1.2 PamCutCollection}. The handling of the actions depend on their position with
103     respect to cuts. In the main analysis code, cuts can be added to the
104     collection; then, an action or more can be added. This sequence of cuts and the
105     actions added after the cuts will be called a bunch. The last bunch may have no
106     actions at its end.
107     \begin{verbatim}
108     .
109     .
110     .
111     Cut1 cut1(``cut1'');
112     Cut2 cut2(``cut2'');
113     Action1 action1(``action1'');
114    
115     Cut1 cut3(``cut3'');
116     Cut2 cut4(``cut4'');
117     Cut1 cut5(``cut5'');
118     Cut2 cut6(``cut6'');
119     Action1 action2(``action2'');
120     Action1 action3(``action3'');
121     .
122     .
123     .
124     \end{verbatim}
125     In the example above, {\bf cut1}, {\bf cut2} and {\bf action1} are the first
126     bunch, {\bf cut3}, {\bf cut4}, {\bf cut5}, {\bf cut6}, {\bf action2} and {\bf
127     action3} are the second bunch and so on. If all the cuts in the bunch are
128     successful, the {\bf SmartCollection} will call {\bf OnGood} for every action
129     in that bunch, and then the analysis will go on with the next bunch for the same
130     event. If a certain cut in a bunch fails, then {\bf OnBad} is called for the
131     actions of the bunch, but successive bunches are ignored; the current event is
132     discarded and the focus switches on the next one (the {\bf
133     SmartBlindCollection} behaves a little different; see the Doxygen
134     documentation for more details.)
135     \\
136 pam-fi 1.1 Loosely speaking, after defining an action one simply has to instantiate it,
137 pam-fi 1.2 add it to a {\bf SmartCollection} and launch the analysis.
138 pam-fi 1.1
139 pam-fi 1.5 \subsection{SmartCollections and memory management}
140     Like the standard collection, SmartCollection can handle the memory management
141     for actions. This works exactly as for cuts. Cuts and actions have to be
142     managed uniformly, ie., one cannot turn off ownership only for cuts or only for
143     actions.
144    
145 pam-fi 1.1 \section{The software organization}
146     The software is organized in a tree of directories. The idea is that each node
147     of this tree must provide the necessary informations about the sub-branches. In
148     each directory, a header file will contain \verb1#include1 directives for all
149     header files in the sub-directories. This way, it is sufficient to include the
150     top-level header in the analysis software to automatically include all the
151     headers of the project. This top-level header is {\it PamCutEnv.h}. Each time a
152     sub-directory is created, the headers it contains must be included in the
153     parent node's header.
154    
155     \subsection{The makefile organization}
156     The instructions to build the software library are encoded in the standard {\bf
157     make} way, following the same tree structure of the software. Each node
158     directory contains an {\it include.mk} file, which includes all the
159     {include.mk} in the sub-directories. This chain of inclusions terminates on the
160     leaves, which most of the times contain the definition and implementation of
161     the classes. Each leaf directory must provide build instructions for the files
162     it contains, in a file called {\it subdir.mk}; this file has to be included in
163     the upper {\it include.mk}. The top level {\it include.mk} is included by
164     the {\it makefile}.\\
165     This way, each directory takes care of its sub-directories, letting a quick and
166     easy addition of the latters.
167    
168     \section{The software architecture}
169     The software is organized in a tree of directories. The root folder contains a
170     {\it PamCutBase} folder, in which the definitions of the base classes ({\bf
171     PamCut} and {\bf PamCutCollection}) are stored, and the general headers {\it
172     PamCutEnv.h} and {\it CommonDefs.h}. {\it PamCutEnv.h} contains the inclusion
173     of all the headers of the framework, allowing for the use of the entire
174     software with only one \verb1#include1 in the code; {\it CommonDefs.h} contains
175     all the definitions which are relevant for the entire framework.
176     To develop a specific cut one has to define a derived class from {\bf PamCut},
177     which at least must provide a concrete implementation of {\bf Check}. One can
178     also define new versions of {\bf OnGood} and {\bf OnBad} for specific
179     post-selection tasks (in the base class these methods do nothing). Be very
180     careful if you decide to redefine {\bf ApplyCut}: remeber that the interface
181     requires them to call the post-selection routines.
182     A good rule for keeping the software distribution clean is to use a different
183     folder for each cut definition, named as the cut itself. When you define a new
184     cut, create a folder named as the cut inside an adequate parent directory and
185     place the {\it .h} and {\it .cpp} files inside it. To speed up development, you may copy
186     and paste an existing cut and then modify it.\\
187     In the same way, actions can be defined.
188    
189     \section{How to define a cut}
190     As said above, to define a cut (let's name it {\bf DummyCut}) it is suggested
191     to create a folder {\it DummyCut} inside, eg., a {\bf DummyDir} directory
192     inside the root folder, and to create inside it two files: {\it DummyCut.h},
193     which will contain all the declarations, and {\it DummyCut.cpp}, where the
194     implementation will be coded. A typical structure for {\it DummyCut.h} would be:
195    
196     \begin{verbatim}
197     #ifndef DUMMYCUT_H_
198     #define DUMMYCUT_H_
199    
200     /*! @file DummyCut.h The DummyCut class declaration file */
201    
202     #include "../../PamCutBase/PamCutBase.h"
203    
204     /*! @brief An example cut. */
205     class DummyCut: public PamCut {
206     public:
207    
208     /*! @brief Constructor. */
209     Dummy(char *cutName):
210     PamCut(cutName) {
211     }
212    
213     /*! @brief Destructor. */
214     ~DummyCut() {
215     }
216    
217     /*! @brief The dummy cut check
218     *
219     * This routine checks the event using the dummy cut criteria.
220     *
221     * @param event The event to analyze.
222     * @return CUTOK if the dummy cut is satisfied
223     */
224     int Check(PamLevel2 *event);
225    
226     };
227    
228     #endif /* DUMMYCUT_H_ */
229    
230     \end{verbatim}
231    
232     Note the inclusion of {\it PamCutBase.h}: this is essential to let {\bf
233     DummyClass} know about its parent class, which in this case is {\bf PamCut}.
234     Another thing to care about is the multiple inclusion protection, ie., the
235     preprocessor directives:
236     \begin{verbatim}
237     #ifndef DUMMYCUT_H_
238     #define DUMMYCUT_H_
239     .
240     .
241     .
242     #endif /* DUMMYCUT_H_ */
243     \end{verbatim}
244     This provide protection against multiple inclusion of a header in a single
245     compilation unit due to circular inclusion; in practice, it will save the
246     developer a lot of double definition errors from the compiler. Be sure that all
247     your headers are encapsulated in such a structure, and that for each header
248     there is a different name after \verb1#ifndef1 (which must match the successive
249     \verb1#define1); otherwise, the compiler could skip the inclusion of some
250     headers. The convention adopted in {\bf PamCut} for such tags is:
251     \verb1<Filename in uppercase>_H_1, which provides univocal tags for each header.
252    
253     The essential redeclarations are: the constructor, the desctructor and {\bf
254     Check}. Constructor deserves some observations: since constructors are not
255     inherited, each class must have its own. However, if a class has a parent
256     class, it is necessary to initialize also the parent class. If the parent class
257     has a default constructor (ie., a constructor without arguments) the compiler
258     will automatically call it and the user has nothing to worry about. However,
259     {\bf PamCut} has no default constructor, so in every derived class there must
260     be an explicit call to the {\bf PamCut} constructor. This is what is done in
261     the lines:
262    
263     \begin{verbatim}
264     Dummy(char *cutName):
265     PamCut(cutName) {
266     }
267     \end{verbatim}
268     {\bf PamCut} objects have a name, internally stored in the {\bf \_cutName}; so
269     our {\bf DummyClass}, since it inherits from {\bf PamCut}, must also have a
270     name. Its constructor, indeed, requires a string containing the cut's name. But
271     in {\bf DummyCut} there's no local variable to store it, and there's no need to
272     define one since it is already present in {\bf PamCut}. The only thing to do is
273     to initialize it exactly as {\bf PamCut} would do: this is the purpose of
274     invoking \verb1PamCut(cutName)1. As an aside, what follows the colon after the
275     {\bf DummyCut} constructor declaration is called \emph{initialization list}:
276     here one can (and must) call the constructors for all the objects and variables
277     contained in the class. These are executed before the class constructor's body,
278     so that one can safely assume that at construction time all the internal
279     objects and variables are properly initialized.
280    
281     To recap and generalize, when you write a derived class you must always call
282     the constructor of its parent class, which will take care about initializing
283     the ``parent core'' of the derived class.
284    
285     One can also redefine {\bf OnGood}, {\bf OnBad}, add new methods, variables and
286     so on. If you plan to build a general purpose cut who could be used by
287     different people please take the time to add the appropriate documentation to
288     the code. In the example, Doxygen comment lines (beginning with \verb1/*!1)
289     are shown; an on-line guide for Doxygen can be found at:
290     \newline
291     \newline
292     \verb1 http://www.stack.nl/~dimitri/doxygen/manual.html1
293    
294     \vspace{.5cm}
295     Once the header has been prepared, it's time to implement the cut into {\it
296     DummyCut.cpp}:
297    
298     \begin{verbatim}
299     /*! @file DummyCut.cpp The DummyCut class implementation file */
300    
301     #include "DummyCut.h"
302    
303     int DummyCut::Check(PamLevel2 * event) {
304    
305     if (<Some condition about the event>)
306     return CUTOK;
307     else
308     return 0;
309     }
310     \end{verbatim}
311    
312     In this very simple implementation a basic feature is noticeable: the interface
313     requires that whenever an event satisfy the selection criterion, {\bf Check}
314     must return the {\bf CUTOK} value. This value is defined in {\it CommonDefs.h}.
315     The return value for a discarded event is implementation-specific, and can be
316     anything but {\bf CUTOK}. This return value could take different values
317     depending on the reason why the event has been discarded: for example, it can
318     be a number associated to a specific detector whose data is missing in a data
319     quality cut. It is then passed to {\bf OnBad}, which can perform some task
320     depending on the specific reason of the cut failure (eg., on which detector has
321     no data).
322     Remember also to include the header file.
323    
324     \section{How to setup the cut's build} \label{sec:build}
325     Once a cut has been declared and implemented, the makefiles have to be adjusted
326     to include it in the cut library. The {\it makefile} provided with the software
327     will build a library called {\it libPamCut.so}, which can be then linked to a
328     specific analysis code. It is based on a submakefile structure. Each folder
329     containing a cut must also include one of these submakefiles (named {\it
330     subdir.mk} by convention), who instructs the main {\it makefile} on how to build
331     the newly added cut. An example is:
332    
333     \begin{verbatim}
334     # Choose a name for the object file. This must coincide with the
335     # .cpp and .h filename, except for the extension which has to
336     # be .o
337     OBJS += ./DummyDir/DummyCut/DummyCut.o
338    
339     # Dependencies file. The extension must be .d, and the name equal
340     # to the cut name.
341     CPP_DEPS += ./DummyDir/DummyCut/DummyCut.d
342    
343     # Rules for compilation. You will likely have only to specify
344     # the path. Put the directory path:
345     # here here
346     ./DummyDir/DummyCut/%.o: ./DummyDir/DummyCut/%.cpp
347     @echo 'Building file: $<'
348     @echo 'Invoking: GCC C++ Compiler'
349     $(C++) -I${ROOTSYS}/include -I${PAM_INC} -I${PAM_INC}/yoda \
350     -Wall -c -MMD -MP -MF"$(@:%.o=%.d)" -MT"$(@:%.o=%.d)"\
351     -o"$@" "$<"
352     @echo 'Finished building: $<'
353     @echo ' '
354    
355     \end{verbatim}
356    
357     Existing files can be used as a template. The first thing you have to modify
358     is the object name (\verb1OBJS1 variable): be careful to append the object name
359     to \verb1OBJS1 using \verb1+=1 instead of overwriting it with \verb1=1. The
360     paths in this file will be relative to the root directory where the {\it
361     makefile} is, eg., \verb1./1 is not the {\it DummyCut} directory where {\it
362     subdir.mk} is, but the root directory which contains the {\it makefile}. The
363     object file ({\it DummyCut.o}) must be named as the cut and the directory that
364     contains the cut. The \verb1CPP_DEPS1 variable must be similarly modified.
365     This variable contains a list of dependency files, which contains all the
366     external headers (eg., PAMELA and ROOT headers) {\bf PamCut} depends on. These
367     lists are automatically generated, and allows {\it make} to rebuild {\bf
368     PamCut} whenever one of these headers are modified. Finally, one has also to
369     put the directory name in the target line which precedes the build command,
370     just below \verb1here1.
371    
372     After creating the {\it subdir.mk}, it must be included in the {\it include.mk}
373     in the parent directory. It looks like this:
374    
375     \begin{verbatim}
376     # Include here the submakefiles for each cut
377    
378     # Cuts
379     -include DummyDir/DummyCut/subdir.mk
380     .
381     .
382     .
383     \end{verbatim}
384    
385    
386     Remember to write paths as relative to the root directory. Going backward along
387     the chain of inclusions leads to the {\it makefile}:
388    
389     \begin{verbatim}
390     # ------------------------ Build options ----------------------#
391    
392     # C++ compiler and linker
393     C++ = g++
394    
395     # Optimization flags.
396     OPTIMIZE = -g3 #-DDEBUGPAMCUT
397    
398     # Library flags
399     EXCLUSIONFLAGS = #-DNO_TOFNUCLEI -DNO_CALONUCLEI -DNO_TRKNUCLEI
400    
401     COMMONDEPS = makefile
402    
403     -include include.mk
404    
405     #------------------------ Make body ---------------------------#
406     #
407     # Below the make commands are defined. There are no options to
408     # set in this section, so it has to be modified only in case of
409     # radical changes to the make procedure.
410    
411     # Remove command for clean
412     RM := rm -rf
413    
414     # Additional dependencies from headers of other software
415     # (PAMELA, ROOT)
416     ifneq ($(MAKECMDGOALS),clean)
417     ifneq ($(strip $(CPP_DEPS)),)
418     -include $(CPP_DEPS)
419     endif
420     endif
421    
422     # All Target
423     all: version libPamCut.so
424    
425     # Tool invocations
426     libPamCut.so: $(OBJS) $(USER_OBJS)
427     @echo 'Building target: $@'
428     @echo 'Invoking: GCC C++ Linker'
429     $(C++) -shared -o"libPamCut.so" $(OBJS)
430     @echo 'Finished building target: $@'
431     @echo ' '
432    
433     # Other Targets
434     clean:
435     -$(RM) $(CPP_DEPS) $(OBJS) libPamCut.so
436     -@echo ' '
437    
438     version:
439     @gcc --version | grep gcc; echo
440    
441     .PHONY: all clean dependents version
442     \end{verbatim}
443    
444     The Build options section is the one that will most likely need a modification
445     if one wants to tweak the compilation; the make body will most of the times be
446     good as it is. The {\it makefile} contains comments to explain how to set the
447     various options.\\
448     The first option to set is the compiler: g++ will work on
449     almost all Linux systems. Next is the optimization level, which could be set to
450     one of the proposed values according to the specific necessities. The {\it
451     makefile} comments contains also instructions on how to enable debug sections
452     in the code.\\
453     Since the PAMELA software is modular, in some setups it may lack some libraries
454     needed by some cuts. When designing such a cut it is a good habit to setup
455     its eventual exclusion from the library. This can be efficiently done using
456     the compiler directive \verb1#ifdef1. For example, encapsulating all the code
457     in the header and implementation files of {\bf DummyCut} in a structure like
458     this:
459    
460     \begin{verbatim}
461     #ifdef NO_DUMMYCUT
462     .
463     .
464     .
465     #endif
466     \end{verbatim}
467    
468     \noindent will completely exclude {\bf DummyCut} from the environment if the
469     flag \verb1NO_DUMMYCUT1 is defined. It can be done passing the parameter
470     \verb1-DNO_DUMMYCUT1 to the compiler. These exclusion flags can be defined in
471     the \verb1EXCLUSIONFLAGS1 variable. A concrete example is given by
472     {\bf TofNucleiZCut}, which requires the {\bf ToFNuclei} library: look at it to
473     see how its exclusion from the code is implemented, allowing to build and use
474     {\bf PamCut} in those environments that do not include {\bf ToFNuclei}. In the
475     end, this is very similar to what is done with the debug sections.\\
476     The \verb1COMMONDEPS1 flag contains the files which, if modified, will trigger
477     a complete rebuild of the library. For example, if you change the {\it
478     makefile} by modifying an optimization option all the modules should be rebuilt
479     so that the whole library will have the same level of optimization. That's why
480     {\it makefile} is in \verb1COMMONDEPS1. Add all the other files that should
481     behave like this.\\
482    
483     If you need some extra modifications to the building system you need to know
484     more about {\it make}; an online guide is at:
485     \newline
486     \verb1http://www.linuxtopia.org/online_books/programming_tool_guides/1
487     \newline
488     \verb1 gnu_make_user_guide/make.html#SEC_Top1
489    
490    
491     \section{How to define an action}
492     Defining an action is very similar to defining a cut. An example header:
493    
494     \begin{verbatim}
495     #ifndef DUMMYACTION_H_
496     #define DUMMYACTION_H_
497    
498     #include "../CollectionAction/CollectionAction.h"
499    
500     /*! @brief A dummy action definition. */
501     class DummyAction: public CollectionAction {
502    
503     public:
504     /*! @brief Constructor.
505     *
506     * @param actionName The action's name.
507     */
508     DummyAction(const char *actionName):
509     CollectionAction(actionName){}
510    
511     /*! @brief Destructor */
512     ~DummyAction() {
513     }
514    
515     /*! @brief The setup procedure.
516     *
517     * @param events The events pointer.
518     */
519     void Setup(PamLevel2 *events);
520    
521     /*! @brief The OnGood procedure.
522     *
523     * @param event The selected event.
524     */
525     void OnGood(PamLevel2 *event);
526    
527     /*! @brief Writes the tree of saved events to the output file. */
528     void Finalize();
529     };
530    
531     #endif /* DUMMYACTION_H_ */
532    
533     \end{verbatim}
534    
535     The {\bf DummyAction} declaration above is a good example. The action classes
536     must inherit from {\bf CollectionAction}, and their constructor have to call
537     the constructor of the ancestor class in the initialization list, similarly to
538     what happens for cuts. Then some of the base class' methods are overridden,
539     specifically {\bf Setup}, {\bf OnGood} and {\bf Finalize}. The last two methods
540     have to be overridden, since they are pure virtual (an action is supposed to do
541     something for good events and something to end the analysis, so making these
542     methods pure virtual is a way to enforce a definition in the derived classes).
543     Conversely, {\bf Setup} and {\bf OnBad} are concrete in the base class, with
544     void implementation: since not all actions would need a setup (it can be done
545     also in the constructor) or a procedure for bad events, this implementation
546     allows to not define them in derived classes. Obviously, the re-declared
547     methods in the header have to have a re-definition in a .cpp file, exactly as
548     for the cuts.
549    
550     \section{How to setup the actions's build}
551     This topic is very similar to that explained in sec. \ref{sec:build}, so it
552     should be straightforward. However, look at the provided concrete
553     implementations of actions if you need some example to setup your build.
554    
555     \section{How to build and use the library}
556 pam-fi 1.3 \subsection{Standard Pamela environment}
557 pam-fi 1.1 If the makefiles are correctly set up, the only remaining thing is to type
558     \verb1make all1. Remember to set the PAMELA environment with the set\_pam\_env
559     script BEFORE invoking \verb1make1. This will generate a {\it libPamCut.so} file
560     which will contain all the cuts. To clean the project and build from scratch
561 pam-fi 1.3 type \verb1make clean all1. The software can then be installed in the usual
562     Pamela environment calling \verb1make install1: this will place all the
563     headers in the folder \verb1$PAM_INC/PamCut1 and the {\it libPamCut.so} file in
564     \verb1$PAM_LIB1. To eliminate the installed files call \verb1make distclean1;
565     note that this will NOT do the work of \verb1make clean1, eg., clean the
566     project, but simply remove the files copied in the Pamela directories. Remember
567     to type \verb1make install1 each time you modify and recompile the software, to
568     upgrade the installed version to the new one.
569    
570     To use the library in an analysis code the
571 pam-fi 1.1 environment header must be included in the code:
572     \verb1#include "<root PamCutdirectory>/PamCutEnv.h"1. With this, all the
573     classes and common definitions will be accessible. A typical usage of {\bf
574     PamCut} inside the analysis code would look like:
575    
576     \begin{verbatim}
577 pam-fi 1.3
578     #include <PamCut/PamCutEnv.h>
579    
580     int main(){
581     .
582     .
583     .
584    
585 pam-fi 1.5 PamCutCollection collection("Collection");
586 pam-fi 1.1
587     DummyCut1 dummy1;
588     collection.AddCut(dummy1);
589     DummyCut2 dummy2(param);
590     collection.AddCut(dummy2);
591    
592     collection.Process(event, 0, event->GetEntries()-1);
593    
594 pam-fi 1.3 .
595     .
596     .
597     }
598    
599 pam-fi 1.1 \end{verbatim}
600    
601     In the simple example above a \verb1DummyCut11 and a \verb1DummyCut21 object
602     (which requires some sort of parameter) are instantiated. They are added to
603     \verb1collection1 which takes care of applying them to all the events.
604    
605 pam-fi 1.3 \subsection{Custom environment}
606     If you don't have access to the Pamela software directories (eg., you don't
607     have write permission on them) you cannot install the software; but you can
608     still use PamCut directly from the source folder.
609    
610     First of all, you have to tell the compiler where to find the {\bf PamCut}'s
611     headers. They are in the main {\bf PamCut} directory, so you may add this
612     option:
613     \newline
614     \verb1 -I<directory>1
615     \newline
616     to the compiler invocation in the {\it makefile} of your main analysis program.
617     This tells the compiler to search for headers in the folder specified after
618     \verb1-I1. So, if {\it <directory>} is the folder which contains the {\bf
619     PamCut}'s main folder, you don't have to change anything in your main analysis
620     file (with respect to what described in the previous subsection), since:
621 pam-fi 1.1 \newline
622 pam-fi 1.3 \verb1 #include <PamCut/PamCutEnv.h>1
623     \newline
624     includes the file {\it PamCutEnv.h} which is in the folder {\it PamCut} in the
625     standard inclusion directories, one of which is the one specified with the
626     \verb1-I1 compiler directive. Obviously, one can play with directories, having
627     care to indicate the right paths to the compiler
628 pam-fi 1.1
629 pam-fi 1.3 The following option must be added to the linker invocation:
630     \newline
631     \verb1 -L<root PamCut directory> -lPamCut1.
632     \newline
633     to tell the linker where the dynamic library is.
634    
635     Then, when the analysis code is compiled and linked against libPamCut.so, to
636 pam-fi 1.1 launch it it's necessary to tell the environment where the library is, so that
637     the program can dynamically access it at runtime. This information is encoded
638     in the environment variable LD\_LIBRARY\_PATH, which contains the paths of the
639     accessible libraries. If libPamCut.so is still in the root PamCut directory one
640     can type:
641     \newline
642     \verb1export LD_LIBRARY_PATH=<root PamCut directory>:$LD_LIBRARY_PATH1
643     \newline
644     This has to be done every time you open a shell; one way to avoid this is to
645     append the above line at the end of your set\_pam\_env script, so that it will
646     be automatically executed every time you set your PAMELA environment.
647    
648     \section{Usage summary}
649     Here's a short summary on how to develop a cut, build and use it. If one only
650     \begin{enumerate}
651     \item Obtain the code (from tarball, repository\ldots) and go in the root
652     code directory.
653     \item Check if the \verb1C++1 option in the Build section of {\it makefile}
654     is correctly set with the C++ compiler name present in your system (for many
655     Linux platform, \verb1g++1 is a safe choice).
656     \item Create a directory named as the cut class you want to develop.
657     \item Place inside the newly created directory a {\it .h} file and a {\it
658     .cpp} file, named as the direcory; edit the files to define and implement the
659     class (one can also cut and paste the files from an existing class and edit
660     them), defining at least the constructor, the distructor and {\bf Check} for
661     the new class.
662     \item Create inside the directory a {\it subdir.mk} file which contains the
663     instructions to build the directory content, as described in \ref{sec:build};
664     as usual, one can cut and paste from an existing class and then edit.
665     \item Modify the {\it makefile} in the root code directory as in
666     \ref{sec:build}, to include the newly developed cut.
667     \item Modify the {\it PamCutEnv.h} file, adding the \verb1#include#1 for the
668     new class header (see examples therein).
669     \item Set the PAMELA environment with the set\_pam\_env script.
670     \item Build the library typing \verb1make all1 (or \verb1make clean all1 to
671     build from scratch); this will produce the library {\it libPamCut.so} in the
672     root code directory, which will contain all the class definitions and
673     implementations.
674     \item Insert \verb1#include ``<root PamCut directory>/PamCutEnv.h#1 in the
675     analysis code, to have access to all the classes in the library.
676     \item Develop the analysis code
677     \end{enumerate}
678    
679     \section{Some advices and suggestions}
680     \begin{itemize}
681     \item Derive your cuts. Try to define a new class every time you need a new
682     cut, instead of modifying an existing one. As an example, you can define a
683     cut with a specific implementation for {\bf Check}, then derive from it many
684     classes which only redefine {\bf OnGood} and {\bf OnBad}. In this way, you
685     can have several post-selection options associated to the same cut; if
686     you'll ever need to modify the cut criteria, you will have to do it only in
687     one place, saving time and reducing code errors opportunities.
688     \item Be consistent with the existing code style. Everyone has its own
689     code style, with its own conventions for naming variables and functions based
690     on personal preferences. Maintaining a uniform code style is a good way to
691     improve the code readability, so it's worth a little effort. The conventions
692     chosen for {\bf PamCut} are:
693     \begin{itemize}
694     \item the names of private and protected members (variables and methods)
695     always begin with an underscore, eg., \verb1_previousOBT1, \verb1_LT1;
696     \item the names of variables usually begin with a lower case; for compound
697     words, the successive initials are upper case, eg., \verb1time1,
698     \verb1liveTime1, \verb1nHitPaddles1;
699     \item the names of classes and methods begin with upper case, eg.,
700     \verb1PamCut1, \verb1GeoFieldCut1, \verb1ApplyCut()1.
701     \end{itemize}
702     Within these conventions, a code row like:
703     \newline
704     \verb1 GeoFieldCut geoFieldCut;1
705     \newline
706     is easily interpreted as the instantiation of a class named
707     \verb1GeoFieldCut1 into an object named \verb1geoFieldCut1. This allows to
708     have objects whose names are almost identical to those of the respective
709     classes, allowing a straightforward type recognition. Also, the discrimination between
710     public and private variables and methods inside a class is immediate.
711     \item Respect the interface. {\bf PamCut} has been designed following precise
712     rules, which allow for a quite general environment that should cover many
713     of the necessities related to data analysis. Try to fit your
714     particular analysis in this scheme; this will result in a much more
715     raedable code. However, someone may need features that are not compatible with
716     the current interface. In this case, the first thing to do is to try a
717     workaround that would leave the interface unchanged. As an example, if the
718     automated post-selection tasks based on {\bf OnGood} and {\bf OnBad} does not
719     satisfy you, you can call directly the {\bf Check} method inside a loop and
720     then do a custom post-processing. In the worst case, the interface could
721     result incompatible with the analysis needs: in this case a redesign
722     hypotheses can be considered if the incompatibility is such that a large
723     piece of the analysis would be compromised. Redesigning is always a tricky
724     task, so it has to be considered as a last option.
725     \item Take care about other people. If you plan to write a code that is
726     likely to be used and modified by other people, please take your time to
727     write the documentation. Documenting the code is a boring and time-consuming
728     task, but can save you and your colleagues a lot of headaches and
729     misunderstandings. The better a code is documented, the lesser are the
730     questions other people will ask you.
731     \end{itemize}
732    
733    
734     \end{document}

  ViewVC Help
Powered by ViewVC 1.1.23