• Tidak ada hasil yang ditemukan

Component Proportion (%) B→ Dτν 1.56 B → Dτν 3.18

B → Dlν 19.80

B→ Dlν 18.11

B→ D∗∗lν 22.74

CombinatorialBB¯ 28.17

Continuum 6.44

Table 7.8: Proportions of each component after reconstruction, evaluated using the generic MC.

• R2All: Second Fox-Wolfram moment.

• Mmiss2 : Square of the missing 4-vector of the event in the CM frame.

• Eextra: Extra neutral energy in the calorimeter.

• cosθT: Cosine of the angle between the thrust and the beam momentum.

• |ptagl |: 3-momentum magnitude of theBtaglepton in the CM frame.

• cosθtag

BD(∗)l: Cosine of the angle between the 3-momentum of the Btag and the 3-momentum sum of its D and lepton daughters in the CM frame.

• cosθtagDl: Cosine of the angle between the 3-momentum of the Dmeson and the lepton daughter in the tag side.

• mtagD : Mass of the Btag Dmeson daughter.

• ∆mtag: Mass difference betweenD andDmeson in the tag side, if exists.

• cosθtagD so f t: Cosine of the angle between theDmesons daughters in the tag side in the CM frame.

• |ptagso f t|: 3-momentum magnitude of the D’s soft daughter in the tag side in the CM frame.

• |psigl |: 3-momentum magnitude of theBsiglepton daughter in the CM frame.

• cosθsigDl: Cosine of the angle between the 3-momentum of the Dmeson and the lepton daughter in the sig side in the CM frame.

• χ2: χ2of theBsigvertex fit.

• mDsig: Mass of theBsig Dmeson daughter.

• ∆msig: Mass difference betweenDandDmeson in the sig side, if it exists.

• cosθsigD so f t: Cosine of the angle between the D mesons’ daughters in the sig side in the CM frame.

• |psigso f t|: 3-momentum magnitude of theD’s soft daughter in the sig side in the CM frame.

• cosθD(∗)lD(∗)l: Cosine of the angle between the twoDl Dlsystems in the CM frame.

• tagl electron PID:Btaglepton daughter’s electron PID level.

• tagl muon PID:Btag lepton daughter’s muon PID level.

• sigl electron PID:Bsiglepton daughter’s electron PID level.

• sigl muon PID:Bsiglepton daughter’s muon PID level.

Figure 7.5: Histograms of variables used for the C1classifier.

Figure 7.6: Histograms of variables used for the C1classifier.

Figure 7.7: Histograms of variables used for the C1classifier.

We use a Gradient Boosting Decision Tree (BDT) [36] for the C1 classifier. The number of decision trees ranges from 20 to 600. The model is implemented using the scikit-learnpackage. The metric used to evaluate the classification performance is the area under the ROC curve. The higher the score, the higher the classification power. Fig. 7.8 shows the relationship between the area under ROC curve and the number of trees. The classification performance becomes stable when the number

of decision trees is above 500. We use a BDT with 600 trees as the final model. Fig.

7.9 shows the importance of variables for classification. Eextra and |psigl | are the most powerful variables to identifyB→ D(∗)(τ/l)ν from all types of backgrounds.

ForEextra, the signal and normalization events usually have near-zero extra neutral energy, while background events have a wide distribution. The normalization decay has an energetic lepton produced by the D decay, leading to a higher |plsig| value than signal events, as well as all types of backgrounds. The output of the BDT score pis transformed using a logit function:

z1= logit(p)= log p 1− p.

The z1 distribution for all types of events is shown in Fig. 7.10. Signal and normalization events tend to have higher z1values than backgrounds.

Figure 7.8: Area under the ROC curve for BDT classifiers with different numbers of trees.

Figure 7.9: Importance of each variable for learning theC1classifier.

Figure 7.10: z1distribution for signal, normalization,D∗∗lν,BB¯combinatorial, and continuum events.

C2classifier

The C2 Classifier aims to classify signal events and normalization events. Similar to theC1classifier, we divide the sample into training and validation samples. The training sample is first used to train the classifier, the validation sample is then applied to evaluate the performance of the classifiers. Signal events are labeled positive and normalization events are labeled negative before training. We use the same variables used for theC1classifier, with the addition of the following quantity:

• cosθsig

BD(∗)l: Cosine of the angle between the 3-momentum of theBsigand the 3-momentum sum of its D and lepton daughters.

The histogram of these variables for signal and normalization events are shown in Fig. 7.11.

Figure 7.11: Histograms of variables used for theC2classifier.

Similar to the C1 classifier, we use a BDT for the C2 classifier. The number of decision trees ranges from 100 to 600. The classification performance is stable when the number of decision trees is above 100, as shown in Fig. 7.12. We use a BDT with 600 trees as the final model. Fig. 7.13 shows the importance of the variables for classification: cosθsigBD(∗)land|psigl |are the most powerful variables to distinguish signal from normalization events. For cosθsigBD(∗)l, normalization events have only a single neutrino, and the value should be from -1 to 1. However, signal

events tend to have more negative values, due to the presence of three neutrinos in the final state. Thee/µproduced from the τdecays in signal events have a softer

|psigl |spectrum than the leptons produced fromDdecays for normalization events.

The output of the BDT score is transformed to z2 using a logit function. The z2 distribution for all types of events is shown in Fig. 7.14. Signal events tend to have a higher z2 score, while normalization events tend to have a lower z2 score.

Backgrounds have z2scores in between.

Figure 7.12: The area under the ROC curve for BDT classifiers with different number of trees.

Figure 7.13: Importance of each variable for learning theC2classifier.

Figure 7.14: z2distribution for signal, normalization,D∗∗lν,BB¯combinatorial, and continuum events.