• Tidak ada hasil yang ditemukan

Signal Extraction

The choice of bandwidth parameterhinvolves minimizing the mean squared error (MISE) between the estimated ˆf and f:

MISE =∫

(fˆ− f)2= ∫

2−2

fˆ· f +∫ f2,

which can be evaluated by summing over the observed data points ofX. Practically, we use cross validation to choose the optimal bandwidth, which gives the same asymptotic accuracy:

CV(h)= ∫

2−2N1Õ

i

i(xi), where ˆfi(xi)= Nh1 Í

i,jK(xihxj)is used to evaluate MISE by removing the contri- bution from pointi.

In this analysis, we use an adaptive KDE to evaluate the 2-dimensional probability density function (PDF) of each event type. Adaptive KDE changes the bandwidth parameters depending on the densities. In lower data density regions, adaptive KDE chooses wider bandwidths to reduce the effect of outliers on the overall KDE, and provides better overall performance. The adaptive KDE of f is:

fˆ(x)= 1 N

ÕN i=1

1

hiK(x−xi hi ) hi= h×λi

λi= ( G f˜(xi))α,

(7.6)

where ˜f(xi)is a pilot estimate of f(xi), λi is local bandwidth factor to control the local smoothing, G is the geometric mean of f(xi), α is a sensitivity parameter between 0 and 1. We normally set its values to be 0.5.

We use an Epanechnikov function as the kernel function, and use grid search to find the optimal bandwidth h. The KDE using an Epanechnikov kernel is claimed to have better mean integrated squared error under the same amount of data, and is one of the popular kernels used in practice. We use thebbrcit_kdelibrary to reduce the time complexity of performing KDE fromO(n2)toO(n). It applies kd-tree [38]

algorithm and applies GPU computation to speedup the KDE computation. The learned densities for all event components for each of four subsets are shown in Fig.

7.15, 7.16, 7.17, and 7.18.

−8 −6 −4 −2 0 2 4 6 8

z2score

−8

−6

−4

−2 0 2 4 6 8

z1score

B¯ν¯τ

0.01 0.03 0.02

0.050.04 0.06

0.01 0.02 0.03 0.04 0.05 0.06

−8 −6 −4 −2 0 2 4 6 8

z2score

−8

−6

−4

−2 0 2 4 6 8

z1score

B¯Dτν¯τ

0.01

0.02

0.03 0.040.05

0.01 0.02 0.03 0.04 0.05 0.06

−8 −6 −4 −2 0 2 4 6 8

z2score

−8

−6

−4

−2 0 2 4 6 8

z1score

B¯Dlν¯l

0.03 0.05

0.100.08 0.12

0.02 0.04 0.06 0.08 0.10 0.12 0.14 0.16

−8 −6 −4 −2 0 2 4 6 8

z2score

−8

−6

−4

−2 0 2 4 6 8

z1score

B¯Dlν¯l

0.02 0.04 0.080.06 0.10

0.02 0.04 0.06 0.08 0.10 0.12

−8 −6 −4 −2 0 2 4 6 8

z2score

−8

−6

−4

−2 0 2 4 6 8

z1score

B¯D∗∗(τ/l)ν¯

0.01 0.02 0.01

0.02 0.03

0.005 0.010 0.015 0.020 0.025 0.030 0.035

−8 −6 −4 −2 0 2 4 6 8

z2score

−8

−6

−4

−2 0 2 4 6 8

z1score

Other bkg.

0.02 0.01 0.02

0.03 0.04 0.05

0.005 0.010 0.015 0.020 0.025 0.030 0.035 0.040 0.045 0.050 0.055

Figure 7.15: KDE learned densities for event components in theD+l subset.

−8 −6 −4 −2 0 2 4 6 8

z2score

−8

−6

−4

−2 0 2 4 6 8

z1score

B¯ν¯τ

0.01 0.02

0.02

0.03 0.050.04 0.06

0.01 0.02 0.03 0.04 0.05 0.06

−8 −6 −4 −2 0 2 4 6 8

z2score

−8

−6

−4

−2 0 2 4 6 8

z1score

B¯Dτν¯τ

0.01 0.02

0.03 0.050.04

0.01 0.02 0.03 0.04 0.05 0.06

−8 −6 −4 −2 0 2 4 6 8

z2score

−8

−6

−4

−2 0 2 4 6 8

z1score

B¯Dlν¯l

0.03 0.06

0.090.12 0.15

0.00 0.02 0.04 0.06 0.08 0.10 0.12 0.14 0.16 0.18

−8 −6 −4 −2 0 2 4 6 8

z2score

−8

−6

−4

−2 0 2 4 6 8

z1score

B¯Dlν¯l

0.03 0.05 0.100.08 0.12

0.02 0.04 0.06 0.08 0.10 0.12 0.14 0.16

−8 −6 −4 −2 0 2 4 6 8

z2score

−8

−6

−4

−2 0 2 4 6 8

z1score

B¯D∗∗(τ/l)ν¯

0.01

0.01 0.01 0.02

0.03 0.03

0.004 0.008 0.012 0.016 0.020 0.024 0.028

−8 −6 −4 −2 0 2 4 6 8

z2score

−8

−6

−4

−2 0 2 4 6 8

z1score

Other bkg.

0.01 0.02 0.02

0.03 0.04

0.000 0.005 0.010 0.015 0.020 0.025 0.030 0.035 0.040 0.045

Figure 7.16: KDE learned densities for event components in theD0l subset.

−8 −6 −4 −2 0 2 4 6 8

z2score

−8

−6

−4

−2 0 2 4 6 8

z1score

B¯Dτν¯τ

0.01 0.02

0.03 0.050.04

0.006 0.012 0.018 0.024 0.030 0.036 0.042 0.048 0.054

−8 −6 −4 −2 0 2 4 6 8

z2score

−8

−6

−4

−2 0 2 4 6 8

z1score

B¯Dlν¯l

0.04 0.08 0.120.16

0.02 0.04 0.06 0.08 0.10 0.12 0.14 0.16 0.18 0.20 0.22

−8 −6 −4 −2 0 2 4 6 8

z2score

−8

−6

−4

−2 0 2 4 6 8

z1score

B¯Dlν¯l

0.03 0.06 0.120.09

0.02 0.04 0.06 0.08 0.10 0.12 0.14 0.16 0.18

−8 −6 −4 −2 0 2 4 6 8

z2score

−8

−6

−4

−2 0 2 4 6 8

z1score

B¯D∗∗(τ/l)ν¯

0.01 0.01

0.02 0.02 0.03 0.03

0.000 0.004 0.008 0.012 0.016 0.020 0.024 0.028 0.032 0.036

−8 −6 −4 −2 0 2 4 6 8

z2score

−8

−6

−4

−2 0 2 4 6 8

z1score

Other bkg.

0.01

0.01

0.02 0.02

0.02 0.040.03

0.000 0.004 0.008 0.012 0.016 0.020 0.024 0.028 0.032 0.036

Figure 7.17: KDE learned densities for event components in theD+l subset.

−8 −6 −4 −2 0 2 4 6 8

z2score

−8

−6

−4

−2 0 2 4 6 8

z1score

B¯Dτν¯τ

0.01

0.03 0.02 0.04 0.050.06

0.01 0.02 0.03 0.04 0.05 0.06

−8 −6 −4 −2 0 2 4 6 8

z2score

−8

−6

−4

−2 0 2 4 6 8

z1score

B¯Dlν¯l

0.04

0.08

0.120.16

0.02 0.04 0.06 0.08 0.10 0.12 0.14 0.16 0.18 0.20 0.22

−8 −6 −4 −2 0 2 4 6 8

z2score

−8

−6

−4

−2 0 2 4 6 8

z1score

B¯Dlν¯l

0.03 0.06 0.09 0.150.12

0.02 0.04 0.06 0.08 0.10 0.12 0.14 0.16 0.18

−8 −6 −4 −2 0 2 4 6 8

z2score

−8

−6

−4

−2 0 2 4 6 8

z1score

B¯D∗∗(τ/l)ν¯

0.01

0.01 0.01

0.02 0.03

0.004 0.008 0.012 0.016 0.020 0.024 0.028

−8 −6 −4 −2 0 2 4 6 8

z2score

−8

−6

−4

−2 0 2 4 6 8

z1score

Other bkg.

0.01

0.02 0.03 0.050.04

0.006 0.012 0.018 0.024 0.030 0.036 0.042 0.048 0.054

Figure 7.18: KDE learned densities for event components in theD0l subset.

Maximum likelihood estimation

We extract the signal and normalization yields from each subset (channel_label) individually, and then combine them to get the overall yields. Given a subset withC components and the corresponding learned densities fj(z1,z2) (j = 1,2, ...,C), the z=(z1,z2)distribution of the subset can be written as

Prob(z)= ÕC

j=1

pjfj(z),

where pj is the proportion of j’s component. If we observeN events in the subset {z}i=1N , the likelihood is

L = ÖN

i=1

Prob(zi)= ÖN

i=1

©

­

« ÕC

j=1

pj · fj(zi

®

¬ .

The MLE estimate of component proportionspj (j = 1,2, ...,C)can be obtained by solving the following convex optimization problem:

p1,pmax2,...,pC

ÖN i=1

©

­

« ÕC

j=1

pj · fj(zi

®

¬ s.t.

Õ

j

pj =1.

(7.7)

We use the CVXOPT [39] python package to implement the above estimation. The procedure is applied to the generic MC to evaluate its performance. To estimate the statistical sensitivity, we bootstrap the generic MC 900 times and solve the MLE for each bootstrap sample. The standard deviations of the 900 estimated proportions are used as absolute statistical uncertainties for the MLE estimation. The extracted yields and the absolute statistical uncertainties for all the four subsets are listed in Table 7.9. The projected fitting results onz1 andz2 scores are shown in Fig. 7.19 and Fig. 7.20.

−8 −6 −4 −2 0 2 4 6

z1score

0.000 0.025 0.050 0.075 0.100 0.125 0.150 0.175

Density ¯B→Dτ¯ντ

B¯→Dτν¯τ B¯→Dlν¯l B¯→Dlν¯l

B→D¯ ∗∗(τ/l)ν¯ Other bkg.

Data

−8 −6 −4 −2 0 2 4 6

z1score

0.000 0.025 0.050 0.075 0.100 0.125 0.150 0.175

Density ¯B→¯ντ

B→¯ Dτν¯τ B→¯ Dlν¯l B→¯ Dlν¯l

B¯→D∗∗(τ/l)ν¯ Other bkg.

Data

−8 −6 −4 −2 0 2 4 6

z1score

0.00 0.05 0.10 0.15 0.20

Density 0.25 B→D¯ τν¯τ

B→Dl¯ ν¯l B→D¯ lν¯l

B→¯ D∗∗(τ/l)¯ν Other bkg.

Data

−8 −6 −4 −2 0 2 4 6

z1score

0.00 0.05 0.10 0.15 0.20

Density ¯B→Dτ¯ντ

B→¯ Dlν¯l B→¯ Dlν¯l

B¯→D∗∗(τ/l)ν¯ Other bkg.

Data

Figure 7.19: Comparison of z1 score distributions of the data with the projections of fit results for (upper left)D+l, (upper right)D0l, (bottom left)D+l, and (bottom right)D0l subsets.

Subset Component Extracted Yield

D+l

B→ Dτν 117±44 B→ Dτν 191±57 B→ Dlν 1965±110 B→ Dlν 1261±146 B→ D∗∗lν 3248±233 Other Bkgs 7258±182

D0l

B→ Dτν 726±273 B→ Dτν 2452±285

B→ Dlν 12012±555 B→ Dlν 22808±648 B→ D∗∗lν 15594±445 Other Bkgs 22065±307 D+l

B→ Dτν 110±17 B→ Dlν 614±127 B→ Dlν 1246±139 B→ D∗∗lν 292±54 Other Bkgs 665±42 D0l

B→ Dτν 175±25 B→ Dlν 479±126 B→ Dlν 2154±149 B→ D∗∗lν 1387±101 Other Bkgs 1415±70

Table 7.9: Fit results for the yields of all the components in four subsets.

−8 −6 −4 −2 0 2 4 6

z2score

0.00 0.02 0.04 0.06 0.08 0.10 0.12 0.14

Density ¯B→Dτ¯ντ

B¯→Dτν¯τ B¯→Dlν¯l B¯→Dlν¯l

B→D¯ ∗∗(τ/l)ν¯ Other bkg.

Data

−8 −6 −4 −2 0 2 4 6

z2score

0.00 0.05 0.10 0.15

Density 0.20 B→¯ ν¯τ

B→¯ Dτ¯ντ B→¯ Dlν¯l B→¯ Dlν¯l

B→¯ D∗∗(τ/l)ν¯ Other bkg.

Data

−8 −6 −4 −2 0 2 4 6

z2score

0.00 0.05 0.10 0.15 0.20 0.25 0.30 0.35 0.40

Density ¯B→Dτ¯ντ

B¯→Dlν¯l B¯→Dlν¯l

B→D¯ ∗∗(τ/l)ν¯ Other bkg.

Data

−8 −6 −4 −2 0 2 4 6

z2score

0.00 0.05 0.10 0.15 0.20 0.25 0.30

Density ¯B→Dτ¯ντ

B→¯ Dlν¯l B→¯ Dlν¯l

B→¯ D∗∗(τ/l)ν¯ Other bkg.

Data

Figure 7.20: Comparison of z2 score distributions of the data with the projections of fit results for (upper left)D+l, (upper right)D0l, (bottom left)D+l, and (bottom right)D0l subsets.

Cross-check on PDFs modeling

Cross-checks are performed to study the modeling of PDF shapes. We first check the PDF modeling of normalization events by comparing the PDFs of MC with data on the normalization enriched region. The normalization enriched region is defined in(z1,z2)space asz1 > 2,z2 < −4, in which 94% events are normalization decays.

The MC and data comparison shown in Fig. 7.21 indicates the normalization PDF shapes are well modeled.

2.0 2.5 3.0 3.5 4.0 4.5 5.0

0.0 0.1 0.2 0.3 0.4 0.5 0.6 0.7

Density

B→¯ D(∗)τν¯τ B→¯ D(∗)lν¯l B→¯ D∗∗(τ/l)ν¯

BB¯combinatorial Continuum

2.0 2.5 3.0 3.5 4.0 4.5 5.0

z1score

−1 0 1

(Data-MC)/Data

−8.0 −7.5 −7.0 −6.5 −6.0 −5.5 −5.0 −4.5 −4.0 0.0

0.1 0.2 0.3 0.4 0.5

Density

B→¯ D(∗)τν¯τ B→¯ D(∗)l¯νl B→¯ D∗∗(τ/l)ν¯

BB¯combinatorial Continuum

−8.0 −7.5 −7.0 −6.5 −6.0 −5.5 −5.0 −4.5 −4.0 z2score

−0.5 0.0 0.5

(Data-MC)/Data

Figure 7.21: MC/data comparison ofz1andz2score distributions in normalization enriched region.

We then check the PDF modeling of backgrounds (after on-peak background cali- bration) by comparing the PDFs of MC with data on background enriched region.

The background enriched region is defined as−4 ≤ z1 ≤ −2, in which 97% events are either B → D∗∗lν, BB¯ combinatorial, or continuum events. The MC and data comparison shown in Fig. 7.22 indicates the PDF shapes of backgrounds are also well-modeled.

−4.00 −3.75 −3.50 −3.25 −3.00 −2.75 −2.50 −2.25 −2.00 0.0

0.2 0.4 0.6 0.8 1.0

Density

B→D¯ (∗)τν¯τ B→D¯ (∗)lν¯l B→D¯ ∗∗(τ/l)¯ν

Bcombinatorial Continuum

−4.00 −3.75 −3.50 −3.25 −3.00 −2.75 −2.50 −2.25 −2.00 z1score

−0.25 0.00 0.25

(Data-MC)/Data

−8 −6 −4 −2 0 2 4 6 8

0.000 0.025 0.050 0.075 0.100 0.125 0.150 0.175

Density

B→D¯ (∗)τν¯τ B→D¯ (∗)lν¯l B→D¯ ∗∗(τ/l)ν¯

Bcombinatorial Continuum

−8 −6 −4 −2 0 2 4 6 8

z2score

−1 0 1

(Data-MC)/Data

Figure 7.22: MC/data comparison ofz1 and z2score distributions in backgrounds enriched region.

7.7 Systematic Uncertainties