Sie sind auf Seite 1von 7

Test 1

Edgar Mauricio Loza Ramrez


March 9, 2016
0.1

Find the constant of normalization N

The pdf for the background noise of events of type b that describes their mass
distribution is
1 cm
,
(1)
N
where c is a constant that describes the slight slope seen in the mass histogram. From probability we know that a pdf f (x) must be normalized to the
unity when integrated all over the interval where it is valid. For this case,
Z mx
1 cm
dm = 1.
(2)
N
mn
Bkgb (m) =

where mx and mn are the upper and lower values of the masses presented
in the histogram. Easily we find the form of N to be
1
N = (mx mn ) c(m2x m2n ).
2

0.2

(3)

Draw the respective histograms of each column

Using R, the mass and decay length histograms were drawn. They are presented
in the next two figures.

0.3

Find the parameters with their respective uncertainties

The Maximum Likelihood method requires a numerical method such as the bisection method to determine the values of the parameters for which a pdf, or
a combination of them, better describes a histogram. The Newton-Raphson
method was used in this case.
The parameter c was approximated by equation (2) and some facts about
the histogram. Looking at the mass histogram
mx 8

mn 2,

so we can easily deduce


N = 6 30c.

Figure 1: Mass distribution

Figure 2: Length distribution


Now the distribution involving the background noise of events b is a line
with a negative slope and positive intersection with the y-axis, that is

cm
< 0,
N

1
> 0.
N

From the inequalities and as m > 0 we conclude


c > 0,

N > 0.

However, at the same time 6 30c > 0 or 1/5 > c. Therefore c [0, 51 ],
giving an idea of a good approximation for c.
The rest of the unknown parameters, sigmas and fractions of the corresponding pdfs as a sum of a total pdf, were determined by means of nested for loops.
The initial parameters used as a approximation for the Maximum Likelihood
method are shown
#s1 = 0 . 5 , s2 = 0 . 2 , a l p h a = 0 . 2 , b e t a = 0 . 3 , gama = 0 . 2
x 0 < c ( 6 . 3 6 , 3 . 2 7 , 4 6 0 , 4 7 0 , 0 . 1 , 1 0 0 , 7 0 0 , 0 . 5 , 0 . 2 , 0 . 2 , 0 . 3 , 0 . 2 )
Thanks to the multivariable Newton-Raphson method these approximations
were improved. To calculate the uncertainty of every parameter the relation
1
2
was used. The table contains the estimators and their uncertainties

log L(
) = log L()

Estimator
m(p1 )
m(p2 )
lambda(p1 )
lambda(p2 )
c
a
b
1
2

n(p1 )
n(p2 )
na
nb

Table 1: Estimators and uncertainties


Code name Central point
l
m p1
6.379841
0.015312
m p2
3.278644
0.000000
lamb p1
241.213819
0.000000
lamb p2
472.462038
0.000000
c
0.088427
0.000000
lamb a
107.025966
0.000000
lamb b
655.319697
0.000000
s1
0.482306
0.000000
s2
0.247797
0.000000
alpha
0.132002
0.000000
beta
0.255972
0.000000
gama
0.239112
0.000000
-2181.341
0
-4229.944
0
-3951.326
0
-6162.389
0

(4)

r
0.015950
0.000328
0.042693
0.047308
NA
NA
0.069552
NA
NA
0.000015
0.000028
0.000025
0.2423712
0.4699938
0.4108689
0.2654876

the number of particles (n(p1 ), n(p2 )) and events from the background (na , nb )
are also included. The estimators whose l is 0 and whose r is NA (not apply)
could be caused by the sensibility of the source code (see the graphs in the
appendix).

Figure 3: Expectation mass

Figure 4: Expectation length decay

0.4

P - value of the full model

The good-of-fit of the model presented here was tested. The Pearsons 2 test
was performed. For a histogram with N bins

2 =

N
X
n2 v 2
i

i=1

vi2

(5)

where vi is the expected number of entries belonging to the i-th bin and ni
is the number of real entries. 2 corresponds to the discrepancy between data
and the hypothesis. Then the P -value is
Z
f (z; nd )dz
(6)
P =
2

where nd is the number of degrees of freedom. The P-value obtained for the
mass distributions was 0. Similarly the P-value for the decay length distribution
was 0. This means the probability of finding a big discrepancy between data
and the hypothesis is small.

0.5

Full and alternative hypothesis

A similar test can be done to test two different hypothesis. Our full model has
a total of 12 parameters. The alternative hypothesis contains just 11, assuming
that lambda(p1 ) = lambda(p2 ). The 2 takes the form
2 = 2 log

L()
L(0 )

(7)

is the likelihood value with full parameters and L(0 ) is the


where L()
function likelihood with restrictions regarding the number of parameters (alternative hypothesis.) The P -value estimated was 0.0078032652348644093721.
Even when this is a tiny number, we cannot reject with a 99.9% confidence the
model with parameters 0 .

0.6

Appendix

Figure 5: No existence of r

Figure 6: No existence of r

Figure 7: No existence of r

Figure 8: No existence of r

Das könnte Ihnen auch gefallen