Print

Print


I also understand the question about the correlations.

Heiko

On Thu, 27 Jul 2006, Wolfgang Menges wrote:

> Hi Concezio,
> 
> 	as Heiko said, very promissing. Can you run with 50% and 100% correlation?
> 
> Cheers,
> 
> 	Wolfgang
> 
> Concezio Bozzi wrote:
> > Hi all, 
> > 
> > I run a test to estimate the systematic uncertainty due to the
> > uncertainty on S/P in mES data fits in the following way: 
> > 
> > 1) Take the S/P ratios determined as 
> > 
> > S/P(data_enriched) = [S/P(MC_enriched) / S/P(MC_depleted)] * S/P
> > (data_depleted)
> > 
> > I have used the values which I have been recently playing with, i.e. 
> > #mx_l mx_h  corr   err_corr
> > 0.00 1.55 1.499 +- 0.495
> > 1.55 1.90 2.688 +- 0.655
> > 1.90 2.20 1.801 +- 0.296
> > 2.20 2.50 1.896 +- 0.611
> > 2.50 2.80 1.165 +- 0.468
> > 2.80 3.10 0.637 +- 0.311
> > 3.10 3.40 19.367+- 34.585
> > 3.40 3.70 1.524 +- 1.610
> > 3.70 4.20 8.180 +- 31.833
> > 4.20 5.00 0.555 +- 6.639
> > 
> > No attempt to fit a n-th order polynomial, just take the values as they
> > come out of the single bin-by-bin fits on data depleted, MC enriched and
> > depleted. 
> > Note that the relative errors are quite large (e.g. 33% on the first
> > bin, 25% on the second, 22% on the third, 32% on the 4th, higher and
> > higher as mX increases). 
> > 
> > 2) Randomize simultaneously the 10 above values according to a gaussian
> > distribution whose mean is the correction (column corr) and whose sigma
> > is the uncertainty (err_corr). The random number is of course different
> > for each mX bin. 
> > 
> > 3) Fit with VVF by using the randomized S/P of point 2) 
> > 
> > 4) go to 2), change the random seed, repeat 100 times
> > 
> > Results of the 100 jobs are in 
> > http://www.slac.stanford.edu/~bozzi/scra/Ibu_SP_*
> > *=1,...,100
> > 
> > Take the (width/mean) ratio of the resulting 100 fits as systematic
> > uncertainty: 
> > 
> > yakut02(~:) grep "BRBR           " ~bozzi/scra/Ibu_SP_*/*dat | awk
> > 'BEGIN{sum=0; sum2=0}{sum+= $3; sum2+=$3*$3; num++}END{print sum/num;
> > print sqrt(sum2/num-sum*sum/num/num)}'
> > 0.0291231
> > 0.00188286
> > 
> > The relative uncertainty is 0.00188/0.02912 = 6.46% i.e. 3.2% on Vub. 
> > This is somewhat lower than a naive argument which can be used (see
> > below) to give the error on the fitted Vub events in the first bin and
> > which give about twice (13.2%) the error on BRBR. I think the reason for
> > this is that the errors on the first 4 bins are comparable, which
> > reduces the lever arm and therefore the variation in the first bin. 
> > 
> > Quite promising, isn't it? 
> > 
> > Concezio. 
> > 
> > 
> > PS: here is the naive argument on the uncertainty on the number of
> > signal events in the first bin, which translates in the uncertainty on
> > BRBR. We have
> > 
> > N_signal = N_data - N_argus - N_peaking 
> > 
> > N_peaking = N_signal * 1/corr
> > (corr is the S/P correction factor)
> > 
> > Solving for N_signal:
> > 
> > N_signal = [corr / (1+ corr)] * [N_data - N_argus] 
> > 
> > Error propagation gives: 
> > 
> > delta(N_signal) / N_signal = [delta(corr) / corr] / (1+corr) 
> > 
> > Taking corr = 1.499, delta(corr)/corr = 0.33 we get 
> > 
> > delta(N_signal) / N_signal = 13.2%
> > 
> 
> -- 
> ------------------------------------------------------------------------
> Wolfgang Menges
> Queen Mary, University of London               SLAC, MS 35
> Mile End Road                                  2575 Sand Hill Road
> London, E1 4NS, UK                             Menlo Park, CA 94025, USA
> +44 20 7882 3753                               ++1 650 926 8503
>                                                 [log in to unmask]
> ------------------------------------------------------------------------
>