Hi,
this is an update on the killer performances.
1st) The code compiles, the exec runs (this is not so obvious ;-) )
2nd) I've performed a trial over 50k ev: it works fine.
3rd) I've ran over D0 run1 data.
I've found 41661 duplicates, I've ran over 506532
The GIGArootfile produced has 464871 entries (right number !!!)
The ran took :
########################
Resource usage summary:
CPU time : 52544.90 sec.
Max Memory : 103 MB
Max Swap : 129 MB
Max Processes : 3
########################
The file produced is 1.59 Gb large (like the ensamble of all the rootfiles
for d0 run1) so we don't save any space. We only get one file instead of
hundred...
So all seems to work fine, but....
a) the job ended with:
###############################
Writing outp3/sx-d0-run1-data.root
*** Break *** bus error
------------------------------------------------------------
Sender: LSF System <lsf@barb0484>
Subject: Job 644308: <./Killer -c chains/d0/sx-d0-run1-data -D outp3 -w
outp3/prova.root -k list_D0> Exited
###################################
When I try to load the file in root I get:
##########################################
CINT/ROOT C/C++ Interpreter version 5.15.07, July 7 2001
Type ? for help. Commands must be C++ statements.
Enclose multiple statements between { }.
root [0] TFile f = TFile("outp3/prova.root");
Warning in <TFile::TFile>: file outp3/prova.root probably not closed,
trying to recover
Warning in <TFile::Recover>: successfully recovered 3 keys
###########################################
The recovering takes a while...
But then I can...
###########################################
root [1] f->ls();
TFile** outp3/prova.root
TFile* outp3/prova.root
KEY: TList StreamerInfo;1 Doubly linked list
KEY: TTree h9;46 CompBRecoNtuple
KEY: TTree h9;45 CompBRecoNtuple
root [2] h9->Print();
...........................
...........................
...........................
*............................................................................*
*Br 476 :energyGam : energyGam[nGam]/F
*
*Entries : 464871 : Total Size= 31445076 bytes File Size =
24358678 *
*Baskets : 4172 : Basket Size= 8000 bytes Compression= 1.29
*
*............................................................................*
*Br 477 :B0RecGam : B0RecGam[nGam]/I
*
*Entries : 464871 : Total Size= 31440904 bytes File Size =
1555677 *
*Baskets : 4172 : Basket Size= 8000 bytes Compression= 20.21
*
*............................................................................*
*Br 478 :chBRecGam : chBRecGam[nGam]/I
*
*Entries : 464871 : Total Size= 31445076 bytes File Size =
3619975 *
*Baskets : 4172 : Basket Size= 8000 bytes Compression= 8.69
*
*............................................................................*
###########################################
......the number of entries is correct and I can make meaningful plots:
Mes
http://www.slac.stanford.edu/~asarti/recoil/killing/prova1.eps
De
http://www.slac.stanford.edu/~asarti/recoil/killing/prova2.eps
I need the expert opinion: Urs, do we have to worry about this problem?
Do you think that the problem will become larger for larger data?
There is a way to find out how much data we can 'pack' in one file without
this problem?
Now I'm waiting for reprocessed run1+run2 TS to work on 50fb-1
Comments and questions are welcomed.
Alessio
______________________________________________________
Alessio Sarti
Universita' & I.N.F.N. Ferrara
tel +39-0532-781928 Ferrara
"Quod non fecerunt barbari, fecerunt Berlusconi"
"Che il bianco sia bianco, e che il nero sia nero
che uno e uno fanno due e che la scienza dice il vero....
DIPENDE !"
|