Hello, I managed to produce tcl files with BbkDatasetTcl and also to split individual runs with e.g. BbkDatasetTcl -ds SP-B0B0bar-Run1 -t 10000 --splitruns This seems to work fine as it increases the first event by 10000 from one tcl file to the next. I suppose we'll definitely need this as the data collections contain ~20M events. Now I have to figure out what datasets to use. Looking at the list of available datasets my guess would have been that we need the following: For generic BB MC: SP-B0B0bar-Run1 SP-B0B0bar-Run2 SP-B0B0bar-Run3 SP-B0B0bar-Run4 SP-BpBm-Run1 SP-BpBm-Run2 SP-BpBm-Run3 SP-BpBm-Run4 Data: AllEventsSkim-Run1-OnPeak-R14 AllEventsSkim-Run2-OnPeak-R14 AllEventsSkim-Run3-OnPeak-R14 (still waiting for run 4) Could someone please confirm this or tell me the names of the correct datasets? Is there a list with description somewhere? (what are the BRecoTo..., BSemiExcl.. and BtoXGamma datasets?) >From Daniele's posting on the Breco HN I figuered we need the following for Signal: cocktail SP-2223 2315 2222 Vub incl SP-3037 3617 SP-3618 2575 Vub excl SP-994 1060 1057 1058 SP-4113 1059 SP-4759 4760 4761 4762 SP-4763 4764 4766 b->sgamma SP-3135 2534 2799 2800 SP-2535 2536 3464 3134 SP-2537 2801 2802 2538 2539 SP-3463 3452 3453 3454 3455 SP-3561 3458 3459 3460 3461 SP-3462 3442 3443 3444 3445 SP-3446 3447 3448 3449 3450 SP-3451 Once I have a runnable executable I can start submitting jobs to see what number of events in data and MC makes good use of the queue length. This would be bfobjy, right? Cheers, Henning On Thu, 15 Apr 2004, Urs Langenegger wrote: > > Hoi Thorsten, > > maybe you two should subscribe to this mailing list. > > > Even though that may be technically possible, it may give you the wrong answer : > > Not that "CM2-converted" data is not really a conversion, it is a re-processing > > starting with the release-12 mini, but using a lot of new reconstruction > > algorithms. Especially for the muon-id, the extrapolation of tracks through > > the IFR is very different in CM2-converted data (or MC). So > > > > "SP5-Objy != CM2-converted SP5" > > > Apparently there was a discussion about this at the PAC meeting and it > seems that "they" considered it OK for analyses to use OBJY SP5. Is > there a solution to this? (Apart from dropping muons? :-) > > > > What has been the strategy so far ? One run per file ? > > > > In CM2, there are usually many runs per collection, and so far I did not manage > > to squeeze out tcl files which process single runs (or a self-defined run range) > > from a given collection. Does anybody know how to do this ? > > Even in the old setup long runs were too CPU-intensive to fit into one > job. The solution was to split the jobs, running only over n events in > the first tcl, while skipping the first n events in the second job. It > has been advertised that BbkDatasetTcl can generate tcl files to cope > with that problem. We have not validated anything, so far. Henning is > working on this. > > > Also, it seems that in it's current state, BbkDataSetTcl delivers somewhat > > unordered tcl-files, in the sense that subsequent "input add" lines do > > not contain collections which are subsequent in terms of run numbers. > > As long as we don't double-count, loose event or cross RUN-{1,2,3,4} > boundaries in one tcl file, there is no problem if the events are not > time (or run) ordered. > > > > > + The size of the tcl files needs to be optimized for the > > > queue length. (kanga?) > > > > You should contact the experts about that. In our PID-tuple production, the > > only reasonnable queue was "bfobjy", which we were told was "illegal". However, > > kanga and xlong had too few machines assigned at that time. Maybe it has changed > > in the meanwhile ??? > > I think the official policy for SP5 OBJY would still be bfobjy, and > CM2 would be non-bfobjy. I am sure they'll tell us when we commit > crimes. > > > > the unskimmed SP5 OBJY (700/fb!) on a relatively(?) short > > > timescale. > > > > One more comment about "unskimmed" data. If you have collections which are > > the output of a "Release12 -> CM2" conversion, you should be aware that > > the tag-part of the data has NOT been converted. So raw "converted CM2" > > data > > still contains the tag-bits as they were in release-12. No re-computation > > of the tag bits is done during conversion. This is why we are supposed to > > use the skims. In the skims (including the "AllEvents" - Skim), the tag > > bits are correct. > > Good to know! The plan is run on skimmed CM2. At least the data. Not > sure yet about SP6. > > > > BTW, how are people using these ntuples ? Are analyses run directly from these > > ntuples, or does each analysis have its own set of "reduced ntuples" which > > are extracted from these "event-store" ntuples ? > > I don't know of anybody who does not produce reduced ntuples. > > Cheers, > --U. > > ------------------------------------------------------------------ | Henning Flaecher | | Physics Department [log in to unmask] | | Royal Holloway Phone: +44 (0)1784 443479 | | University of London Fax: +44 (0)1784 472794 | | Egham, Surrey TW20 0EX, UK http://www.pp.rhul.ac.uk/~henning | | | | SLAC M/S 35 | | 2575 Sand Hill Road [log in to unmask] | | Menlo Park, CA 94025, USA Phone: (+1) (650) 926 5269 | ------------------------------------------------------------------