I am also still seeing a problem in the recon files where we have clusters but nothing recorded in the ECalTrackerHits collection. I was specifically looking in the beam-tri pass3/test recon files. On Mon, Nov 9, 2015 at 2:30 PM, Sho Uemura <[log in to unmask]> wrote: > It looks like the bad files are all pairs1 now? It looks like you reran > recon but not readout for those, and I see that the readout files are also > small. I think the pairs1 files went bad in the readout step, not recon. > > For example, > /work/hallb/hps/mc_production/pass3/logs/readout/beam-tri/1pt05/egsv3-triv2-g4v1_HPS-EngRun2015-Nominal-v3-fieldmap_3.4.1_pairs1_41.out > shows 85 events written (consistent with the count of recon events in your > test job). If you look at the EventMarkerDriver output in the .err file you > can see that the readout sim stopped partway through the 4th SLIC output > file (out of 100). We get ~20 pairs1 triggers per beam-tri SLIC output > file, so 85 events is consistent with the readout sim running normally > until it stopped at that point. > > I checked that the SLIC output file > (egsv3-triv2-g4v1_HPS-EngRun2015-Nominal-v3-fieldmap_4003) where the > readout sim stopped has the right number of events and is of typical file > size, and there's nothing unusual in its logfiles. I don't think this is a > stdhep problem, and I think the SLIC output file is fine. > > Maybe this was just some glitch on the farm node running the readout job? > Not sure. Can you try rerunning readout? > > BTW, I am sure SLIC had nothing to do with the singles1 problem that now > seems to be fixed. I don't think it's responsible for this problem, either. > > > On Mon, 9 Nov 2015, Bradley T Yale wrote: > > The beam-tri test files, which should have used the updated SLIC, still >> seems to have the problem: >> >> cat >> /work/hallb/hps/mc_production/pass3/test/data_quality/recon/beam-tri/1pt05/*.txt >> | grep "^Read " >> Read 111 events >> Read 1987 events >> Read 2014 events >> Read 2013 events >> Read 2094 events >> Read 2094 events >> Read 1989 events >> Read 2083 events >> Read 2070 events >> Read 1887 events >> Read 2007 events >> Read 1955 events >> Read 2037 events >> Read 2013 events >> Read 1991 events >> Read 1900 events >> Read 2002 events >> Read 1996 events >> Read 1835 events >> Read 85 events >> Read 1914 events >> Read 111 events >> Read 98 events >> Read 202 events >> Read 114 events >> Read 155 events >> Read 2007 events >> Read 59 events >> Read 1800 events >> Read 2052 events >> Read 42253 events >> Read 42023 events >> Read 41771 events >> Read 42099 events >> Read 41720 events >> Read 42093 events >> Read 41917 events >> Read 41522 events >> Read 41876 events >> Read 41711 events >> Read 42058 events >> Read 42478 events >> Read 41945 events >> Read 42078 events >> Read 42181 events >> Read 42177 events >> Read 42189 events >> Read 42061 events >> Read 42117 events >> >> singles1 should have ~42000 events, and pairs1 should have ~2000. >> This sample contains the original problem files (5* and 9*), and they >> seem fixed, but now files like 15, 41, and 45 are bad. >> >> So this may point to a random stdhep read error then, maybe from that >> header issue? >> >> >> >> ________________________________________ >> From: McCormick, Jeremy I. <[log in to unmask]> >> Sent: Friday, November 6, 2015 3:48 PM >> To: Bradley T Yale >> Subject: RE: beam-tri singles MC problems >> >> Okay, that's great. Appreciate your hard work and responsiveness on this. >> >> Can you let me know when these have finished? I'll try to go over the >> data very carefully to make sure no more dumb bugs are lurking in there. >> >> -----Original Message----- >> From: Bradley T Yale [mailto:[log in to unmask]] >> Sent: Friday, November 06, 2015 12:45 PM >> To: McCormick, Jeremy I.; Holly Vance; Uemura, Sho >> Cc: hps-software >> Subject: Re: beam-tri singles MC problems >> >> Working on it. The SLIC beam-tri looks ok now, but will verify at recon. >> >> I'm going to put these samples at every stage of production in >> /mss/hallb/hps/production/pass3/test/ >> >> and all logs, dq, etc. in >> /work/hallb/hps/mc_production/pass3/test/ >> >> >> ________________________________________ >> From: [log in to unmask] <[log in to unmask]> on >> behalf of McCormick, Jeremy I. <[log in to unmask]> >> Sent: Friday, November 6, 2015 3:39 PM >> To: Holly Vance; Uemura, Sho >> Cc: Bradley T Yale; hps-software >> Subject: RE: beam-tri singles MC problems >> >> Can we rerun a limited subset of all the data including all the MC event >> types for pass3 with these fixes? >> >> Then I think we should do QA before relaunching all the jobs to fix the >> scoring plane issue. >> >> -----Original Message----- >> From: Holly Vance [mailto:[log in to unmask]] >> Sent: Friday, November 06, 2015 12:33 PM >> To: Uemura, Sho >> Cc: Bradley T Yale; McCormick, Jeremy I.; hps-software >> Subject: Re: beam-tri singles MC problems >> >> Hi all, >> >> The ECal scoring plane needs to be working. It's a good check on the >> track projection to the Ecal face (which was not working in Pass3), and >> it's the only way to verify certain cluster property corrections in the >> Ecal. >> >> -Holly >> >> On Fri, Nov 6, 2015 at 3:27 PM, Sho Uemura <[log in to unmask]> >> wrote: >> >> >> Looks good to me. >> >> For people who care: This is nothing to do with SLIC, since the >> error was coming from one of the standalone stdhep utilities >> (/u/group/hps/hps_soft/stdhep/bin/beam_coords) that reads in a stdhep file >> from tape. But neither the utility nor the input stdhep files have changed >> since pass2, and this error did not happen in pass2. My theory is that the >> cache copy of the stdhep file was corrupt. >> >> Anyway, I think this problem is fixed now. The affected beam-tri >> and tritrig-beam-tri must be rerun, but we should decide if the ECal >> scoring plane fix merits rerunning all of the pass3 MC. >> >> >> On Fri, 6 Nov 2015, Bradley T Yale wrote: >> >> >> >> Re-running the problem files in quarantine, it looks like >> the same stdhep files are being read now: >> >> /work/hallb/hps/mc_production/pass3/test/logs/slic/beam-tri/1pt05/egsv3-triv2-g4v1_HPS-EngRun2015-Nominal-v3-fieldmap_952.out >> >> Maybe the latest SLIC update worked. I can redo the >> (tritrig)-beam-tri and see if it's fixed. >> As mentioned, I don't see this problem in the other MC >> components, only beam-tri. >> If you REALLY want to be safe, I can re-run everything, >> but would at least like to do it with a post-release jar (3.4.2-SNAPSHOT or >> 3.4.2-20151014.013425-5) so we can test current things. >> >> ________________________________________ >> From: Sho Uemura <[log in to unmask]> >> Sent: Thursday, November 5, 2015 9:39 PM >> To: Bradley T Yale >> Subject: Re: beam-tri singles MC problems >> >> I tried running beam_coords on the farm >> (/work/hallb/hps/uemura/bradtest/beam-tri_100.xml, >> logfiles and output in >> same directory) and it works fine. >> >> I looked at beam-tri logs for pass2 for the same files, >> and they are fine. >> >> So this stuff worked in pass2, broke in pass3, works again >> now, but >> nothing has changed - same stdhep file, and the >> beam_coords binary hasn't >> changed. >> >> Can you try rerunning the slic beam-tri job? It could be >> something weird >> like jcache screwing up and not copying the file correctly >> from tape - >> that would affect every job in that run but not runs >> before or after. >> >> On Thu, 5 Nov 2015, Sho Uemura wrote: >> >> >> >> It looks like the problem is that beam_coords is >> having trouble reading the >> beam.stdhep file and crashes, and so the >> beam-tri.stdhep file that goes into >> SLIC is missing all the beam background, and the >> trigger rate ends up being >> ridiculously low. Of course this affects every >> SLIC run that uses that >> beam.stdhep file. >> >> I get that from looking at >> >> /work/hallb/hps/mc_production/pass3/logs/slic/beam-tri/1pt05/egsv3-triv2-g4v1_HPS-EngRun2015-Nominal-v3-fieldmap_952.err >> and >> >> /work/hallb/hps/mc_production/pass3/logs/slic/beam-tri/1pt05/egsv3-triv2-g4v1_HPS-EngRun2015-Nominal-v3-fieldmap_952.out >> and comparing to other runs - you'll see that the >> .out file is missing some >> printouts after "Rotating Beam" and >> rot_beam.stdhep is missing from the file >> list. For example, one of the first things >> beam_coords should print is the >> number of events in the input file. >> >> So there must be something wrong with that stdhep >> file, but it has nothing to >> do with SLIC. Is it possible that this has always >> been happening, in pass2 >> and earlier? I'll look at log files. >> >> Weirdly I have no difficulty running beam_coords >> on egsv3_10.stdhep on ifarm. >> Maybe there's something different about the batch >> farm environment? >> >> The bad news is that this must affect every MC >> that has beam background or >> beam-tri mixed in. >> >> On Thu, 5 Nov 2015, Bradley T Yale wrote: >> >> >> >> First, I submitted a report about those >> otherwise successful jobs not being >> written to tape, and it turned out to be a >> system glitch. It appears fixed >> now and unrelated to the following, >> which only affects ~15% of Pass3 beam-tri >> and tritrig-beam-tri files but no >> other Pass3 MC components. >> >> The beam-tri files that were readout >> 10-to-1 have the same problem with an >> inconsistent # of events, so it wasn't a >> problem with time/space allottment >> for the jobs. >> A few recon files with no time limit set >> for the jobs (100-to-1, labelled >> 'NOTIMELIMIT') made it through before the >> tape-writing glitch as well, and >> have the same problem. >> >> Digging a little further, it appears that >> this issue with readout event >> inconsistency is likely related to the >> stdhep file-reading problem that >> Jeremy found while fixing SLIC for >> v3-fieldmap, so I brought him into this. >> Let me motivate that conclusion... >> >> About 85% of Pass3 beam-tri readout files >> look fine, and then: >> cat >> >> /work/hallb/hps/mc_production/pass3/data_quality/readout/beam-tri/1pt05/egsv3-triv2-g4v1_10to1_HPS-EngRun2015-Nominal-v3-fieldmap_3.4.1_singles1_*.txt >> | grep "^Read " >> .......... >> Read 41911 events >> Read 42775 events >> Read 41551 events >> Read 42055 events >> Read 42556 events >> Read 9 events >> Read 7 events >> Read 7 events >> Read 3 events >> Read 9 events >> Read 10 events >> Read 2 events >> Read 13 events >> Read 7 events >> Read 41529 events >> Read 8 events >> Read 42149 events >> Read 42141 events >> Read 41933 events >> Read 41856 events >> Read 41711 events >> Read 42038 events >> Read 42004 events >> Read 41997 events >> Read 42029 events >> Read 41764 events >> Read 42156 events >> Read 42245 events >> Read 41732 events >> Read 42060 events >> Read 42070 events >> Read 42060 events >> Read 41962 events >> Read 41967 events >> Read 42071 events >> Read 42067 events >> Read 42017 events >> Read 42046 events >> Read 42614 events >> Read 42655 events >> Read 42337 events >> Read 42342 events >> Read 42503 events >> Read 42454 events >> Read 42237 events >> Read 42338 events >> Read 42607 events >> Read 41791 events >> Read 42309 events >> Read 3 events >> Read 4 events >> Read 7 events >> Read 7 events >> Read 4 events >> Read 6 events >> Read 7 events >> Read 7 events >> Read 4 events >> Read 41993 events >> >> The affected 10-to-1 readout files are >> #51-60 and #91-100, which were made >> from SLIC files #501-600, and #901-1000. >> For example: >> cat >> >> /work/hallb/hps/mc_production/pass3/data_quality/readout/beam-tri/1pt05/egsv3-triv2-g4v1_10to1_HPS-EngRun2015-Nominal-v3-fieldmap_3.4.1_singles1_96.txt >> | grep "^Read " >> >> /work/hallb/hps/mc_production/pass3/logs/readout/beam-tri/1pt05/egsv3-triv2-g4v1_10to1_HPS-EngRun2015-Nominal-v3-fieldmap_3.4.1_singles1_96.out >> >> Looking at the SLIC files that were used >> for readout (e.g. #951-960): >> >> /work/hallb/hps/mc_production/pass3/logs/slic/beam-tri/1pt05/egsv3-triv2-g4v1_HPS-EngRun2015-Nominal-v3-fieldmap_952.out >> >> This shows that stdhep is not reading the >> events from 1 out of every 25 >> beam.stdhep files using the Pass3 setup. >> The actual beam.stdhep files from this >> problem (#51-60 and #91-100 in >> >> /mss/hallb/hps/production/stdhep/beam/1pt05/) look fine. >> >> Also, Pass3 tritrig-beam-tri, which are >> readout 1-to-1, have occasional >> files which contain no events. This means >> that when the beam-tri files are >> readout in larger quantities, these files >> without events shave off ~4000 >> events for each affected SLIC file used. >> This is probably why some of the >> original 100-to-1 beam-tri files appear >> light on events, and are a lot >> worse with 10-to-1. >> >> The corresponding Pass2 readout/recon, >> which used the same seed and files >> as the problem ones, seem correct though: >> cat >> >> /work/hallb/hps/mc_production/data_quality/readout/beam-tri/1pt05/egsv3-triv2-g4v1_s2d6_HPS-EngRun2015-Nominal-v1_3.4.0-20150710_singles1_9*.txt >> | grep "^Read " >> cat >> >> /work/hallb/hps/mc_production/pass2/logs/readout/beam-tri/1pt05/egsv3-triv2-g4v1_s2d6_HPS-EngRun2015-Nominal-v3_3.4.0_singles1_*.out >> | grep "events " >> >> In summary, this inconsistency at readout >> is due to beam.stdhep files >> occasionally not being able to be read >> during Pass3 SLIC jobs. >> It only affects beam-tri made using the >> updated SLIC and v3-fieldmap >> detector. >> I'll make a Jira item about it. >> >> ________________________________________ >> From: Nathan Baltzell <[log in to unmask]> >> Sent: Thursday, November 5, 2015 4:26 AM >> To: Bradley T Yale >> Cc: Sho Uemura; Omar Moreno; Matthew Solt; >> Mathew Thomas Graham >> Subject: Re: beam-tri singles MC problems >> >> Probably should submit a CCPR on the >> failing to write to tape (including >> an example failed jobid/url). I don't >> notice any related CCPRs in the >> system, >> and no corresponding errors in the >> farm_outs. >> >> >> On Nov 5, 2015, at 9:00 AM, Bradley T Yale >> <[log in to unmask]> wrote: >> >> >> >> Ok, I'll do those 10to1 as well to >> match everything else. >> >> By the way, the "failed" job >> status you see is because the trigger plots >> fail for some reason and so the >> entire job gets classified that way. >> All other output is fine though, >> and just can't be written to tape. That >> has never been an issue before, >> but I disabled the trigger plots for the >> latest batch just in case. >> It could just be something with >> the system. I'll see if it's resolved >> tomorrow. >> >> >> ________________________________________ >> From: Sho Uemura < >> [log in to unmask]> >> Sent: Thursday, November 5, 2015 >> 1:49 AM >> To: Bradley T Yale >> Cc: Omar Moreno; Matthew Solt; >> Mathew Thomas Graham; Nathan Baltzell >> Subject: Re: beam-tri singles MC >> problems >> >> pairs1 seems better - there are >> still quite a few files that run under, >> but maybe 75% have the right >> number (1 ms/file * 100 files * 20 kHz = >> 2000) of events. >> >> cat >> >> /work/hallb/hps/mc_production/pass3/data_quality/recon/beam-tri/1pt05/egsv3-triv2-g4v1_HPS-EngRun2015-Nominal-v3-fieldmap_3.4.1_pairs1_*.txt >> | grep "^Read " >> Read 111 events >> Read 1987 events >> Read 2014 events >> Read 2013 events >> Read 2094 events >> Read 2094 events >> Read 1989 events >> Read 2083 events >> Read 2070 events >> Read 1887 events >> Read 2007 events >> Read 1955 events >> Read 2037 events >> Read 2013 events >> Read 1991 events >> Read 1900 events >> Read 2002 events >> Read 1996 events >> Read 1835 events >> Read 85 events >> Read 1914 events >> Read 111 events >> Read 98 events >> Read 202 events >> Read 114 events >> Read 155 events >> Read 2007 events >> Read 59 events >> Read 1800 events >> Read 2052 events >> >> >> On Thu, 5 Nov 2015, Bradley T Yale >> wrote: >> >> >> >> Everything is failing to >> write to tape. >> >> Maybe this is also the >> cause of the badly cached dst files you were >> seeing as well. >> >> I have no idea what is >> causing this. That's why I included Nathan in >> this. >> >> >> On a side note, are you >> seeing the same inconsistency in pairs1 beam-tri, >> or just singles? >> >> >> >> ________________________________ >> From: Bradley T Yale >> Sent: Thursday, November >> 5, 2015 1:13 AM >> To: Omar Moreno; Sho Uemura >> Cc: Matthew Solt; Mathew >> Thomas Graham; Nathan Baltzell >> Subject: Re: beam-tri >> singles MC problems >> >> >> So, the 10to1 readout jobs >> successfully completed, but failed to write to >> tape: >> >> >> http://scicomp.jlab.org/scicomp/#/jasmine/jobs?requested=details&id=115214062 >> >> >> I'm trying again after >> setting 'Memory space' back to "1024 MB", which is >> what it had been before. >> >> Is there anything else >> that could be causing this? >> >> >> >> ________________________________ >> From: Bradley T Yale >> Sent: Wednesday, November >> 4, 2015 7:41 PM >> To: Omar Moreno; Sho Uemura >> Cc: Matthew Solt; Mathew >> Thomas Graham >> Subject: Re: beam-tri >> singles MC problems >> >> >> Sorry. The latest ones are >> being reconstructed now and labelled >> 'NOTIMELIMIT'. They >> shouldn't take long once active. Their readout did >> not have a time limit to >> try to fix the problem, but just in case, I'm >> also reading out others >> 10-to-1 (labelled '10to1') and will probably >> start doing it that way so >> readout doesn't take forever. >> >> >> >> >> ________________________________ >> From: [log in to unmask] >> <[log in to unmask]> on behalf of Omar >> Moreno <[log in to unmask]> >> Sent: Wednesday, November >> 4, 2015 4:24 PM >> To: Sho Uemura >> Cc: Bradley T Yale; Omar >> Moreno; Matthew Solt; Mathew Thomas Graham >> Subject: Re: beam-tri >> singles MC problems >> >> Any news on this? I'm >> transferring all of the beam-tri files over to >> SLAC and I'm noticing that >> they are still all random sizes. >> >> On Fri, Oct 23, 2015 at >> 3:33 PM, Sho Uemura >> <[log in to unmask] >> <mailto:[log in to unmask]>> wrote: >> Hi Brad, >> >> 1. readout files seem to >> be really random lengths: >> >> cat >> >> /work/hallb/hps/mc_production/pass3/data_quality/readout/beam-tri/1pt05/egsv3-triv2-g4v1_HPS-EngRun2015-Nominal-v3-fieldmap_3.4.1_singles1_*|grep >> "^Read "|less >> >> Read 52 events >> Read 16814 events >> Read 17062 events >> Read 12543 events >> Read 328300 events >> Read 355896 events >> Read 12912 events >> Read 309460 events >> Read 306093 events >> Read 313868 events >> Read 325727 events >> Read 298129 events >> Read 417300 events >> Read 423734 events >> Read 308954 events >> Read 365261 events >> Read 301648 events >> Read 316249 events >> Read 340949 events >> Read 319316 events >> Read 424033 events >> Read 308746 events >> Read 317204 events >> Read 12363 events >> Read 355813 events >> Read 329739 events >> Read 298601 events >> Read 29700 events >> Read 12675 events >> Read 287237 events >> Read 311071 events >> Read 12406 events >> Read 12719 events >> Read 30428 events >> Read 324795 events >> Read 345850 events >> Read 25765 events >> Read 29806 events >> Read 77 events >> Read 12544 events >> Read 372642 events >> Read 12779 events >> >> which makes it seem like >> jobs are failing randomly or something - I think >> normally we see most files >> have the same length, and a minority of files >> (missing some input files, >> or whatever) are shorter. In this case I think >> the expected number of >> events (number of triggers from 100 SLIC output >> files) is roughly 420k, >> and as you can see only a few files get there. >> >> I looked at log files and >> I don't see any obvious error messages, but >> maybe you have ideas? I'll >> keep digging. >> >> 2. Looks like the singles >> recon jobs are running into the job disk space >> limit, so that while >> readout files can have as many as 420k events, recon >> files never have more than >> 240k. Looks like the disk limit is set to 5 GB >> (and a 240k-event LCIO >> recon file is 5.5 GB), but it needs to be at least >> doubled - or the number of >> SLIC files per readout job needs to be >> reduced? >> >> cat >> >> /work/hallb/hps/mc_production/pass3/data_quality/recon/beam-tri/1pt05/egsv3-triv2-g4v1_HPS-EngRun2015-Nominal-v3-fieldmap_3.4.1_singles1_*|grep >> "^Read "|less >> Read 1 events >> Read 16814 events >> Read 17062 events >> Read 242359 events >> Read 243949 events >> Read 242153 events >> Read 12776 events >> Read 242666 events >> Read 244165 events >> Read 243592 events >> Read 243433 events >> Read 242878 events >> Read 241861 events >> Read 242055 events >> Read 30428 events >> Read 243156 events >> Read 241638 events >> Read 4 events >> Read 241882 events >> >> >> >> From >> >> /work/hallb/hps/mc_production/pass3/logs/recon/beam-tri/1pt05/egsv3-triv2-g4v1_HPS-EngRun2015-Nominal-v3-fieldmap_3.4.1_singles1_22.err: >> >> >> >> >> java.lang.RuntimeException: Error writing LCIO file >> at >> org.lcsim.util.loop.LCIODriver.process(LCIODriver.java:116) >> at >> org.lcsim.util.Driver.doProcess(Driver.java:261) >> at >> org.lcsim.util.Driver.processChildren(Driver.java:271) >> at >> org.lcsim.util.Driver.process(Driver.java:187) >> at >> >> org.lcsim.util.DriverAdapter.recordSupplied(DriverAdapter.java:74) >> at >> >> org.freehep.record.loop.DefaultRecordLoop.consumeRecord(DefaultRecordLoop.java:832) >> at >> >> org.freehep.record.loop.DefaultRecordLoop.loop(DefaultRecordLoop.java:668) >> at >> >> org.freehep.record.loop.DefaultRecordLoop.execute(DefaultRecordLoop.java:566) >> at >> org.lcsim.util.loop.LCSimLoop.loop(LCSimLoop.java:151) >> at >> org.lcsim.job.JobControlManager.run(JobControlManager.java:431) >> at >> org.hps.job.JobManager.run(JobManager.java:71) >> at >> org.lcsim.job.JobControlManager.run(JobControlManager.java:189) >> at >> org.hps.job.JobManager.main(JobManager.java:26) >> Caused by: >> java.io.IOException: File too large >> at >> java.io.FileOutputStream.writeBytes(Native Method) >> at >> java.io.FileOutputStream.write(FileOutputStream.java:345) >> at >> >> hep.io.xdr.XDROutputStream$CountedOutputStream.write(XDROutputStream.java:103) >> at >> java.io.DataOutputStream.write(DataOutputStream.java:107) >> at >> >> hep.io.sio.SIOWriter$SIOByteArrayOutputStream.writeTo(SIOWriter.java:286) >> at >> hep.io.sio.SIOWriter.flushRecord(SIOWriter.java:208) >> at >> hep.io.sio.SIOWriter.createRecord(SIOWriter.java:83) >> at >> org.lcsim.lcio.LCIOWriter.write(LCIOWriter.java:251) >> at >> org.lcsim.util.loop.LCIODriver.process(LCIODriver.java:114) >> ... 12 more >> >> >> Thanks. No rush on these, >> I imagine that even if the problems were fixed >> before/during the >> collaboration meeting we would not have time to use the >> files. >> >> >> >> >> >> >> ######################################################################## >> Use REPLY-ALL to reply to list >> >> To unsubscribe from the HPS-SOFTWARE list, >> click the following link: >> >> https://listserv.slac.stanford.edu/cgi-bin/wa?SUBED1=HPS-SOFTWARE&A=1 >> >> >> >> >> >> >> ######################################################################## >> Use REPLY-ALL to reply to list >> >> To unsubscribe from the HPS-SOFTWARE list, click the >> following link: >> >> https://listserv.slac.stanford.edu/cgi-bin/wa?SUBED1=HPS-SOFTWARE&A=1 >> >> >> >> >> >> ######################################################################## >> Use REPLY-ALL to reply to list >> >> To unsubscribe from the HPS-SOFTWARE list, click the following >> link: >> >> https://listserv.slac.stanford.edu/cgi-bin/wa?SUBED1=HPS-SOFTWARE&A=1 >> >> >> -- >> >> >> >> >> >> ######################################################################## >> Use REPLY-ALL to reply to list >> >> To unsubscribe from the HPS-SOFTWARE list, click the following link: >> https://listserv.slac.stanford.edu/cgi-bin/wa?SUBED1=HPS-SOFTWARE&A=1 >> >> ######################################################################## >> Use REPLY-ALL to reply to list >> >> To unsubscribe from the HPS-SOFTWARE list, click the following link: >> https://listserv.slac.stanford.edu/cgi-bin/wa?SUBED1=HPS-SOFTWARE&A=1 >> >> > ######################################################################## > Use REPLY-ALL to reply to list > > To unsubscribe from the HPS-SOFTWARE list, click the following link: > https://listserv.slac.stanford.edu/cgi-bin/wa?SUBED1=HPS-SOFTWARE&A=1 > > > -- > BEGIN-ANTISPAM-VOTING-LINKS > ------------------------------------------------------ > > NOTE: This message was trained as non-spam. If this is wrong, > please correct the training as soon as possible. > > Teach CanIt if this mail (ID 03PDvvqnA) is spam: > Spam: > https://www.spamtrap.odu.edu/canit/b.php?i=03PDvvqnA&m=e95fc18dd1f4&t=20151109&c=s > Not spam: > https://www.spamtrap.odu.edu/canit/b.php?i=03PDvvqnA&m=e95fc18dd1f4&t=20151109&c=n > Forget vote: > https://www.spamtrap.odu.edu/canit/b.php?i=03PDvvqnA&m=e95fc18dd1f4&t=20151109&c=f > ------------------------------------------------------ > END-ANTISPAM-VOTING-LINKS > > ######################################################################## Use REPLY-ALL to reply to list To unsubscribe from the HPS-SOFTWARE list, click the following link: https://listserv.slac.stanford.edu/cgi-bin/wa?SUBED1=HPS-SOFTWARE&A=1