The xrootd fix appears to have done the trick :) I'm running Jacek's load test on the cluster. On 12/03/15 03:30, Fabrice Jammes wrote: > Hello, > > John, you're branch is now running on the cluster with latest xrootd > version, and .... it works much better :-) > > fjammes@ccosvms0070:~/src/qserv-cluster/shmux (master=)$ time mysql > --host ccqserv125 --port 4040 --user qsmaster LSST -e "SELECT ra, decl > FROM Object WHERE deepSourceId = 2322920177142607;"; > +------------------+-------------------+ > | ra | decl | > +------------------+-------------------+ > | 29.3088063472755 | -86.3088404611897 | > +------------------+-------------------+ > > real 0m0.835s > user 0m0.008s > sys 0m0.023s > fjammes@ccosvms0070:~/src/qserv-cluster/shmux (master=)$ > fjammes@ccosvms0070:~/src/qserv-cluster/shmux (master=)$ > fjammes@ccosvms0070:~/src/qserv-cluster/shmux (master=)$ > fjammes@ccosvms0070:~/src/qserv-cluster/shmux (master=)$ time mysql > --host ccqserv125 --port 4040 --user qsmaster LSST -e "SELECT count(*) > FROM Object;"; > +----------------+ > | SUM(QS1_COUNT) | > +----------------+ > | 1889695615 | > +----------------+ > > real 0m29.622s > user 0m0.010s > sys 0m0.025s > > Feel free to test it extensively. > > Use qserv-cluster/shmux/run.sh to reinstall Qserv from scratch. > > Cheers > > On 12/02/2015 07:02 PM, John Gates wrote: >> >> The branch for me is tickets/DM-2699. >> >> I made the change to admin/templates/configuration/etc/lsp.cf and >> pushed it to github. >> >> Thank you :) >> >> On 12/02/15 09:30, Fabrice Jammes wrote: >>> Hi John, >>> >>> Andy and you did great job chasing the bug :-) >>> I'll try to re-install all of it tonight, during our meeting. Could >>> you please: >>> - recall me your branch name, >>> - add "ssi.trace all debug" in lsp.cf server section (please watch >>> ccqserv126 lsp.cf for example) in this branch if still needed >>> - push to github the latest version of your branch, it'll run on the >>> cluster. >>> >>> Thanks, >>> >>> Fabrice >>> >>> On 12/02/2015 05:25 PM, John Gates wrote: >>>> Hi Fabrice, >>>> >>>> Can you rebuild the containers with the latest xrootd code so we >>>> can test his change? Or tell me what I need to do to build the >>>> containers and distribute them. >>>> >>>> Thank you, >>>> John >>>> >>>> >>>> -------- Forwarded Message -------- >>>> Subject: Re: int2p3 cluster problem >>>> Date: Wed, 2 Dec 2015 01:45:34 -0800 >>>> From: Andrew Hanushevsky <[log in to unmask]> >>>> To: Gates, John H <[log in to unmask]> >>>> CC: Fabrice Jammes <[log in to unmask]>, Becla, Jacek >>>> <[log in to unmask]>, Fritz Mueller <[log in to unmask]> >>>> >>>> >>>> >>>> Hi John, >>>> >>>> I have pushed the fix to the main xrdssi branch. That eans that it has to >>>> be merged into the LSST close of the branch and xroot rebuilt. Hopefully, >>>> the build is not part of the container at this point. Also, I rolloed in >>>> all of the accumulated patches to the base xrootd/cmsd. That tag should >>>> correspond to 4.3.0-rc4 (we currently have RC3 but this includes new >>>> patches since then). >>>> >>>> Andy >>>> >>>> On Tue, 1 Dec 2015, Gates, John H wrote: >>>> >>>> > There's a little bit more information from the czar-consol.log file. For >>>> > chunk 4483, the last thing in the log is "Sending a fcntl command for >>>> > ...". It looks like it should be followed by a "Sending read command ... >>>> > ", which never happens. >>>> > >>>> > -John >>>> > >>>> > ---- czar-consol.log grep -- 4483 (failed) >>>> > 151130 18:26:38 281 SsiSched: TaskXeqEvent: [[2015-11-30 18:26:38.644839 >>>> > +0000][Debug ][File ] >>>> > [0xf0076270@xroot://ccqserv125.in2p3.fr:1094//chk/LSST/1034] Sending a >>>> > read command for handle 0x4 to 10.158.37.135:1094 >>>> > SessRelTask: [0x7fdef03bce00] [2015-11-30 18:26:40.448304 +0000][Debug >>>> > ][File ] >>>> > [0xf039db40@xroot://ccqserv125.in2p3.fr:1094//chk/LSST/3668] Sending a >>>> > read command for handle 0x38 to 10.158.37.126:1094 >>>> > SessProcReq: [0x7fdef03effd0] Task=0x7fde3c006230 processing >>>> > id=0[2015-11-30 18:26:40.544837 +0000][Debug ][File ] >>>> > [0xf03ef770@xroot://ccqserv125.in2p3.fr:1094//chk/LSST/3846] Sending a >>>> > fcntl command for handle 0x3d to 10.158.37.133:1094 >>>> > TaskKill: [0x7fdec0007300] Status = isReady mhPend=0 id=0[2015-11-30 >>>> > 18:26:40.704483 +0000][Debug ][File ] >>>> > [0xf0435710@xroot://ccqserv125.in2p3.fr:1094//chk/LSST/4006] Sending an >>>> > open command >>>> > SessOpen: [0x7fdef051ff20] Opening >>>> > xroot://ccqserv125.in2p3.fr:1094//chk/LSST/4483 >>>> > [2015-11-30 18:26:41.189518 +0000][Debug ][File ] >>>> > [0xf0520060@xroot://ccqserv125.in2p3.fr:1094//chk/LSST/4483] Sending an >>>> > open command >>>> > [151130 18:26:41 268 SsiSched: running TaskReal0x7fde24007df0] Status = >>>> > isReady mhPend=0 id=[2015-11-30 18:26:41.198308 +0000][Debug >>>> > ][File ] >>>> > [0xf0520060@xroot://ccqserv125.in2p3.fr:1094//chk/LSST/4483] Open has >>>> > returned with status [SUCCESS] >>>> > [2015-11-30 18:26:41.198334 +0000][Debug ][File ] >>>> > [0xf0520060@xroot://ccqserv125.in2p3.fr:1094//chk/LSST/4483] >>>> > successfully opened at 10.158.37.126:1094, handle: 0x51, session id: 1 >>>> > TaskXeqEvent: [0x7fde30007210] [2015-11-30 18:26:41.204531 +0000][Debug >>>> > ][File ] >>>> > [0xf0520060@xroot://ccqserv125.in2p3.fr:1094//chk/LSST/4483] Sending a >>>> > write command for handle 0x51 to 10.158.37.126:1094 >>>> > [2015-11-30 18:26:41.208073 +0000][Debug ][File ] >>>> > [0xf0520060@xroot://ccqserv125.in2p3.fr:1094//chk/LSST/4483] Sending a >>>> > fcntl command for handle 0x51 to 10.158.37.126:1094 >>>> > : running TaskRealTaskXeqEvent: inq= >>>> > >>>> > >>>> > ----- czar-consul.log grep 4435 (successful) >>>> > TaskSetBuff: [0x7fde90006d30[2015-11-30 18:26:41.044359 +0000][Debug >>>> > ][File ] >>>> > [0xf04c44f0@xroot://ccqserv125.in2p3.fr:1094//chk/LSST/4307] Sending a >>>> > truncate command for handle 0x4b to 10.158.37.129:1094 >>>> > SessOpen: [0x7fdef04fe520] Opening >>>> > xroot://ccqserv125.in2p3.fr:1094//chk/LSST/4435 >>>> > [2015-11-30 18:26:41.131018 +0000][Debug ][File ] >>>> > [0xf04fe660@xroot://ccqserv125.in2p3.fr:1094//chk/LSST/4435] Sending an >>>> > open command >>>> > TaskXeqEvent: [2015-11-30 18:26:41.154194 +0000][Debug >>>> > ][File ] >>>> > [0xf04fe660@xroot://ccqserv125.in2p3.fr:1094//chk/LSST/4435] Open has >>>> > returned with status [SUCCESS] >>>> > [0x7fde080081c0] sess=ok id=0 Status = [2015-11-30 18:26:41.154214 >>>> > +0000][Debug ][File ] >>>> > [0xf04fe660@xroot://ccqserv125.in2p3.fr:1094//chk/LSST/4435] >>>> > successfully opened at 10.158.37.126:1094, handle: 0x4f, session id: 1 >>>> > TaskXeqEvent: [0x7fded400d5a0[2015-11-30 18:26:41.156974 +0000][Debug >>>> > ][File ] >>>> > [0xf04fe660@xroot://ccqserv125.in2p3.fr:1094//chk/LSST/4435] Sending a >>>> > write command for handle 0x4f to 10.158.37.126:1094 >>>> > TaskXeqEvent: [2015-11-30 18:26:41.205640 +0000][Debug >>>> > ][File ] >>>> > [0xf04fe660@xroot://ccqserv125.in2p3.fr:1094//chk/LSST/4435] Sending a >>>> > fcntl command for handle 0x4f to 10.158.37.126:1094 >>>> > [2015-11-30 18:26:41.212676 +0000][Debug ][File ] >>>> > [0xf04fe660@xroot://ccqserv125.in2p3.fr:1094//chk/LSST/4435] Sending a >>>> > read command for handle 0x4f to 10.158.37.126:1094 >>>> > 151130 18:26:41 311 SsiSchedSessProcReq: [2015-11-30 18:26:41.255720 >>>> > +0000][Debug ][File ] >>>> > [0xf04fe660@xroot://ccqserv125.in2p3.fr:1094//chk/LSST/4435] Sending a >>>> > read command for handle 0x4f to 10.158.37.126:1094 >>>> > TaskXeqEvent: [0x7fde0c006af0] sess=ok id=0[2015-11-30 18:26:41.301306 >>>> > +0000][Debug ][File ] >>>> > [0xf04fe660@xroot://ccqserv125.in2p3.fr:1094//chk/LSST/4435] Sending a >>>> > truncate command for handle 0x4f to 10.158.37.126:1094 >>>> > >>>> > >>>> > >>>> > On 12/01/15 02:15, Andrew Hanushevsky wrote: >>>> >> Hi John, >>>> >> >>>> >> Well, assuming the IP address is for 126 the first question is why >>>> >> doesn't resolve via DNS. Not that it should matter as the cmsd doesn't >>>> >> really care that a node has a dns name. What I see i the log is that >>>> >> that IP address did log in. So, other than not having a DNS name >>>> >> nothing seems unusual. Are the missing chunks always related to 126? >>>> >> >>>> >> Andy >>>> >> >>>> >> On Mon, 30 Nov 2015, John Gates wrote: >>>> >> >>>> >>> Hi Andy, >>>> >>> >>>> >>> >>>> >>> I'm hoping you can shed some light to the cause of this. The problem >>>> >>> is that ProcessResponse is not being called for a few, seemingly >>>> >>> random, chunks. I've not completed the query at all on the cluster >>>> >>> and it only occurs on the cluster. >>>> >>> >>>> >>> I've tried the query select count(*) from Object; on the cluster an >>>> >>> it failed with 9 queries in flight. I've only looked at jobId=1945 in >>>> >>> depth, which corresponds to chunk 4483. I've included relevant parts >>>> >>> of the log files below. >>>> >>> >>>> >>> >>>> >>> The cmsd.log entry for ccqserv126 was a little odd in that most >>>> >>> workers seem to have 2 sets of entries in the log file and ccqserv126 >>>> >>> is missing the set where it is mentioned by name and only has the set >>>> >>> where it is mentioned by ip address. The qserv-czar.log file looked >>>> >>> fine except that process responses is never called for chunk4483. >>>> >>> >>>> >>> The worker log looks fine for chunk 4483 but is missing entries I see >>>> >>> for other chunks (flagged below with "*****"). I expect to see >>>> >>> something like the following for 4483: >>>> >>> ssi_fctl: 0:/chk/LSST/XXXX query resp status >>>> >>> ssi_fctl: 0:/chk/LSST/XXXX resp ready >>>> >>> ssi_Finalize: 0:/chk/LSST/XXXX [bound odRsp] Calling Finished(0) >>>> >>> >>>> >>> Thanks, >>>> >>> John >>>> >>> >>>> >>> Failed executive job ids. >>>> >>> < 1945 >>>> >>> < 2280 >>>> >>> < 4788 >>>> >>> < 5409 >>>> >>> < 5522 >>>> >>> < 5572 >>>> >>> < 6765 >>>> >>> < 7063 >>>> >>> < 7422 >>>> >>> >>>> >>> >>>> >>> ------- 1945 -> chunk 4483 --- found on ccqserv126 >>>> >>> czar log: >>>> >>> 2015-11-30T18:26:41.188Z [0x7fdeef7fe700] DEBUG root >>>> >>> (core/modules/qproc/TaskMsgFactory2.cc:153) - SELECT count(*) AS >>>> >>> QS1_COUNT FROM LSST.Object_4483 AS QST_1_ >>>> >>> 2015-11-30T18:26:41.189Z [0x7fdeef7fe700] INFO root >>>> >>> (core/modules/ccontrol/MergingHandler.cc:226) - setError: code: 0, >>>> >>> message: >>>> >>> 2015-11-30T18:26:41.189Z [0x7fdeef7fe700] DEBUG qdisp.Executive >>>> >>> (core/modules/qdisp/Executive.cc:106) - Executive::add(job(id=1945 >>>> >>> payload.len=123 ru=/chk/LSST/4483)) >>>> >>> 2015-11-30T18:26:41.189Z [0x7fdeef7fe700] DEBUG root >>>> >>> (core/modules/qdisp/JobQuery.h:104) - JobQuery JQ_jobId=1945 >>>> >>> desc=job(id=1945 payload.len=123 ru=/chk/LSST/4483) >>>> >>> 2015-11-30T18:26:41.189Z [0x7fdeef7fe700] INFO qdisp.Executive >>>> >>> (core/modules/qdisp/Executive.cc:129) - Executive: Add job with >>>> >>> path=/chk/LSST/4483 >>>> >>> 2015-11-30T18:26:41.189Z [0x7fdeef7fe700] DEBUG root >>>> >>> (core/modules/qdisp/MessageStore.cc:53) - Add msg: 4483 1200 >>>> >>> Executive: Add job with path=/chk/LSST/4483 >>>> >>> 2015-11-30T18:26:41.189Z [0x7fdeef7fe700] DEBUG root >>>> >>> (core/modules/qdisp/JobQuery.cc:55) - runJob {job(id=1945 >>>> >>> payload.len=123 ru=/chk/LSST/4483) : 2015-11-30T18:26:41+0000, >>>> >>> Unknown, 0, } >>>> >>> 2015-11-30T18:26:41.189Z [0x7fdeef7fe700] INFO root >>>> >>> (core/modules/ccontrol/MergingHandler.cc:226) - setError: code: 0, >>>> >>> message: >>>> >>> 2015-11-30T18:26:41.189Z [0x7fdeef7fe700] DEBUG root >>>> >>> (core/modules/qdisp/QueryResource.cc:51) - QueryResource JQ_jobId=1945 >>>> >>> 2015-11-30T18:26:41.189Z [0x7fdeef7fe700] INFO qproc.QuerySession >>>> >>> (core/modules/qproc/QuerySession.cc:383) - Non-subchunked >>>> >>> 2015-11-30T18:26:41.189Z [0x7fdeef7fe700] DEBUG root >>>> >>> (core/modules/qproc/TaskMsgFactory2.cc:151) - no nextFragment >>>> >>> 2015-11-30T18:26:41.198Z [0x7fde727fc700] DEBUG root >>>> >>> (core/modules/qdisp/QueryRequest.cc:57) - jobId=1945 New QueryRequest >>>> >>> with payload(123) >>>> >>> 2015-11-30T18:26:41.198Z [0x7fde727fc700] DEBUG root >>>> >>> (core/modules/qdisp/QueryRequest.cc:74) - jobId=1945 Requesting, >>>> >>> payload size: [123] >>>> >>> 2015-11-30T18:26:41.204Z [0x7fde727fc700] DEBUG root >>>> >>> (core/modules/qdisp/QueryResource.cc:55) - ~QueryResource() >>>> >>> JQ_jobId=1945 >>>> >>> 2015-11-30T18:26:41.208Z [0x7fde6bfff700] DEBUG root >>>> >>> (core/modules/qdisp/QueryRequest.cc:81) - jobId=1945 RelRequestBuffer >>>> >>> >>>> >>> ----- czar cmsd.log: - missing second server login message for >>>> >>> ccqserv126 >>>> >>> 151130 18:21:57 225 Protocol: Primary >>>> >>>server.199:[log in to unmask]:1094 logged in. >>>> >>> =====> Routing for 10.158.37.126: local pub4 prv4 >>>> >>> =====> Route all4: 10.158.37.126 Dest=[::10.158.37.126]:1094 >>>> >>> *** there should be something like "Protocol: Primary >>>> >>> server.199@ccqserv126" but there's nothing >>>> >>> >>>> >>> 151130 18:21:35 221 Protocol: Primary >>>> >>>server.198:[log in to unmask]:1094 logged in. >>>> >>> =====> Routing for 10.158.37.127: local pub4 prv4 >>>> >>> =====> Route all4: 10.158.37.127 Dest=[::10.158.37.127]:1094 >>>> >>> 151130 18:22:11 205 Protocol: Primary server.199:28@ccqserv127:1094 >>>> >>> logged in. >>>> >>> =====> Routing for 10.158.37.127: local pub4 prv4 >>>> >>> =====> Route all4: 10.158.37.127 Dest=[::10.158.37.127]:1094 >>>> >>> >>>> >>> >>>> >>> >>>> >>> ------ ccqserv126 cmsd.log looks a lot like the other cmsd.log files >>>> >>> [2015-11-30T18:26:41.198Z] [0x7f054c250700] INFO root >>>> >>> (core/modules/xrdsvc/SsiService.cc:105) - Got provision call where >>>> >>> rName is: /chk/LSST/4483 >>>> >>> 151130 18:26:41 237qserv.232:[log in to unmask] ssi_open: /chk/LSST/4483 >>>> >>> 151130 18:26:41 237qserv.232:[log in to unmask] ssi_write: >>>> >>> 0:/chk/LSST/4483 rsz=123 wsz=123 >>>> >>> 151130 18:26:41 237qserv.232:[log in to unmask] ssi_Activate: >>>> >>> 0:/chk/LSST/4483 [new wtReq] oucbuff rqsz=123 >>>> >>> 151130 18:26:41 257qserv.232:[log in to unmask] ssi_DoIt: >>>> >>> 0:/chk/LSST/4483 [begun xqReq] Calling session Process >>>> >>> [2015-11-30T18:26:41.205Z] [0x7f05447f8700] INFO root >>>> >>> (core/modules/xrdsvc/SsiSession.cc:61) - ProcessRequest, >>>> >>> service=/chk/LSST/4483 >>>> >>> 151130 18:26:41 257qserv.232:[log in to unmask] ssi_GetRequest: >>>> >>> 0:/chk/LSST/4483 [begun xqReq] sz=123 >>>> >>> [2015-11-30T18:26:41.205Z] [0x7f05447f8700] INFO root >>>> >>> (core/modules/xrdsvc/SsiSession.cc:68) - GetRequest took 6.6e-05 seconds >>>> >>> [2015-11-30T18:26:41.205Z] [0x7f05447f8700] INFO root >>>> >>> (core/modules/xrdsvc/SsiSession.cc:99) - Decoding TaskMsg of size 123 >>>> >>> [2015-11-30T18:26:41.205Z] [0x7f05447f8700] DEBUG root >>>> >>> (core/modules/wbase/Task.cc:111) - Task(...) tSeq=80 :count=81 0, 1, >>>> >>> 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20, >>>> >>> 21, 22, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, >>>> >>> 38, 39, 40, 41, 42, 43, 44, 45, 46, 47, 48, 49, 50, 51, 52, 53, 54, >>>> >>> 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 67, 68, 69, 70, 71, >>>> >>> 72, 73, 74, 75, 76, 77, 78, 79, 80 >>>> >>> 151130 18:26:41 257qserv.232:[log in to unmask] ssi_BindDone: >>>> >>> 0:/chk/LSST/4483 [begun xqReq] Bind called; session set >>>> >>> 151130 18:26:41 257qserv.232:[log in to unmask] ssi_RelReqBuff: >>>> >>> 0:/chk/LSST/4483 [bound xqReq] called >>>> >>> [2015-11-30T18:26:41.205Z] [0x7f05447f8700] DEBUG root >>>> >>> (core/modules/wsched/BlendScheduler.cc:75) - BlendScheduler::queCmd >>>> >>> tSeq=80 >>>> >>> [2015-11-30T18:26:41.205Z] [0x7f05447f8700] DEBUG BlendSched >>>> >>> (core/modules/wsched/BlendScheduler.cc:93) - Blend chose group >>>> >>> [2015-11-30T18:26:41.205Z] [0x7f05447f8700] DEBUG BlendSched >>>> >>> (core/modules/wsched/BlendScheduler.cc:100) - Blend queCmd tSeq=80 >>>> >>> [2015-11-30T18:26:41.205Z] [0x7f05447f8700] INFO root >>>> >>> (core/modules/xrdsvc/SsiSession.cc:137) - BindRequest took 0.000167 >>>> >>> seconds >>>> >>> [2015-11-30T18:26:41.205Z] [0x7f0544ff9700] DEBUG BlendSched >>>> >>> (core/modules/wsched/BlendScheduler.cc:156) - >>>> >>> BlendScheduler::_ready() groups(r=1, q=1, flight=0) scan(r=0, q=0, >>>> >>> flight=0) >>>> >>> [2015-11-30T18:26:41.206Z] [0x7f05447f8700] INFO root >>>> >>> (core/modules/xrdsvc/SsiSession.cc:138) - Enqueued TaskMsg for >>>> >>> Resource(/chk/LSST/4483) in 0.000167 seconds >>>> >>> ... >>>> >>> [2015-11-30T18:26:41.206Z] [0x7f0544ff9700] DEBUG BlendSched >>>> >>> (core/modules/wsched/BlendScheduler.cc:112) - >>>> >>> BlendScheduler::commandStart tSeq=80 >>>> >>> [2015-11-30T18:26:41.206Z] [0x7f0544ff9700] DEBUG Foreman >>>> >>> (core/modules/wdb/QueryRunner.cc:137) - Exec in flight for Db = >>>> >>> q_fd5b7faeb8710396bed8ae38be7ad9ef >>>> >>> [2015-11-30T18:26:41.206Z] [0x7f0544ff9700] WARN Foreman >>>> >>> (core/modules/wdb/QueryRunner.cc:115) - QueryRunner overriding dbName >>>> >>> with LSST >>>> >>> [2015-11-30T18:26:41.208Z] [0x7f0544ff9700] DEBUG root >>>> >>> (core/modules/wdb/QueryRunner.cc:242) - _transmit last=1 tSeq=80 >>>> >>> [2015-11-30T18:26:41.208Z] [0x7f0544ff9700] DEBUG root >>>> >>> (core/modules/wdb/QueryRunner.cc:263) - _transmitHeader >>>> >>> [2015-11-30T18:26:41.208Z] [0x7f0544ff9700] INFO root >>>> >>> (core/modules/proto/ProtoHeaderWrap.cc:52) - msgBuf size=256 -> >>>> >>> [[0]=40, [1]=13, [2]=2, [3]=0, [4]=0, ..., [251]=48, [252]=48, >>>> >>> [253]=48, [254]=48, [255]=48] >>>> >>> 151130 18:26:41 237qserv.232:[log in to unmask] ssi_fctl: >>>> >>> 0:/chk/LSST/4483 query resp status >>>> >>> 151130 18:26:41 237qserv.232:[log in to unmask] ssi_fctl: >>>> >>> 0:/chk/LSST/4483 resp not ready >>>> >>> [2015-11-30T18:26:41.208Z] [0x7f0544ff9700] INFO root >>>> >>> (core/modules/xrdsvc/SsiSession_ReplyChannel.cc:85) - sendStream, >>>> >>> checking stream 0 len=256 last=0 >>>> >>> 151130 18:26:41 237qserv.232:[log in to unmask] ssi_Done: >>>> >>> 0:/chk/LSST/4483 [bound xqReq] wtrsp sent; resp here >>>> >>> 151130 18:26:41 184qserv.232:[log in to unmask] ssi_ProcessResponse: >>>> >>> 0:/chk/LSST/4483 [bound xqReq] Response presented wtr=0 >>>> >>> 151130 18:26:41 184qserv.232:[log in to unmask] ssi_ProcessResponse: >>>> >>> 0:/chk/LSST/4483 [bound doRsp] Resp strm >>>> >>> ***** the above is the last mention of 4483 in the log file. >>>> >>> Should there be lines like the one below for chunk 4435 ??? >>>> >>> ***** 151130 18:26:41 237qserv.232:[log in to unmask] ssi_fctl: >>>> >>> 0:/chk/LSST/4435 query resp status >>>> >>> ***** 151130 18:26:41 237qserv.232:[log in to unmask] ssi_fctl: >>>> >>> 0:/chk/LSST/4435 resp ready >>>> >>> ***** 151130 18:26:41 158qserv.232:[log in to unmask] ssi_Finalize: >>>> >>> 0:/chk/LSST/4435 [bound odRsp] Calling Finished(0) >>>> >>> [2015-11-30T18:26:41.208Z] [0x7f0544ff9700] INFO root >>>> >>> (core/modules/xrdsvc/ChannelStream.cc:91) - last=0 [[0]=40, [1]=13, >>>> >>> [2]=2, [3]=0, [4]=0, [5]=0, [6]=21, [7]=47, [8]=0, [9]=0, ..., >>>> >>> [246]=48, [247]=48, [248]=48, [249]=48, [250]=48, [251]=48, [252]=48, >>>> >>> [253]=48, [254]=48, [255]=48] >>>> >>> [2015-11-30T18:26:41.208Z] [0x7f0544ff9700] INFO root >>>> >>> (core/modules/xrdsvc/ChannelStream.cc:94) - trying to append message >>>> >>> (flowing) >>>> >>> [2015-11-30T18:26:41.208Z] [0x7f0544ff9700] DEBUG root >>>> >>> (core/modules/wdb/QueryRunner.cc:253) - _transmit last=1 tSeq=80 >>>> >>> resultString=[[0]=8, [1]=0, [2]=16, [3]=1, [4]=26, ..., [42]=48, >>>> >>> [43]=50, [44]=55, [45]=16, [46]=0] >>>> >>> [2015-11-30T18:26:41.208Z] [0x7f0544ff9700] INFO root >>>> >>> (core/modules/xrdsvc/SsiSession_ReplyChannel.cc:85) - sendStream, >>>> >>> checking stream 0x7f0538007a20 len=47 last=1 >>>> >>> [2015-11-30T18:26:41.209Z] [0x7f0544ff9700] INFO root >>>> >>> (core/modules/xrdsvc/ChannelStream.cc:91) - last=1 [[0]=8, [1]=0, >>>> >>> [2]=16, [3]=1, [4]=26, [5]=29, [6]=10, [7]=27, [8]=10, [9]=9, ..., >>>> >>> [37]=10, [38]=6, [39]=50, [40]=54, [41]=56, [42]=48, [43]=50, >>>> >>> [44]=55, [45]=16, [46]=0] >>>> >>> [2015-11-30T18:26:41.209Z] [0x7f0544ff9700] INFO root >>>> >>> (core/modules/xrdsvc/ChannelStream.cc:94) - trying to append message >>>> >>> (flowing) >>>> >>> [2015-11-30T18:26:41.209Z] [0x7f0544ff9700] DEBUG BlendSched >>>> >>> (core/modules/wsched/BlendScheduler.cc:132) - >>>> >>> BlendScheduler::commandFinish tSeq=80 >>>> >>> >>>> >>> >>>> >>> >>>> >>> >>>> > >>>> > >>>> >>>> >>> >> >> >> ------------------------------------------------------------------------ >> >> Use REPLY-ALL to reply to list >> >> To unsubscribe from the QSERV-L list, click the following link: >> https://listserv.slac.stanford.edu/cgi-bin/wa?SUBED1=QSERV-L&A=1 >> > ######################################################################## Use REPLY-ALL to reply to list To unsubscribe from the QSERV-L list, click the following link: https://listserv.slac.stanford.edu/cgi-bin/wa?SUBED1=QSERV-L&A=1