Friday, 2019-04-26

pabelangerlooking00:01
corvusno failures so far, and looking at the times, individual test runtimes look to be about 25% less than the current typical (while the entire job is about 20% longer)00:01
corvusi'd trade an extra 5 minutes for more reliability00:01
corvuswe're losing way more than 5m each time it bombs out00:02
pabelangeryah00:02
pabelanger++00:02
corvuswe've got 12 test runs under the belt so far, i just did a recheck to get 10 more00:02
pabelangerI kinda like that devstack has a dstat log, that you can also look at to see what the node is doing. Maybe we should consider a zuul jobs to do the same thing00:04
*** mattw4 has quit IRC00:07
*** jamesmcarthur has quit IRC00:08
*** jamesmcarthur has joined #zuul00:20
*** jamesmcarthur has quit IRC00:36
SpamapSHrm.. I rotated my github private key yesterday and now I'm getting 401's...00:48
corvuspabelanger: feel free to play around with https://review.opendev.org/61010000:49
pabelangercorvus: will do00:50
SpamapSah.. we forgot to bounce the merger00:50
pabelangermnaser: w00t! zuul streaming logs via ipv6 now00:51
pabelangernext up, ipv6 ansible00:51
pabelangerbut that is for another day00:51
pabelangerneed to prep for avengers!00:51
fungiokay, just finished confirming Revert "Prepend path with bin dir of ansible virtualenv" fixed openstack's release jobs01:15
fungiso that definitely seems to have been the cause for the regression we experienced01:16
mnaserpabelanger: nice!01:17
mnaserpabelanger: I hope to one day run OpenStack with iptables blocking all ipv4 traffic :)01:17
mnaserbut we've got some work still to do01:18
pabelangermnaser: seems like a great goal01:18
*** jamesmcarthur has joined #zuul01:27
pabelangerhttps://review.opendev.org/655808/ looks happy now, that is patch from corvus to fix race condition in testing01:29
pabelangerI +2'd01:29
*** jamesmcarthur has quit IRC01:34
*** jamesmcarthur has joined #zuul01:35
openstackgerritMerged zuul/zuul master: web: add triggers information to pipeline list  https://review.opendev.org/63767002:09
*** jamesmcarthur has quit IRC02:18
*** jamesmcarthur has joined #zuul02:26
*** bhavikdbavishi has joined #zuul02:28
*** jamesmcarthur has quit IRC02:43
*** jamesmcarthur has joined #zuul02:45
*** bhavikdbavishi has quit IRC02:59
*** jamesmcarthur has quit IRC03:15
*** jamesmcarthur has joined #zuul03:45
*** bhavikdbavishi has joined #zuul03:50
openstackgerritTristan Cacqueray proposed zuul/zuul master: trigger: add job filter event  https://review.opendev.org/63990503:51
*** jamesmcarthur has quit IRC03:51
openstackgerritTristan Cacqueray proposed zuul/zuul master: webtrigger: add initial driver and event  https://review.opendev.org/55515303:59
openstackgerritTristan Cacqueray proposed zuul/zuul master: webtrigger: add web route and rpclistener  https://review.opendev.org/55483903:59
openstackgerritTristan Cacqueray proposed zuul/zuul master: web: add build button to trigger job  https://review.opendev.org/63571603:59
openstackgerritTristan Cacqueray proposed zuul/zuul master: webtrigger: add support for parameterized trigger  https://review.opendev.org/64448403:59
openstackgerritTristan Cacqueray proposed zuul/zuul master: web: add build modal with a parameter form  https://review.opendev.org/64448503:59
openstackgerritTristan Cacqueray proposed zuul/zuul master: web: add support for checkbox and list parameters  https://review.opendev.org/64866103:59
*** bhavikdbavishi1 has joined #zuul04:01
*** bhavikdbavishi has quit IRC04:03
*** bhavikdbavishi1 is now known as bhavikdbavishi04:03
*** jamesmcarthur has joined #zuul04:20
*** pcaruana has joined #zuul04:23
*** jamesmcarthur has quit IRC04:27
*** threestrands has joined #zuul04:38
*** bhavikdbavishi has quit IRC04:52
*** bhavikdbavishi has joined #zuul04:53
*** jamesmcarthur has joined #zuul05:00
*** jamesmcarthur has quit IRC05:05
*** jamesmcarthur has joined #zuul05:34
*** jamesmcarthur has quit IRC05:39
*** quiquell has joined #zuul06:06
*** electrofelix has joined #zuul06:16
*** jamesmcarthur has joined #zuul06:35
*** jamesmcarthur has quit IRC06:40
*** pcaruana has quit IRC06:48
*** pcaruana has joined #zuul06:55
*** jamesmcarthur has joined #zuul07:13
*** jamesmcarthur has quit IRC07:25
*** jpena|off is now known as jpena07:43
*** threestrands has quit IRC07:50
openstackgerritMerged zuul/zuul master: Fix race in test_job_pause_pre_skipped_child  https://review.opendev.org/65580809:03
*** panda|off is now known as panda09:34
*** threestrands_ has joined #zuul10:03
*** threestrands_ has quit IRC10:16
*** bhavikdbavishi has quit IRC10:36
*** panda is now known as panda|lunch11:10
*** jpena is now known as jpena|lunch11:33
*** jamesmcarthur has joined #zuul11:45
*** bhavikdbavishi has joined #zuul11:49
*** jamesmcarthur has quit IRC11:50
*** quiquell is now known as quiquell|lunch11:51
*** panda|lunch is now known as panda11:57
*** maxamillion has quit IRC12:02
*** maxamillion has joined #zuul12:03
openstackgerritTobias Henkel proposed zuul/zuul master: Recover cached repos from corrupt object files  https://review.opendev.org/65589012:11
*** jamesmcarthur has joined #zuul12:17
*** jamesmcarthur has quit IRC12:30
*** EmilienM is now known as EvilienM12:31
*** jpena|lunch is now known as jpena12:32
*** gtema has joined #zuul12:36
*** quiquell|lunch is now known as quiquell12:40
openstackgerritMerged zuul/zuul-jobs master: Don't repeat the etc/alias setup for buildset registry pushes  https://review.opendev.org/65580212:47
*** jamesmcarthur has joined #zuul12:50
*** jamesmcarthur has quit IRC13:16
*** gtema has quit IRC13:17
*** gtema has joined #zuul13:18
*** openstackgerrit has quit IRC13:27
Shrewspabelanger: endgame was sooo good13:59
pabelangerShrews: yes! No spoilers but was great14:01
pabelangeralso, http://paste.openstack.org/show/749814/14:01
pabelangerhow can we get nodepool to give more info on that failure?14:01
Shrewspabelanger: you can't. that's all you're given back from openstacksdk (and iirc, that's all it gets back from nova)14:02
pabelangerboo14:02
pabelangerokay, asking mnaser to help see what happened14:03
*** openstackgerrit has joined #zuul14:46
openstackgerritSean McGinnis proposed zuul/zuul-jobs master: Revert "Add environment debugging to ensure-twine role"  https://review.opendev.org/65591614:46
*** gtema has quit IRC14:49
*** ericbarrett has quit IRC14:55
*** ianychoi_ has joined #zuul15:00
*** ianychoi has quit IRC15:03
*** zbr|rover is now known as zbr|over15:16
openstackgerritSean McGinnis proposed zuul/zuul-jobs master: Revert "Add environment debugging to ensure-twine role"  https://review.opendev.org/65591615:35
openstackgerritMerged zuul/zuul master: Recover cached repos from corrupt object files  https://review.opendev.org/65589015:43
openstackgerritPaul Belanger proposed zuul/zuul master: WIP: Run dstat and generate graphs in unit tests  https://review.opendev.org/61010015:52
*** electrofelix has quit IRC16:00
corvusi've restarted our scheduler with only a revert of 3704095c7927568a1f32317337c3646a9d15769e to see if it is the cause of the memory leak, or the other change (02b07a362b201382f62bb5dd0bb82e3bce35e4cc)16:05
corvus(so if our memory usage is steady, then the problem was the cancel jobs patch; if it grows now, then it was the missing project patch)16:06
corvushttp://cacti.openstack.org/cacti/graph.php?action=view&local_graph_id=64792&rra_id=all16:06
clarkbcorvus: 9fc2 wasthe other change16:07
corvuser yes, that16:09
corvus9f7c642ae1dc5ac7de1cb0ff5c7e32d6426bd1b316:10
*** chandankumar is now known as raukadah16:12
openstackgerritMerged zuul/zuul-jobs master: Revert "Add environment debugging to ensure-twine role"  https://review.opendev.org/65591616:20
*** mattw4 has joined #zuul16:20
*** jpena is now known as jpena|off16:42
*** panda is now known as panda|off16:56
*** maxamillion has quit IRC16:58
*** maxamillion has joined #zuul16:58
*** zxiiro has quit IRC17:03
*** zxiiro has joined #zuul17:04
mordredShrews: looks like zuul-preview has gone south again - I wonder if we're running our latest fixes there17:24
mordredcorvus: ^^17:25
mordredthis is obviously not urgent, and I don't think we should spend much effort on it today17:25
mordredif any effort at all17:25
corvusoh probably not, i've been completely neglecting it17:25
corvusprobably not running latest fixes17:25
mordredbut: http://site.4f816af6b10540b1b99d19fca3adc551.opendev.zuul-preview.opendev.org/ is hanging, which is the symptom that caused the fix patch17:25
mordredcool17:25
mordredshould I just do a pull/restart?17:25
corvusyeah, seems easy and no reason not to17:26
mordreddone17:26
mordredworks now17:26
*** jamesmcarthur has joined #zuul17:30
*** tjgresha has joined #zuul17:31
Shrewsmordred: that's exciting17:33
Shrewsglad it's working now tho17:34
*** tjgresha has quit IRC17:36
openstackgerritJames E. Blair proposed zuul/zuul master: Fix memory leak in job cancelation  https://review.opendev.org/65598217:49
corvustobiash, clarkb, mordred, Shrews: i left some commentary on https://review.opendev.org/640609 on changes i observed and the possible culprit.  655982 ^ should be a fix if i'm right.17:49
corvusi have not experimentally verified any of that; that's just from the old mental python interpreter.17:50
clarkbcool I'll take a look soon. Happu I could help narrow things down17:50
openstackgerritJames E. Blair proposed zuul/zuul master: Fix memory leak in job cancelation  https://review.opendev.org/65598217:52
openstackgerritJames E. Blair proposed zuul/zuul master: Fix memory leak in job cancellation  https://review.opendev.org/65598217:53
tobiashcorvus: oh, I fixed the first one in the fail-fast patch too: https://review.opendev.org/#/c/652764/7/zuul/scheduler.py17:53
corvustobiash: whoops.  :)17:54
Shrewscorvus: that change had me doing a double take there17:54
Shrewsthe new type hints threw me17:55
corvusShrews: yeah, sorry, the fix is just a few characters, the type hints are penance for allowing the error.  :)17:55
tobiashyeah, the type hints would have prevented that bug...17:56
Shrewslgtm though17:56
corvusi'm conflicted on whether i like them for python, but i have to admit that they can help in this case.  alternatively, we could be more consistent on taking job *objects* and then job.name will naturally crash if you give it the wrong thing.17:57
tobiashcorvus: is https://review.opendev.org/#/c/655982/3/zuul/manager/__init__.py a drive by fix or is there a relation I don't see?17:58
*** mattw4 has quit IRC17:58
tobiashah, looks like it was an unused parameter17:59
corvusyeah, i updated the commit to mention that in the latest ps18:00
corvuscommit msg18:00
mnaseryum, type-hint'd python...18:01
*** mattw4 has joined #zuul18:02
mordredcorvus: yay for type hints being useful!18:12
*** gtema has joined #zuul18:16
*** gtema has quit IRC18:20
*** jamesmcarthur has quit IRC18:34
*** jamesmcarthur has joined #zuul18:34
*** jamesmcarthur has quit IRC18:36
*** mattw4 has quit IRC18:45
*** mattw4 has joined #zuul18:48
*** jamesmcarthur has joined #zuul18:49
*** jamesmcarthur has quit IRC18:56
*** jamesmcarthur has joined #zuul18:59
*** mattw4 has quit IRC19:03
openstackgerritTobias Henkel proposed zuul/zuul master: Fix stale job dir deletion on startup  https://review.opendev.org/65600319:55
openstackgerritJames E. Blair proposed zuul/zuul master: Fix race in test_job_pause_retry  https://review.opendev.org/65600419:56
tobiashcorvus: I thought I remembered that the executor was supposed to cleanup the job dirs on startup, but it never did in my deployment. This fixes it ^19:56
corvustobiash: https://review.opendev.org/620697 is where i left that20:04
corvustobiash: looks like the same fix, though i added a non-working test20:06
corvusit's the test that was holding up that patch20:06
tobiashoh, looks like I overlooked that20:06
corvusconsidering the way the test was failing, i think it might be worth looking into that20:06
corvus(i haven't looked into it to find out if it would cause a production issue, or only tests)20:07
openstackgerritTobias Henkel proposed zuul/zuul master: Fix deletion of stale build dirs on startup  https://review.opendev.org/62069720:08
openstackgerritJames E. Blair proposed zuul/zuul master: Add more debugging to tests  https://review.opendev.org/65600620:13
openstackgerritMerged zuul/zuul master: Fix memory leak in job cancellation  https://review.opendev.org/65598220:42
*** nickx-intel has joined #zuul20:45
nickx-intelhi zuulies. I'm working through the quickstart guide ..... I'm not seeing zuul approve my reviews though :(20:45
nickx-intelI replied to gerrit change @zuul "looks ok to me, a human - what do you think zuul?"20:49
nickx-intel+1 +1 +1 ... just waiting for zuul to approve20:49
nickx-intelI think it's probably because of noop jobs but the guide doesn't really point to that; I'm expecting it to work as is20:49
clarkbnickx-intel: you have to leave a workflow +1 vote for zuul to pick it up looking at the quickstart example pipeline. Did you do that?20:50
corvusi think https://zuul-ci.org/docs/zuul/_images/review-1003.png is the relevant screenshot20:51
nickx-intelI left workflow +1 vote ... 2 +1 in fact ... and I'm actually in the Test Zuul Pipelines subheader right now20:51
corvusso if you click code-review +2, and click workflow +1, then click send20:51
nickx-intelI have to +2 code-review?20:51
nickx-intellet me try that20:51
nickx-intelok, +2, submitted, merged20:52
mordred\o/20:52
*** tjgresha has joined #zuul20:53
corvusnickx-intel: ah, if you at "Test Zuul Pipelines", the only thing we're expecting is for zuul to report on the change on patchset upload.  so i think you were working ahead of the rest of the class.  :)20:53
nickx-inteloh? I was expecting zuul to +1? huh20:54
nickx-intelverified +1 vote?20:54
nickx-intelerrata since quick-start was written or so? :120:55
corvusnickx-intel: yeah, roughly speaking, "Test Zuul Pipelines" should get you to the point where zuul should leave a +1 when you upload a patch.  the next section ("Configure a Base Job") gets all the way to merging it.20:55
clarkbunrelated but I'm about to shut down the desktop where I took these notes. The window where the memory leak arose was between 2019-04-16 03:02:39 UTC Restarted zuul-scheduler and zuul-web on commit 0bb220c and 2019-03-19 21:11:31 UTC restarted all of zuul at commit 77ffb70104959803a8ee70076845c185bd17ddc1 just in case we haev to do more debugging of that20:56
nickx-intel........ do I need to reload docker after the pipelines merge? for zuul to get its new config?20:56
corvusnickx-intel: nope that should all be automatic20:57
nickx-intelso I've done something wrong with the bootstrap corvus? I've been over and over this .... same results same results20:58
openstackgerritTobias Henkel proposed zuul/zuul master: Fix deletion of stale build dirs on startup  https://review.opendev.org/62069720:58
corvusnickx-intel: i'm not sure, i thought you said it worked?  what's not happening as you expect?20:58
nickx-intelthe merge of check and gate pipelines was successful vis a vis bootstrapping20:59
nickx-intelI'm expecting the basic noop check and gate configuration to evaluate proposed changes for untrusted projects21:00
nickx-intelI see quul is added as a reviewer but not verified +121:01
corvusso when you pushed up your test change, you did not see this?  https://zuul-ci.org/docs/zuul/_images/check1-1002.png21:01
nickx-intel^quul^zuul21:01
nickx-intelis that ...... when you upload via html?21:01
corvusnickx-intel: that should be what happens after you run "git review"21:01
nickx-intelI'm going to type out some text here, 1sec:21:02
nickx-intel# git review21:03
nickx-intel> could not connect to gerrit.21:03
nickx-intel> enter your gerrit username: [username]21:03
tobiashcorvus: latest PS on 620697 should fix the tests. As far as I can see there was no problem with the log streamer but just some places in the tests that didn't account for the changed job dir location21:03
nickx-intel> trying again with ssh://[username]@localhost:29418/test121:03
nickx-intel> Enter passphrase for key [...]:21:03
tobiashnickx-intel: zuul cannot handle ssh keys with pass phrases atm21:04
nickx-intel> creating a remote called "gerrit" that maps to: ssh://[username]@localhost:29418/test121:04
nickx-inteluhh21:04
nickx-intelhmm21:04
corvustobiash: that's not the issue21:04
nickx-intelso anyway, after all that21:04
nickx-intel> remote: SUCCESS21:04
corvustobiash: nickx-intel is running through the quickstart docs -- that's the git-review command being used to push up changes to the embedded gerrit21:04
nickx-intelremote: http://localhost:8080/c/test1/+/1002 Add Auul test1 job21:05
tobiashoops, missed that detail21:05
nickx-intelyeah so review succeeds as [user]21:05
corvusnickx-intel: can you go to the  http://localhost:8080/c/test1/+/1002 make a screenshot and paste the link here?21:05
corvus(firefox has built-in screenshot support if you're using it; should be under the "..." menu in the url bar)21:06
nickx-intelwhat's a good screenshot hosting site again? not pastebin .. uh ..21:07
corvusimgur works21:07
nickx-intelword corvus, sec21:07
nickx-intelhttps://imgur.com/a/ZY2shOY21:08
nickx-intel800x600 - apologies21:09
corvusnickx-intel: can you click "expand all" in the bottom right hand corner, scroll to the bottom, and screenshot that?21:10
nickx-intelyup yup21:10
nickx-intelhttps://imgur.com/a/2wK35z821:13
corvusnickx-intel: sorry i meant to expand the comments at the bottom21:14
nickx-inteloh haha ok 1sec21:14
nickx-intelhttps://imgur.com/a/GqLo96O21:15
nickx-intelsee me in the comments? "zuul pls"21:16
*** pcaruana has quit IRC21:16
nickx-intelcorvus, ^21:17
corvusnickx-intel: can you visit http://localhost:9000/t/example-tenant/status  and paste a screenshot?21:17
nickx-intelhmm port 9000 is closed21:19
nickx-intelI'm clearly not operating on localhost:####21:19
nickx-intelI see netstat listening on 900021:19
tjgreshawhich cloud are you on @intel21:20
corvusnickx-intel: how about the output of "docker ps" -- you can use paste.openstack.org to paste it here if you want21:20
nickx-inteliptables shows 9000 open21:20
tjgreshaneed to change the security in the tenant to open port 9000 if it is not21:20
nickx-inteltjgresha, if you have an internal address pm me21:21
nickx-intelooooo let me check that, although, I've added those rules, but let's see21:21
corvusnickx-intel: ah, you're sshing into a cloud vm where you're running docker?  and i guess you have port 8080 open so you can reach gerrit?21:22
nickx-intelyeahhhhh I'm just missing 9000 from the "cloud" security group rules, sec21:22
nickx-intel9000 open, ss incoming21:23
tjgreshasee - i know things21:23
SpamapSfeature idea: it would be cool if we could make checks cancelled if a job goes in to gate. I waste a *lot* of compute on parallel check/gate runs when we're paired up and reviewing fast.21:23
nickx-intelcorvus, check: 0 gate: 0; 0 events; 0 management events; 0 results21:24
nickx-inteltjgresha, +1  :D21:24
openstackgerritTobias Henkel proposed zuul/zuul master: Fix test race in test_job_pause_post_fail  https://review.opendev.org/65601921:25
tobiashcorvus: another test race fix ^21:25
corvusnickx-intel: is there a little bell in the top right hand corner?21:25
nickx-intelnope corvus21:25
corvusnickx-intel: at this point i think i would look at the scheduler log to see if there are errors there, and if it saw the event from gerrit.21:26
corvusnickx-intel: you should be able to see that with "docker logs examples_scheduler_1" i think21:27
nickx-intelI can see noop jobs in zuul-config master check/gate jobs ....21:27
corvusSpamapS: yeah we could probably use that in our new opendev and zuul tenants (which don't have clean check either)21:28
SpamapScorvus:right, I was thinking something like "cancels: {pipeline}" which would be able to cancel something in the other pipeline with the same change ID.21:28
corvusSpamapS: yeah i think that's the way to do it21:29
nickx-inteluh, binary file (standard input) matches .. even with xargs :1  just tryina grep error21:29
SpamapScorvus:I'll do it with one of my free minutes.21:30
corvusSpamapS: shouldn't take more than one.  ;)21:30
nickx-intellol21:30
openstackgerritTobias Henkel proposed zuul/zuul master: Increase zookeeper priority during tests  https://review.opendev.org/65602121:32
nickx-intelcorvus, handscanning this log, I see no errors, would be nice to grep but "people in hell want ice water" :921:32
corvusnickx-intel: oh you might have to "docker logs examples_scheduler_1 2>&1 | grep"21:33
corvusi think the logs may go to stdout and docker "helpfully" maintains that21:33
corvuser stderr21:33
nickx-intelsame binary file (standard input) matches21:34
nickx-intelI'm working on the concept that stuff is succeeding improperly21:34
corvusnickx-intel: the logs should have a lot of information including all of the events received from gerrit, starting and completing jobs, and reporting to gerrit, though obviously at least the last one will be missing.21:34
corvusnickx-intel: another thing you can check in the zuul web interface is the builds tab --  http://localhost:9000/t/example-tenant/builds21:35
corvusnickx-intel: if you see something there, then zuul got the event from gerrit and ran jobs; if not, then we've narrowed it down to either zuul not receiving the event, or it did receive the event but didn't match it to any projects or pipelines, and therefore ran no jobs.21:36
openstackgerritTobias Henkel proposed zuul/zuul master: Fix test race in test_job_pause_post_fail  https://review.opendev.org/65601921:36
corvustobiash: http://logs.openstack.org/06/656006/1/check/tox-py35/b15bf22/testr_results.html.gz looks like maybe we have a similar race in plain old "test_job_pause".  i know there's a bunch of unrelated errors there, but i think the failure on test_job_pause is a legitimate test race.21:37
nickx-intelcorvus,  https://pastebin.com/9V6ZU81k21:37
nickx-intelbuilds appears to be empty :o21:38
tobiashcorvus: yes, looks like exactly the same race, fixing21:38
corvusnickx-intel: then the answer is going to be in the scheduler log.  if you want to paste that, i can help analyze it.21:39
*** mattw4 has joined #zuul21:39
openstackgerritTobias Henkel proposed zuul/zuul master: Fix test race in test_job_pausee  https://review.opendev.org/65602421:40
nickx-intelthanks corvus, let me check with someone here about that, and I'll get back to you with a "we figured it out" or "ugh wat"21:40
openstackgerritTobias Henkel proposed zuul/zuul master: Fix test race in test_job_pause  https://review.opendev.org/65602421:41
corvusnickx-intel: good luck.  several other folks here can help pinpoint problems too, if i'm not around.  the log is very verbose, but if you haven't run a zuul you might not know what's *not* being printed.21:42
nickx-intelthanks for tracking corvus, we're reviewing internally at present22:12
nickx-intelcorvus, there were a couple of issues: 1. I made myself admin in gerrit (force of habit) 2. verified review as user and took zuul's job22:35
nickx-intelso zuul was sulking without throwing any errors XD22:36
nickx-intel<zuul> "ok fine u do it then"22:36
*** sshnaidm has joined #zuul22:37
*** sshnaidm is now known as sshnaidm|off22:43
*** sshnaidm|off has quit IRC22:46
*** EvilienM is now known as EmilienM22:51
*** sshnaidm has joined #zuul22:53
*** sshnaidm is now known as sshnaidm|off22:54
*** jamesmcarthur has quit IRC23:05
pabelangerhttps://review.opendev.org/656024/ could use another +2 / +A, thanks to tobiash / corvus. I was looking into them this afternoon, but got distracted23:13
*** sshnaidm|off has quit IRC23:24
pabelangerso, I've now see across multiple tests that we don't seem to be shutting down all the threads properly: http://logs.openstack.org/04/656004/1/gate/tox-py36/1e59b10/testr_results.html.gz23:38
pabelangerI can see the thread ID, eg: Thread: 140429056014080 then traceback23:39
pabelangerhowever, I am unsure how it maps to: http://paste.openstack.org/show/749842/23:40
pabelanger<Thread(Thread-2929, started daemon 140429056014080)>, doesn't really explain what it is23:40
pabelangerwhere others are:  <Thread(Gearman client poll, started daemon 140429592884992)>23:40
corvuspabelanger: that happens when tests timeout; it's a symptom not a cause23:51
pabelangercorvus: ah, okay. thanks for info23:51
corvuspabelanger: usually what i do is look for the first Traceback; if it's in the middle of a test and it relates to a disconnect, it means the system is too busy and we had a connection timeout23:53
corvuspabelanger: if it's a timeout exception, then that's straightforward -- it's just a test timeout, also probably because the test is too busy23:53
corvuspabelanger: if one of those happens, then the tests that run afterwords are suspect -- they're probably going to output errors like that even if they ran okay23:54
corvuspabelanger: if the first error is an actual test assertion error though, that's probably debuggable.23:54
pabelangerk23:56
pabelangerI just rechecked https://review.opendev.org/656019/ it failed in gate, but 656024 is the fix.23:57
pabelangerthen will look at output of 655805 again, which seems to show we are still loose gearman connections some times23:58

Generated by irclog2html.py 2.15.3 by Marius Gedminas - find it at mg.pov.lt!