Thursday, 2019-09-19

mnaserpabelanger: neat!00:02
mnaseri dont think 8 might be as bad00:03
pabelangerI think DIB is doing RHEL8, but haven't looked. ianw would know00:04
ianwyes, but only from upstream images00:04
pabelangeryah, figured will be some time for centos-minimal00:04
pabelangerfor v800:04
ianwi think we know the pain points from rhel8 & fedora28, so it should hopefully not be too bad00:05
pabelangersounds like a good time to try fedora-30 images00:10
openstackgerritJames E. Blair proposed zuul/zuul master: WIP: Fix gerrit errors from production  https://review.opendev.org/68300600:10
*** jamesmcarthur has quit IRC00:14
*** mattw4 has quit IRC00:21
openstackgerritIan Wienand proposed zuul/zuul master: Support nodes setting 'auto' python-path  https://review.opendev.org/68227500:29
*** Goneri has quit IRC00:45
*** jamesmcarthur has joined #zuul00:58
*** jamesmcarthur has quit IRC01:10
*** jamesmcarthur has joined #zuul01:10
*** jamesmcarthur has quit IRC01:15
*** jamesmcarthur has joined #zuul01:49
*** jamesmcarthur has quit IRC01:58
*** rlandy|bbl is now known as rlandy02:01
*** rlandy has quit IRC02:01
*** jamesmcarthur has joined #zuul02:28
*** roman_g has quit IRC02:35
*** bhavikdbavishi has joined #zuul02:51
*** bhavikdbavishi1 has joined #zuul02:54
*** bhavikdbavishi has quit IRC02:56
*** bhavikdbavishi1 is now known as bhavikdbavishi02:56
*** jamesmcarthur has quit IRC03:38
*** jamesmcarthur has joined #zuul03:40
*** jamesmcarthur has quit IRC03:44
*** jamesmcarthur has joined #zuul04:09
*** jamesmcarthur has quit IRC04:16
*** pcaruana has joined #zuul04:42
*** bolg has joined #zuul04:43
*** kerby has quit IRC04:45
*** jamesmcarthur has joined #zuul04:51
*** jamesmcarthur has quit IRC04:56
*** swest has joined #zuul05:00
*** bolg has quit IRC05:00
*** jamesmcarthur has joined #zuul05:17
*** jamesmcarthur has quit IRC05:22
*** bhavikdbavishi has quit IRC05:26
*** bhavikdbavishi has joined #zuul05:27
*** noorul has joined #zuul06:31
noorulI am trying to sync project src folders to /tmp/ using http://paste.openstack.org/show/777607/06:32
noorulI gave delegate_to: "{{ inventory_hostname }}"06:33
noorulI thought sync directory will be mapped to the node rather than on the executor06:34
*** themroc has joined #zuul06:53
noorulAny idea on the remote node where will be the projects src copied to ?06:58
AJaegerfor project openstack/horizon in opendev.org server, they are copied to "{{ ansible_user_dir }}/{{ zuul.projects['opendev.org/openstack/horizon'].src_dir }}"07:16
AJaegerThat's a variable you can use in ansible07:17
AJaegerAh, you use zuul_projects already, so that should be fine. You can look at the inventory log file to see what is where07:18
*** jamesmcarthur has joined #zuul07:18
*** jamesmcarthur has quit IRC07:23
*** tosky has joined #zuul07:23
*** jpena|off is now known as jpena07:33
mnasiadkaseems Zuul job queue is ridiculously long - and very long running tripleo jobs at the top...07:41
*** roman_g has joined #zuul07:42
*** noorul has quit IRC07:53
*** zbr is now known as zbr|ruck07:57
*** noorul has joined #zuul08:05
*** hashar has joined #zuul08:09
*** arxcruz|ruck is now known as arxcruz|rover08:20
*** lennyb has quit IRC08:21
mordredmnasiadka: the joys of release crunch time I believe08:22
mnasiadkamordred: might be, but still over 5 hours jobs don't look too good :)08:25
mordredmnasiadka: agree, it's never awesome. just be glad openstack is less trendy than a few years ago where release crunch would lead to 48 hour backed up queues :)08:28
mnasiadkamordred: well, non-trendiness has it's perks it seems :)08:28
mordredmy expectations may be skewed - I'm now thrilled that release crunch is ONLY doing 5 hour gate backups08:28
mordredright?08:28
mordredI LOVE being less trendy :)08:29
*** gtema has joined #zuul08:31
*** lennyb has joined #zuul08:34
*** noorul has quit IRC08:38
*** noorul has joined #zuul08:38
noorulmordred: hi09:05
noorulmordred: Y'day we discussed about required_projects and how zuul pulls in code from those09:05
noorulmordred: If the current project has different release scheme number scheme, how can we handle that09:06
*** jangutter has joined #zuul09:09
*** tobias-urdin has joined #zuul09:12
*** gtema has quit IRC09:13
mordrednoorul: that's the situtation where you'll need to use override checkout - because you'll need to tell zuul which branches should be tested with each other09:26
noorulmordred: Yes, but the dependency will vary release to release09:35
mordrednoorul: yes - you'll likely want to make some child jobs that each have a branch matcher and then an override checkout09:37
mordredlet me make you a for-instance real quick09:37
mordrednoorul: http://paste.openstack.org/show/777615/ <-- if you do that, it creates variants of the job definition that match on various different branch values of the first project09:40
mordrednoorul: so it says, "if the patch to the project is on release_1.0.0, please check out release_2.0 of some/other/project - and if the patch is on release_1.2.0 then please check out release_3.0 of some/other/project - but for other ones, check out matching branches"09:42
openstackgerritIan Wienand proposed zuul/zuul master: Add a more conversational overview to README.rst  https://review.opendev.org/68308509:45
AJaegernoorul: there's also https://zuul-ci.org/docs/zuul/user/config.html#attr-pragma.implied-branches09:47
mordredoh yeah. that too. thanks AJaeger :)09:53
*** noorul has quit IRC09:54
*** hashar has quit IRC09:59
*** hashar has joined #zuul10:00
*** bhavikdbavishi has quit IRC10:03
openstackgerritFabien Boucher proposed zuul/zuul master: Add reference pipelines file for Github driver  https://review.opendev.org/67271210:04
*** hashar has quit IRC10:21
*** noorul has joined #zuul10:42
*** jangutter has quit IRC10:44
*** noorul has quit IRC10:47
*** pcaruana has quit IRC10:54
*** hashar has joined #zuul10:55
*** jangutter has joined #zuul11:00
*** noorul has joined #zuul11:03
*** jpena is now known as jpena|lunch11:04
*** jangutter has quit IRC11:05
noorulmordred: Cool :)11:05
*** pcaruana has joined #zuul11:17
*** bhavikdbavishi has joined #zuul11:41
*** bhavikdbavishi1 has joined #zuul11:43
*** bhavikdbavishi has quit IRC11:45
*** bhavikdbavishi1 is now known as bhavikdbavishi11:45
noorulIf I have a job running for a PR and I push a change, shouldn't the old job get killed?11:48
*** jangutter has joined #zuul11:53
*** rfolco has joined #zuul12:04
*** Goneri has joined #zuul12:07
*** noorul has quit IRC12:08
*** jamesmcarthur has joined #zuul12:18
*** bhavikdbavishi has quit IRC12:18
*** bhavikdbavishi has joined #zuul12:19
*** bhavikdbavishi has quit IRC12:21
*** noorul has joined #zuul12:23
*** jpena|lunch is now known as jpena12:23
*** bhavikdbavishi has joined #zuul12:23
*** jamesmcarthur has quit IRC12:30
*** jamesmcarthur has joined #zuul12:30
*** rlandy has joined #zuul12:31
mnasernoorul: yes12:35
*** jamesmcarthur has quit IRC12:36
tobias-urdinhow could i best build a job to test an upgrade of a multi-repo project, what i need is project1-version1 install and test, project2 run upgrade, project1-version2 run test12:37
tobias-urdinwhere project1-version2 needs to support depends-on checkout12:37
*** armstrongs has joined #zuul12:38
pabelangercan have job that runs from project1-version2, but first switch branch, install, switch back12:39
pabelangerdepends-on will be correct12:40
armstrongsi noticed a bug in the new console output. If you have an ansible playbook fail with undefined variable error, zuul will fail but on the console log it as a pass in the ui. I checked the ara report and that logs it correctly as a fail. Is that something you guys are aware of or have come across. im on Zuul version: 3.10.2.dev23 b53b6bad12:40
noorulmnaser: That is not happening in my case. Is this functionality tightly coupled with source driver?12:40
tobias-urdinpabelanger: oh, not sure why i was thinking about checking out the project twice instead of swapping branches, thanks12:42
pabelangernoorul: yes, we sometimes see that happen in github driver (haven't fixed it yet), the key logic is in https://opendev.org/zuul/zuul/src/branch/master/zuul/driver/github/githubmodel.py#L41 for example12:44
pabelangersame would be with bitbucket12:44
pabelangertobias-urdin: yah, that is the method grenade in openstack does IIRC12:45
tobias-urdinpabelanger: i'll sneak peak grenade, that might help me getting started12:46
*** fdegir has quit IRC12:47
mordredtobias-urdin: zuul prepares appropriate state including depends-on for all of the branches in the repo to support that use case - so even the branches zuul isn't checking out for you should be set up properly if you change branch to them12:47
*** fdegir has joined #zuul12:48
mordred(which is just me agreeing with pabelanger with too many words)12:48
tobias-urdinthanks :)12:49
*** bhavikdbavishi1 has joined #zuul12:50
*** bhavikdbavishi has quit IRC12:51
*** bhavikdbavishi1 is now known as bhavikdbavishi12:51
mnaseri can probably look this up but has the zuul base library moved on from past 2.6 ?13:00
mnaseri ask because https://opendev.org/zuul/zuul-base-jobs/commit/7deaf1142f8755063f58591aa5e5367e7621207513:01
mnaser(uploading base-jobs into a repo on github with their scanning on makes it unhappy because of CVE-2019-10156)13:01
pabelangerno, we still have zuul that run ansible 2.513:07
pabelangerso we need to keep zuul-jobs to support it13:07
pabelangersorry, 2.6*13:07
pabelangermnaser: https://review.opendev.org/650431/ drop 2.5 in zuul13:09
pabelangerand stack adds 2.9 support13:09
pabelangerwhich I need to rebase13:09
*** pcaruana has quit IRC13:28
*** recheck has quit IRC13:28
*** recheck has joined #zuul13:29
*** jamesmcarthur has joined #zuul13:36
*** jamesmcarthur has quit IRC13:36
*** jamesmcarthur has joined #zuul13:36
noorulIs ansible 2.9 released?13:44
pabelangerno13:51
pabelangersoon(tm)13:51
pabelangerrc1 is likely today13:51
*** armstrongs has quit IRC13:54
*** pcaruana has joined #zuul13:54
*** saneax has joined #zuul14:01
*** avass has joined #zuul14:02
*** saneax has quit IRC14:04
*** saneax has joined #zuul14:05
pabelangernow that we have 3.10.2, we got some UI fixes.  I noticed this today14:09
pabelangeransible-test-network-integration-vyos-python37 (3. attempt)14:09
pabelangershould that say (3rd attempt)?14:09
*** saneax has quit IRC14:12
mordredmaybe?14:15
*** noorul has quit IRC14:15
pabelangerk, wasn't sure if . was a magic bit that auto did it14:16
mordredI also don't know14:17
*** hashar has quit IRC14:20
*** themroc has quit IRC14:24
fungitobias-urdin: i believe grenade does (or did) maintain separate clones of repositories under /opt/stack/old and /opt/stack/new, but as long as they're cloned from the copies zuul has prepared on the node they can still checkout those branches and get the same prepared heads14:26
fungimnaser: fwiw, i completely ignore/disable all of github's vulnerability notification features. the signal-to-noise ratio is so low as to be basically useless for all the projects i've gotten notified about in the past14:28
*** avass has quit IRC14:32
*** pcaruana has quit IRC14:33
*** michael-beaver has joined #zuul14:35
*** mmedvede has quit IRC14:36
*** mmedvede has joined #zuul14:37
flaper87what var should I use to get the src dir on the worker node?14:37
flaper87`zuul.project.src_dir` ?14:38
mordredflaper87: yes. that will be the src dir of the project that triggered the job14:38
mordredflaper87: (obviously it won't have as important a meaning for multi-repo jobs that can be triggered by different projects)14:38
corvusflaper87: that will be a relative path -- if you need an absolute path on the worker, you can prepend ansible_user_dir.14:39
mordredflaper87: it shoudl be noted that the zuul.project.src_dir value is a relative path14:39
mordredcorvus: jinx14:39
flaper87sweet, perfect, thanks! :)14:42
* flaper87 will use ansible_user_dir + zuul....14:42
AJaegerflaper87: example: "{{ ansible_user_dir }}/{{ zuul.projects['opendev.org/openstack/horizon'].src_dir }}"14:56
*** tosky has quit IRC14:57
*** tosky has joined #zuul14:58
openstackgerritDavid Shrewsbury proposed zuul/nodepool master: WIP: Do not overwrite image upload zk data on delete  https://review.opendev.org/68185714:59
openstackgerritKerby proposed zuul/nodepool master: AWS driver: add ability to determine AMI id using filters  https://review.opendev.org/68318315:25
*** jangutter has quit IRC15:29
*** Goneri has quit IRC15:47
openstackgerritDavid Shrewsbury proposed zuul/nodepool master: Do not overwrite image upload ZK data on delete  https://review.opendev.org/68185715:53
*** noorul has joined #zuul15:55
ShrewsI think that image leak fix ^^^ is finally ready. The test was... annoying. I was hoping to verify image deletion at the provider level, but that's not possible to do right now since we don't share provider data (in this case, images) across threads. PS2 shows the failure before the fix, PS3 should pass with the fix.15:57
*** mgoddard has quit IRC15:58
*** mgoddard has joined #zuul15:59
Shrewsoh, the test comment is a bit off16:00
noorulI have nodepool config http://paste.openstack.org/show/777709/16:01
mnasernoorul: neat! that seems functional as long as the right users are there?16:02
noorulI would like to run a job on "10.29.12.20", but without making any changest to nodepool16:02
noorul*changes16:02
mnaseri think you'll have to make a specific node label for that use case to be possible16:02
noorulI thought there will be another way16:03
pabelangeryes, need specific label16:03
openstackgerritDavid Shrewsbury proposed zuul/nodepool master: Do not overwrite image upload ZK data on delete  https://review.opendev.org/68185716:03
pabelangerthen create a nodeset in zuul, to use it16:03
pabelangerwait16:04
pabelangernoorul: just create a new nodeset, that used 10.29.12.20 as the label16:05
*** openstackgerrit has quit IRC16:06
noorulpabelanger: Let me try that16:07
noorulIs there a way to cancel the running job?16:07
clarkbyou can use the zuul dequeue command, or push a new patchset or abandon the change (though I don't know if these last two things work with the bitbucket driver)16:08
*** jamesmcarthur has quit IRC16:12
sean-k-mooneyyou might also be able to configure jobs to look at labels like workflow.16:14
sean-k-mooney*pipelines16:14
sean-k-mooneyif it a sepereate zuul install.16:15
sean-k-mooneybut i know i have used workfow to ignore WIP patches locally in the past not sure if it would kick out a job that is in the pipeline queue after it was added but before it started16:16
*** jamesmcarthur has joined #zuul16:21
*** noorul has quit IRC16:22
*** noorul has joined #zuul16:24
mnaserso GitHub allows you to provide a link with pre-populated fields for creating a new app16:27
mnaserwould it be beneficial to include that in our docs?16:27
mnaserso "go create a new app and check all these boxes" or "click this link and fill out the few missing fields"16:27
fungiseems like a reasonable improvement to me, as long as that api is fairly stabilized16:28
*** Goneri has joined #zuul16:29
nooruldequeue brought my zuul instance down16:30
clarkbnoorul hrm I hust used it yesterday with no problems on a freshly installed 3.10.2 zuul16:31
funginoorul: likely there will be a traceback in the log which will tell you why it crashed16:33
*** mattw4 has joined #zuul16:34
rlandyhello - zuul timer question: https://zuul-ci.org/docs/zuul/admin/drivers/timer.html says 'The first weekday is Monday.' iiuc, cron considers Sunday to be day 0. As such, if I want a pipeline to trigger on Saturday and Sunday, would the line be - time: '0 6 * * 5,6'?16:34
noorulIt came up after several restarts16:35
funginoorul: it's more that i'm hoping you can find the reason it crashed so we can determine if it's a bug with, say, interactions between the dequeue rpc subcommand and the bitbucket driver16:36
*** kerby has joined #zuul16:37
noorulfungi: I understand that16:38
pabelangerrlandy: I believe you want 0 6 * * 6,016:38
pabelangerwhat you have is Friday, saturday16:39
noorulfungi: Right now I am trying to bring up zuul to a state where I can demo its capabilities16:39
*** mattw4 has quit IRC16:39
*** mattw4 has joined #zuul16:39
rlandypabelanger: ok - 'The first weekday is Monday.' is confusing - if that is considered day 0 or day 116:40
noorulIs there a plan to merge stash driver soon. I have not seen ofosos for long16:40
pabelangerrlandy: I find contrab.guru helpful too!16:40
pabelangerhttps://crontab.guru/#0_6_*_*_6,016:40
rlandythanks16:40
*** Goneri has quit IRC16:41
*** jamesmcarthur has quit IRC16:42
*** jamesmcarthur has joined #zuul16:43
*** igordc has joined #zuul16:44
noorulpabelanger: This http://paste.openstack.org/show/777719/ is the error I see at that time16:44
pabelangeris port 19885 open on servers for zuul_console?16:45
pabelangertcp16:45
*** hashar has joined #zuul16:46
noorulYes16:46
noorulI could see console log streaming16:46
noorulIs there a way to make the Duration column in the builds UI more readable. In seconds/minutes ?16:47
pabelangeryah, I'd also like to see that updated too.16:48
noorulWhat is the difference between builds and buildsets?16:49
fungia buildset is all the builds associated with a particular change item16:51
fungiso for example when you propose a pr and zuul runs builds for each of the jobs, that set of builds it ran together constitute a buildset16:51
noorulThat means only relevant to gerrit16:51
fungiwhy would it only be relevant to gerrit?16:51
fungican you not run multiple jobs when using other code review systems?16:52
noorulok, multiple jobs16:52
fungiyes, "change item" was shorthand for gerrit change, github pr, git commit, timer event, et cetera16:52
fungiit's the set of builds which were triggered together16:53
noorulI got it http://zuul.opendev.org/t/zuul/buildset/42f6cc28945e4f32a2d985c904f2eaa416:54
noorulIs there a concept of build numbers like Jenkins in Zuul?16:59
noorulIs there a migration path?16:59
*** openstackgerrit has joined #zuul16:59
openstackgerritKerby proposed zuul/nodepool master: AWS driver: add ability to determine AMI id using filters  https://review.opendev.org/68320516:59
clarkbnoorul: each build and buildset is identified by a uuiid (the last portion of that link you just pasted is one), but there is no monotonically increasing counter for them17:01
*** pcaruana has joined #zuul17:01
clarkbnoorul: migration path of the build data? or of the jobs or?17:01
*** Goneri has joined #zuul17:01
noorulToday in Jenkins we use running number17:03
noorulx.y.z-build-number17:03
noorulpabelanger: http://paste.openstack.org/show/777723/ I tried that17:07
noorulpabelanger: but it still picking other node from the nodepool17:07
*** zbr|ruck is now known as zbr17:07
pabelangerlabel needs to be '10.29.12.20'17:08
pabelangerthat is what you have in nodepool17:08
noorulhttp://paste.openstack.org/show/777709/17:09
noorulYou mean name translates to label in nodeset?17:09
pabelangeryes17:10
pabelangerhttps://zuul-ci.org/docs/zuul/user/config.html#nodeset17:10
noorulnodeset.nodes.label (required)17:12
noorulThe Nodepool label for the node. Zuul will request a node with this label.17:12
noorulIn my nodepool config label is central-ut17:12
pabelangerOh, wait.17:12
pabelangersorry, this is static17:12
pabelangeryah, so you need per region labels17:13
pabelangerlike central-ut-a17:13
pabelangercentral-ut-b17:13
pabelangeretc17:13
noorulIn nodeset?17:13
pabelangerno, in nodepool17:13
noorulSo I have to change nodepool17:13
pabelangeryes17:13
*** Goneri has quit IRC17:14
*** pcaruana has quit IRC17:14
noorulok, then for the time being I will comment out all other nodes17:14
pabelangeryou can keep central-ut, just also create per region labels, if you want that17:14
noorulWhere can I find the change-id of the running job?17:20
noorulso that I can pass it as parameter to dequeue17:21
noorulzuul show has no ref / change id information17:22
*** jpena is now known as jpena|off17:22
*** pcaruana has joined #zuul17:25
noorulIn openstack python jobs, are you caching python packages somewhere? As the new nodes are created every time package install could take significant time17:33
pabelangeropendev uses regional caches to help speed up somethings17:34
*** hashar has quit IRC17:34
pabelangerbut, you can also cache things into our base images17:34
noorulhmm17:35
funginoorul: yes, in opendev's case we have multiple layers of caching. we run a caching apache proxy in each environment which our job nodes are configured to connect to instead of pypi.org, but also for packages which lack viable prebuilt wheels we maintain a separate wheel cache where we build platform-specific wheels for each of the operating systems we test on and then configure nodes to look there as well as17:36
fungithe pypi proxy17:36
fungibasically every region where we run jobs has a local to it cache of these things17:37
fungito cut down on trying to fetch them over the internet repeatedly17:38
fungiwe do similarly for distro packages, dockerhub images, and so on17:38
noorulhow can I determine change id of a running job?17:49
corvusnoorul: https://zuul-ci.org/docs/zuul/user/jobs.html#var-zuul.change17:50
ShrewstristanC: the nodepool openshift job seems to be failing quite often. see https://review.opendev.org/683205 and https://review.opendev.org/681857 for examples.17:51
corvusnoorul: oh you wanted that for dequeue -- look on the status page17:51
noorulcorvus: Is it this command separated value "4,01a7e7e135e04405e1493786fb6887351516b12f"17:53
corvusnoorul: yep17:53
nooruls/command/comma17:53
corvusnoorul: the 4 is the PR #, the sha is the current tip commit for that PR17:54
noorulcorvus: Thank you!17:54
noorulcorvus: It worked17:54
noorulI tried this zuul dequeue --tenant central --pipeline check --project ac/monitoring --ref 01a7e7e135e04405e1493786fb6887351516b12f17:55
nooruland zuul crashed17:55
fungithat definitely shouldn't be able to crash the scheduler, so sounds like a definite bug17:55
*** pcaruana has quit IRC17:55
SpamapSShrews: is there a quick way to make a request for nodepool without running a whole zuul scheduler?17:56
fungimay be we need better error handling around refs mentioned in rpc commands17:56
ShrewsSpamapS: not without a small bit of coding17:57
ShrewsSpamapS: the nodepool unit test suite is filled with examples of creating a manual NodeRequest17:57
AJaegerzuul-jobs reviewers, the 2 weeks waiting period for https://review.opendev.org/567696 is over - shall we merge it?17:57
corvusnoorul: can you elaborate on "zuul crashed"?17:57
*** avass has joined #zuul17:58
funginoorul: the tracebacks in http://paste.openstack.org/show/777719/ don't look related to the crash. i suppose it's possible the scheduler dies before it can write the relevant traceback to its log17:58
noorulcorvus: The UI was not responding17:58
funginoorul: did the scheduler process stop?17:58
ShrewsSpamapS: https://opendev.org/zuul/nodepool/src/branch/master/nodepool/tests/unit/test_launcher.py#L49-L52 is the gist of it17:58
noorulLooks like it stopped, because UI was returning 50017:59
fungidid you have to start the scheduler again to get it working?17:59
noorulYes17:59
*** Goneri has joined #zuul18:12
tristanCShrews: it seems like both failed happen in fortnebula-regionone, where the instance ip  192.168.48 is not reachable ?18:14
openstackgerritJames E. Blair proposed zuul/zuul master: Fix gerrit errors from production  https://review.opendev.org/68300618:19
*** bhavikdbavishi has quit IRC18:19
corvusclarkb, mordred: ^ that's the omnibus fix-gerrit-errors-from-yesterday patch18:20
*** igordc has quit IRC18:20
*** jamesmcarthur has quit IRC18:21
corvusShrews: did all the expected autohold patches land?18:21
corvusShrews: apparently not, looks like maybe lost in restart; i'll nudge them18:21
*** jamesmcarthur has joined #zuul18:22
Shrewscorvus: oh, i assume they had18:23
corvusmy fault, sorry18:23
*** noorul has quit IRC18:23
corvusShrews: they're ahead of the fix i just pushed up, so we should get them in the next restart18:24
Shrewscorvus: the last one in that chain needs a +3 https://review.opendev.org/67905718:24
Shrewscorvus: no worries. i forgot about them when i returned to the image leak fix   :)18:25
*** michael-beaver has quit IRC18:25
corvusShrews, tristanC, mhu: that change lgtm, but you can specify the request method in the route map, so you don't need an extra method to dispatch based on that -- maybe we should tidy that up in a later change18:26
*** avass has quit IRC18:50
*** Diabelko has quit IRC19:09
*** Diabelko has joined #zuul19:15
daniel2How would you modify the cloud config stuff for a nodepool build?19:32
Shrewscorvus: wow, such failure on the autohold revamp changes. and i can't seem to pull up logs19:33
pabelangerdaniel2: you mean on the one nodepool-builder uses?19:33
daniel2Yeah19:34
pabelangeryou can use any cfgmgmt tool19:34
pabelangereg: ansible19:34
corvusShrews: yeah, i dunno what happened with the unit tests; i was just looking at the quickstart19:34
Shrewslooks like qs failed on swift upload?19:35
corvusShrews: where do you see that?19:35
Shrewslast entry in https://zuul.opendev.org/t/zuul/build/927433d151a642049c85322a387d535d/log/job-output.txt19:35
Shrewsno error message, but that was the last task19:36
corvusShrews: no that's just where the logs end (we don't get any logs in swift after we start the swift upload, because of restrictions imposed by the space-time continuum) -- if it had failed there, it would be a post_failure19:36
corvusShrews: try the console tab, it'll take you right to the failed task: https://zuul.opendev.org/t/zuul/build/927433d151a642049c85322a387d535d/console19:37
Shrewscorvus: ah. we need ubiquitous logging. make that happen  :)19:37
corvusShrews: i've been reading this book "A Journey Through Time" which has some interesting ideas about how we could solve that19:38
Shrewscorvus: i find that error confusing, given the message is "OK"19:38
Shrewscorvus: are you able to access the py3* logs?19:39
corvusShrews: no argument there, but that tells us that something went wrong waiting for zuul to report, and we can dig into the zuul logs for that19:40
corvusShrews: yeah, but i can tell you they won't be useful.19:41
Shrewsi think the node was hit by a cosmic ray19:42
corvusif the job timed out with no test-results.html output, then it's probably some catastrophe that causes all the tests not to work19:42
corvusShrews: both, apparently19:42
corvuslike something happened to zk or somesuch19:42
corvusShrews: this looks like the actual error with qs: https://zuul.opendev.org/t/zuul/build/927433d151a642049c85322a387d535d/log/container_logs/scheduler.log#203719:44
corvusit seems very unlikely related to that change; i'm trying to come up with a theory19:46
corvusthe same version of gerrit was used in the last successful gate run19:49
*** pcaruana has joined #zuul19:51
Shrewsnothing unusual in the gerrit log19:52
corvusi'm running that job locally to see if it's reproducible19:56
*** jamesmcarthur has quit IRC20:01
openstackgerritJames E. Blair proposed zuul/zuul-jobs master: Remove parent from validate-zone-db  https://review.opendev.org/68322520:02
corvusnot a fatal error, but that does appear in the errer logs ^20:02
*** jamesmcarthur has joined #zuul20:04
corvusShrews: i got nothin20:15
corvusShrews: official cause of death: very bad luck20:15
Shrewsseems about right20:15
corvusShrews: i'm reapproving them all20:16
corvusShrews: my change failed with the same 40420:18
*** jamesmcarthur has quit IRC20:20
Shrewscorvus: it's catching20:21
*** jamesmcarthur has joined #zuul20:21
*** jamesmcarthur has quit IRC20:22
corvusShrews: all three jobs (the last successful gate job, yours, and mine) have the same sha256 for the gerrit container image20:24
Shrewssome sort of race? can't imagine how20:25
Shrewsor on what20:25
corvusmaybe on secondary index updates?20:26
*** pcaruana has quit IRC20:31
*** igordc has joined #zuul20:33
Shrewshrm, reading https://gerrit-review.googlesource.com/Documentation/config-gerrit.html#index seems there may be potential for some sort of race, but i'm not familiar enough with how/when it is used and if that could potentially affect these ops20:35
corvusit should have had plenty of time to update (it ran an actual job, plus we tried 4 times)20:35
corvusi've also verified that the zuul-scheduler image i tested with locally matches what was used in the most recent passing gate job20:39
*** sshnaidm is now known as sshnaidm|off20:42
corvusShrews: dockerhub has the image used for your change, maybe i'll try locally with that https://hub.docker.com/layers/zuul/zuul-scheduler/change_661114_latest/images/sha256-d0b18f587bc0235faff43a050dbf5c37671359efeb1ac2235ce1d39b3892a45420:42
openstackgerritMerged zuul/zuul-jobs master: Remove parent from validate-zone-db  https://review.opendev.org/68322520:47
corvusShrews: whoah-hoh!  i reproduced locally20:48
Shrewsmiddle-school-me wants to make a joke there20:50
Shrewsbut yay!20:50
Shrewsi wonder if it's an artifact of the way the image is built, or just chance20:51
corvusShrews: well, 2 image builds since the last one show the problem; and i don't thitk we have an image built since the last change merged which doesn't show it.20:52
corvusi'll see about doing an image build of master right now to see if it shows up20:53
SpamapSShrews: on the topic of how to create a request.. it's weird to me that we don't have CLI tools to CRUD requests.20:55
corvusSpamapS: we have RD not CU20:58
ShrewsSpamapS: RUD is supported by current nodepool cli. only the C20:58
Shrewsyeah, i guess not U since hold request was moved to zuul20:59
corvusthe only kind of request you could create from a cli would be a hold20:59
corvus(like, request a node and immediately hold it)20:59
ShrewsSpamapS: if you think of the way nodepool actually works, requesting nodes asynchronously via CLI doesn't make sense because after the request, you'd have to have another command to remove the fulfilled NodeRequest. And yet another to mark the nodes as USED.21:01
corvusso, potentially a useful debugging and development tool, but i guess not one that has seemed more useful than autoholding from zuul.  and when i've hacked on the drivers before, i just use min-ready to generate the requests.21:01
Shrewsit's not quite made for such manual (prone to error) interactions21:01
Shrewscorvus: looks like it failed again21:05
Shrewswow, this revamp is the pain that just keeps on giving21:06
Shrewsoh look... is that beer over there???21:06
corvusShrews: yeah, at this point i expect there's a real problem we don't understand yet21:07
corvusi just build new images based on current master, trying that now21:07
corvusyep that fails too21:08
corvusit's super weird that this failed but not the unit tests21:09
corvus(i mean, they did fail, but for other reasons -- not reliably)21:10
SpamapSShrews: k.. just makes it really hard to test in isolation other than via the test suite. :-P21:13
ShrewsSpamapS: yeah, but that's why we use zuul so we can do crossproject testing and not have to test in isolation   :)21:14
Shrewsbut yeah, hard to experiment with it in isolation w/o some coding21:15
corvusSpamapS: you may want to try the min-ready trick -- make a launcher config with 'min-ready: 1' and the launcher will always try to keep a node available.  delete the node to get it to run again.21:18
corvus^C and restart as you make changes21:19
SpamapScorvus: that's true.. I forgot about that. ;)21:20
SpamapSkerby: ^^21:20
Shrewsit's also not hard to write a 20 or so line python script that uses the nodepool zk library to request a node. i think i have one somewhere...21:20
corvusShrews: okay, i've got a curl command working that should be doing exactly what zuul says it's doing but getting a 404 -- and it's getting that 404 even after my curl command succeeds21:23
ShrewsSpamapS: http://paste.openstack.org/show/777952/21:24
corvusShrews: i'm assuming, for the moment, that some low-level system or python library has changed in some very subtle way21:24
SpamapSShrews:maybe that should be in tools. ;)21:25
ShrewsSpamapS: i would not want to give folks the impression that is a great way to dynamically request nodes21:25
Shrewsit does not account for leaving the node request around after it is fulfilled21:26
Shrewsit's a sloppy, quick hack21:26
SpamapSyeah, min-ready trick is better21:26
Shrewscorvus: that's wonderful news21:27
Shrewsthat script is actually quite old. it should be using storeNodeRequest() instead of the direct client call21:29
ShrewsSpamapS: ^^21:29
Shrewsbut a hack is a hack21:29
corvusShrews: i'm positive it's not zuul related now -- i have a short python script that reproduces the problem; it works in my py36 venv and fails in the image21:37
corvusnow i have 2 containers, one where the simple script works and one where it fails21:44
corvussame python version, same requests version21:44
corvus-urllib3==1.25.421:45
corvus+urllib3==1.25.321:45
corvusthat sounds plausible21:45
corvusi'd just like to call out urllib3's tagline: "Sanity-friendly HTTP client."21:46
corvusShrews: is that resonating for you right now like it is for me? :)21:46
Shrewscorvus: that’s the only difference?21:47
corvusShrews: in python libs, yes21:48
Shrewsnot doing much for MY sanity, personally21:48
corvusalso the only difference in system package21:49
corvuspackages21:49
corvusi still don't know what it's doing wrong.  i'm assuming it's doing something weird with "~"21:50
SpamapShttps://github.com/urllib3/urllib3/releases/tag/1.25.421:51
SpamapS" sethmlarson released this 5 hours ago "21:51
SpamapShttps://github.com/urllib3/urllib3/compare/1.25.3...1.25.421:53
SpamapSLots of changes21:53
corvushttps://github.com/urllib3/urllib3/pull/1673 looks promising21:53
ShrewsFix for Python 4 (#1669)21:55
Shrewsi wasn't aware that was a thing21:55
SpamapSSince you have a reproducer... bisect ftw?21:56
corvusyeah, let me see if i can get that happening on my workstation21:57
corvusyep, that works in my py36 venv21:57
corvusbisect confirms it's https://github.com/urllib3/urllib3/pull/1673/commits/fd8a95af871a4328c474086b897accef620661eb22:00
*** mattw4 has quit IRC22:01
*** mattw4 has joined #zuul22:02
corvusthey seem to be going against the "should not" in https://tools.ietf.org/html/rfc3986#section-2.322:05
corvusbecause it does look like they are percent-encoding the tilde22:06
*** armstrongs has joined #zuul22:27
corvushttps://github.com/urllib3/urllib3/pull/168422:30
*** rlandy is now known as rlandy|biab22:30
openstackgerritJames E. Blair proposed zuul/zuul master: Don't use urllib3 1.25.4  https://review.opendev.org/68325022:34
corvuszuul-maint: ^ can we merge that asap to unblock the gate?22:34
openstackgerritJames E. Blair proposed zuul/zuul master: Fix gerrit errors from production  https://review.opendev.org/68300622:35
*** armstrongs has quit IRC22:37
*** igordc has quit IRC22:44
*** igordc has joined #zuul23:01
*** mattw4 has quit IRC23:03
*** mattw4 has joined #zuul23:03
*** tosky has quit IRC23:12
*** rlandy|biab is now known as rlandy|bbl23:18
*** kerby has quit IRC23:59

Generated by irclog2html.py 2.15.3 by Marius Gedminas - find it at mg.pov.lt!