Monday, 2018-01-29

*** elyezer has quit IRC00:11
*** elyezer has joined #zuul00:23
*** Wei_Liu has joined #zuul02:16
*** sshnaidm has quit IRC02:40
*** harlowja has joined #zuul03:11
*** xinliang has quit IRC03:16
*** xinliang has joined #zuul03:16
*** harlowja has quit IRC03:18
*** elyezer has quit IRC03:35
*** elyezer has joined #zuul05:48
*** xinliang has quit IRC06:02
*** xinliang has joined #zuul06:14
*** xinliang has quit IRC06:14
*** xinliang has joined #zuul06:14
*** yolanda_ has quit IRC06:44
*** elyezer has quit IRC07:10
*** elyezer has joined #zuul07:15
Wei_LiutristanC: I found out that the zuul scheduler is required to be restarted if I change main.yaml, is there anyway let zuul scheduler load main.yaml automatically?07:33
tristanCWei_Liu: it's not automatic, but you can signal SIGHUP to trigger a dynamic reconfiguration07:57
*** AJaeger has quit IRC08:26
*** AJaeger has joined #zuul08:31
*** sshnaidm has joined #zuul08:37
Wei_LiutristanC: would this action affect the running job?08:44
tristanCtristanC: it should not affect running jobs08:45
*** jpena|off is now known as jpena08:48
*** threestrands_ has joined #zuul10:33
*** threestrands has quit IRC10:36
*** jappleii__ has joined #zuul10:37
*** jappleii__ has quit IRC10:38
*** jappleii__ has joined #zuul10:39
*** threestrands_ has quit IRC10:41
rcarrillocruzodyssey4me: hey jesse, how's going10:42
rcarrillocruzturns out i started writing a nodepool ansible inventory10:42
rcarrillocruziirc you wanted to deploy a just nodepool CI (with maybe jenkins)10:42
rcarrillocruzcurious if that plugin would be of interest of you, as i'm writing it and happy to put $things to return per hosts10:43
rcarrillocruzthinking on returning now groups by label10:43
rcarrillocruzand state10:43
rcarrillocruz[ready[10:43
rcarrillocruz[used]10:43
rcarrillocruzetc10:43
odyssey4me@rcarrillocruz hmm, actually that'd be pretty nifty I think10:47
odyssey4meyeah, there needs to be some way of communicating between jenkins and nodepool essentially...10:48
odyssey4meI guess if there's a dynamic inventory, jenkins could just run a playbook once per minute - effectively being a poll10:49
odyssey4mebut then how would jenkins inform nodepool when it's using the node, and when it's done with a node?10:49
*** evrardjp has joined #zuul10:51
rcarrillocruzjenkins would need to request the lock on a node, then release it when done10:53
rcarrillocruzi.e. do what zuul does in zk.py10:53
rcarrillocruzi've always thought it would be useful to have a 'lock' verb on nodepool cli10:53
rcarrillocruzlike, when zuul takes a node10:53
rcarrillocruzit sets the state of the node to 'in-use'10:53
rcarrillocruzwould be nice if you could do that from CLI10:54
rcarrillocruzon non-zuul CIs10:54
odyssey4meyeah, that'd be a simple way of implementing it actually10:54
rcarrillocruzin the uber rough CI we have in place, we simply have one node per platform10:55
rcarrillocruzand we run a cronjob10:55
rcarrillocruzto run int tests for each one of them10:55
rcarrillocruzthat way, we don't need to schedule10:55
rcarrillocruzthere's always one ready10:55
rcarrillocruzonce we finish10:55
rcarrillocruzwe delete the node10:55
rcarrillocruzand nodepool recreates10:55
rcarrillocruzbut that's super janky10:55
rcarrillocruzshould be simple enough to do a kazoo script that requests a node, then release post job10:56
odyssey4meat this point we're kinda ok to go with janky, but obviously we'd like to work towards something better and more maintainable10:56
rcarrillocruzodyssey4me: if janky is ok, we use that in prod , it works10:56
odyssey4meyeah, that's our current plan10:56
rcarrillocruzexcept we use the openstack inventory10:56
rcarrillocruzi figure the nodepool inventory will open more interesting things10:56
rcarrillocruzanyway, i'll ping you when i have it up10:56
rcarrillocruzcurious to see what kind of info you think would be useful to pull10:57
odyssey4meFor the moment we're on hold with this - I'm hoping we can get back to it next month. Unfortunately other work was prioritised above it.10:57
odyssey4meI did a rather janky implementation which got me to a point... https://gist.github.com/odyssey4me/d1a202d6e340d165513f9cec1d19d5f010:58
odyssey4meit implemented a little CLI thing to request/release nodes and a daemon to watch changes in state10:59
odyssey4methe idea being that when the state changes, do something (like register the node in jenkins)10:59
odyssey4meShrews very rightly said that we should be using the same API as zuul, instead of directly changing the data in zookeeper... hence my notes about how hacky it is.11:01
*** elyezer has quit IRC11:04
* odyssey4me wonders if he needs to remind mordred to republish https://review.openstack.org/520657 to the master branch :)11:04
*** electrofelix has joined #zuul11:06
*** elyezer has joined #zuul11:06
*** JasonCL has joined #zuul11:14
*** JasonCL has quit IRC11:15
*** JasonCL has joined #zuul11:15
*** _ari_|DevConf is now known as _ari_|conf11:32
*** _ari_|conf is now known as _ari_|brno11:32
*** elyezer has quit IRC12:23
*** elyezer has joined #zuul12:25
*** sshnaidm has quit IRC12:28
*** jpena is now known as jpena|lunch12:33
*** sshnaidm has joined #zuul12:43
*** sshnaidm_ has joined #zuul13:11
*** sshnaidm has quit IRC13:14
openstackgerritTobias Henkel proposed openstack-infra/zuul master: Move github webhook from webapp to zuul-web  https://review.openstack.org/53571113:16
openstackgerritTobias Henkel proposed openstack-infra/zuul master: Move status_url from webapp to web section  https://review.openstack.org/53677313:16
openstackgerritTobias Henkel proposed openstack-infra/zuul master: Remove webapp  https://review.openstack.org/53678013:16
*** rlandy has joined #zuul13:30
*** jpena|lunch is now known as jpena13:33
rcarrillocruzodyssey4me, Shrews : the fact there's interest for non-zuul interactions with nodepool, yet the api for interacting for it is within zuul tree (zk.py), it begs the question if there should be at some point a very light nodepool-lib13:55
odyssey4meyeah, that'd be nice - obviously it'd make it much easier to implement the right bits in various code bases, even if just in nodepool and zuul to start with13:57
Shrewsrcarrillocruz: odyssey4me: that was actually my plan (the lib idea) when we started zuulv3. but it was decided that we'd just break it out later if we needed to.14:07
odyssey4meShrews it'd be super-nice if nodepool standalone had the CLI options to take nodes through their life cycle... that'd make integration with third party tooling far simpler14:08
Shrewsodyssey4me: can't do that because of the requirement to hold locks during the lifecycle14:09
*** dmsimard is now known as dmsimard|afk14:17
*** myoung is now known as myoung|reboot14:20
*** myoung|reboot is now known as myoung14:29
openstackgerritTobias Henkel proposed openstack-infra/nodepool master: Fix relaunch attempts when hitting quota errors  https://review.openstack.org/53693014:58
*** elyezer has quit IRC15:03
*** elyezer has joined #zuul15:05
*** myoung is now known as myoung|brb15:24
pabelangerI'm not sure if I am going to explain it properly, but I'm seeing a lot of dynamic reload configurations related to a single project this morning in zuul.  openstack/networking-odl15:30
pabelangerour results queue grows large, due to reloads being processed, eventually we finish that, results contain new info, and reloads happen again with openstack/networking-odl15:31
pabelangeris it possible that job results (pass / fail) could be triggering a reload for openstack/networking-odl ? I new when a new patchset was added into check, it would do so. But I am seeing a lot of Re-enqueing change ... in queue ..15:32
pabelanger2018-01-29 15:33:18,172 DEBUG zuul.Pipeline.openstack.gate: Re-enqueing change <Change 0x7fbdb66afd30 528717,52> in queue <ChangeQueue gate: openstack/networking-odl> for example15:33
pabelangerhttp://paste.openstack.org/show/657071/15:35
*** myoung|brb is now known as myoung15:37
Shrewsclarkb: i left a comment on https://review.openstack.org/538529. I think there's something else happening there.15:47
clarkbShrews: that is the behavior I saw with extra logging added to the test15:48
clarkbbasially there would be an extra node in a deleting state and we'd double delete it never deleting the second node so quota would not free up15:49
clarkbhappened about one in 10 test runs locally15:49
Shrewsclarkb: but... there's no node15:49
Shrewsso i'm confused15:49
Shrewsclarkb: oh! we create an artificial one for leaked instances15:52
Shrewsso yeah, we could be racing that one15:52
Shrewsclarkb: changed to +3  :)15:54
clarkbthanks, good to know nodepool isnt being extra weird here :)15:54
Shrewswe could have also just compared the node id's there, but state also works15:54
clarkbfwiw I added a print(nodes) just above the loop to figure it out15:54
Shrewsnodepool is just complex. even though I wrote most of it, i have to continually re-inspect code to remind myself of how it works  :/15:55
Shrewss/wrote/re-wrote/15:56
corvusi'm not feeling well today -- would anyone else like to chair the meeting?15:58
corvusor we could cancel it16:02
Shrewscorvus: i think i can stumble my way through the commands. but i think we can probably skip this one maybe?16:02
Shrewscorvus: we'll work it out. go rest16:03
clarkbI'm taking care of kids all this week due to some last minute travel plans by my wife. Happy to help run a meeting if the kids will let me sit still long enough16:03
corvusShrews: ok thanks16:03
*** jappleii__ has quit IRC16:05
ShrewsI don't see anything on https://wiki.openstack.org/wiki/Meetings/Zuul so maybe we can just skip it today16:05
pabelanger+116:10
openstackgerritMerged openstack-infra/nodepool master: Fix race in test_failed_provider  https://review.openstack.org/53852916:21
*** jpena is now known as jpena|brb16:44
*** JasonCL has quit IRC17:05
*** JasonCL has joined #zuul17:08
*** sshnaidm_ is now known as sshnaidm17:09
*** JasonCL has quit IRC17:10
*** elyezer has quit IRC17:23
*** elyezer has joined #zuul17:23
*** jpena|brb is now known as jpena17:31
corvustristanC: i have one small request (for another test) on 535553 -- then i think we can merge it (and strike an item off the 3.0 blocker list!)17:40
*** myoung is now known as myoung|food17:42
*** JasonCL has joined #zuul17:48
*** weshay|ruck is now known as weshay|ruck|brb18:09
*** myoung|food is now known as myoung18:14
openstackgerritFabien Boucher proposed openstack-infra/zuul-jobs master: Propose to move submit-log-processor-jobs and submit-logstash-jobs in zuul-jobs  https://review.openstack.org/53784718:17
*** jpena is now known as jpena|off18:18
*** weshay|ruck|brb is now known as weshay18:20
*** weshay is now known as weshay|ruck18:20
openstackgerritDavid Shrewsbury proposed openstack-infra/nodepool master: Partial revert for disabled provider change  https://review.openstack.org/53899518:28
ShrewsCan I get some eyes on 538995? I don't want a just merged change to accidentally hit production. It's also a good demonstration for "always include tests".  :)18:30
* Shrews reprimands himself18:30
clarkbShrews: looks like more than a partial revert as the max servers stuff is removed entirely?18:34
Shrewsclarkb: no, it keeps the "accept max-servers less than 0" part18:34
Shrewsclarkb: which was something you asked for a while back, iirc18:35
clarkbits only checking max_concurrency not max_servers18:35
Shrewscorrect18:36
clarkboh I see it is supposed to fall through a bit further18:37
Shrewsright18:37
Shrewsi mean, i can revert the entire thing and add the new test and -1 support separately if that's a cleaner solution18:38
Shrewsbut that's 2, maybe 3, reviews compared to this single one18:39
clarkbno I see it now just had to pull up the old change and do mental diffing18:39
pabelangertobiash: looking at 536930, the test for test_over_quota might also need to be updated. If I understand, request 2 would never be fulfilled now, so we need a 3rd?18:39
tobiashpabelanger: yeah it also needs updating18:41
tobiashbut have to look at this closer18:41
tobiashprobably tomorrow morning18:41
pabelangerack18:41
tobiashI think request 2 might be paused now as it is intended and we need to check for this, delete request 1 and check if request 2 now gets processed18:42
tobiashat least that's what I remember from a few months ago18:42
tobiashbut will take a closer look tomorrow18:42
clarkbI'm not sure how many of us here knew spearce but https://groups.google.com/forum/#!topic/repo-discuss/B4P7G1YirdM is worth a read if you knew him18:48
clarkbtl;dr cancer sucks18:48
Shrewsclarkb: oh wow18:54
*** sshnaidm is now known as sshnaidm|afk19:12
openstackgerritMatthieu Huin proposed openstack-infra/zuul master: [WIP] zuul web: add admin endpoint, enqueue commands  https://review.openstack.org/53900419:23
*** harlowja has joined #zuul19:25
Shrewsmordred: pabelanger: clarkb: SpamapS: Any thoughts on skipping today's Zuul meeting in 2hrs? Nothing on the agenda and corvus will be afk. Unless someone has something to discuss, I vote push to next week.20:00
SpamapS+1, I can't attend20:01
pabelangerthe only real thing I wanted to discuss was migrating zuul.o.o to zookeeper cluster, due to outage this weekend. But more a heads up of something we want to do sooner then later, and if there is any concerns20:01
pabelangerand how we might do that, but that is more an openstack-inra thing20:01
clarkbI don't have anything urgent to go over20:01
SpamapSpabelanger: oh, did you have to dump the zk db?20:01
SpamapSor super long snap recovery?20:02
Shrewspabelanger: yeah, we could cover that at tomorrow's infra meeting20:02
pabelangerSpamapS: we haven't done it yet. But we did loose zookeeper host for 4 hours yesterday20:02
pabelangerShrews: yup, wfm20:02
ShrewsSpamapS: iirc, we greatly reduced our snapshot count a while back20:03
Shrewsdue to diskspace issues20:03
Shrewsthough for the life of me, i cannot find where we set that  :/20:10
pabelangersystem-config?20:10
clarkbits a puppet module setting iirc20:10
Shrewscodesearch.o.o of snapCount doesn't pull up anything20:12
clarkbpurge_interval => 6,20:14
clarkbin system-config/manifests/site.pp20:14
clarkbthe puppet module defaulted to higher snapcount out of the box iirc20:17
openstackgerritMatthieu Huin proposed openstack-infra/zuul master: zuul autohold: allow filtering per commit  https://review.openstack.org/53699320:19
openstackgerritTobias Henkel proposed openstack-infra/zuul master: Remove webapp  https://review.openstack.org/53678020:21
clarkbthe length of the outage was due to rax saying leave the node alone20:21
clarkbthen they brought back the hardware for it in a broken way aiui20:21
pabelangeryes20:26
SpamapSbummer20:44
*** elyezer has quit IRC21:08
*** elyezer has joined #zuul21:09
openstackgerritMerged openstack-infra/nodepool master: Partial revert for disabled provider change  https://review.openstack.org/53899521:43
openstackgerritMatthieu Huin proposed openstack-infra/zuul master: [WIP] zuul web: add admin endpoint, enqueue commands  https://review.openstack.org/53900421:43
*** myoung is now known as myoung|bbl21:46
fungino meeting today?22:02
Shrewsfungi: nope22:02
fungithanks!22:03
*** threestrands has joined #zuul22:05
*** threestrands has quit IRC22:05
*** threestrands has joined #zuul22:05
*** threestrands_ has joined #zuul22:07
*** threestrands has quit IRC22:08
*** threestrands_ has quit IRC22:08
*** threestrands has joined #zuul22:08
*** dtruong has quit IRC22:13
*** dmellado has quit IRC22:17
*** dmellado has joined #zuul22:20
*** threestrands_ has joined #zuul22:21
*** threestrands has quit IRC22:23
*** jappleii__ has joined #zuul22:35
*** jappleii__ has quit IRC22:36
*** jappleii__ has joined #zuul22:37
*** threestrands_ has quit IRC22:37
openstackgerritMatthieu Huin proposed openstack-infra/zuul master: zuul autohold: allow filtering per commit  https://review.openstack.org/53699322:43
*** dtruong has joined #zuul23:04
*** weshay|ruck is now known as weshay|ruck|afk23:07
*** rlandy is now known as rlandy|bbl23:11
*** Wei_Liu has quit IRC23:13
*** logan- has quit IRC23:45
*** clarkb has quit IRC23:45
*** smyers has quit IRC23:45
*** logan- has joined #zuul23:51
*** clarkb has joined #zuul23:51
*** smyers has joined #zuul23:51

Generated by irclog2html.py 2.15.3 by Marius Gedminas - find it at mg.pov.lt!