Monday, 2020-03-09

*** threestrands has joined #openstack-infra00:12
*** lbragstad_ has joined #openstack-infra00:38
*** jamesmcarthur has quit IRC00:39
*** lbragstad_ has quit IRC00:44
*** dayou has quit IRC00:46
*** dayou has joined #openstack-infra00:48
fungigmann: as far as gerrit is concerned, x/foo and openstack/foo are repository names, so a repository can be renamed from x/foo to openstack/foo using that process00:56
fungithat way the original review history is retained in the renamed project00:57
fungiand redirects are created in gitea and so on00:57
*** lbragstad_ has joined #openstack-infra01:09
*** matt_kosut has joined #openstack-infra01:25
*** matt_kosut has quit IRC01:30
*** imacdonn has quit IRC01:53
*** lbragstad_ has quit IRC01:57
*** dannins has joined #openstack-infra02:26
*** dave-mccowan has joined #openstack-infra02:26
*** dave-mccowan has quit IRC02:32
*** lbragstad_ has joined #openstack-infra02:57
*** jamesmcarthur has joined #openstack-infra03:11
*** igordc has joined #openstack-infra03:19
*** lbragstad_ has quit IRC03:20
*** ramishra has joined #openstack-infra03:27
*** matt_kosut has joined #openstack-infra03:27
*** jamesmcarthur has quit IRC03:31
*** matt_kosut has quit IRC03:31
*** jamesmcarthur has joined #openstack-infra03:35
*** jamesmcarthur has quit IRC03:40
*** dave-mccowan has joined #openstack-infra03:42
*** ricolin has quit IRC03:43
*** armax has quit IRC03:49
*** ykarel|away is now known as ykarel04:24
*** ricolin has joined #openstack-infra04:27
*** dchen has quit IRC04:27
*** dave-mccowan has quit IRC04:30
*** dchen has joined #openstack-infra04:46
*** matt_kosut has joined #openstack-infra05:27
*** matt_kosut has quit IRC05:32
*** evrardjp has quit IRC05:35
*** evrardjp has joined #openstack-infra05:35
*** matt_kosut has joined #openstack-infra06:26
*** igordc has quit IRC06:31
*** threestrands has quit IRC06:42
*** admcleod has quit IRC06:47
*** ricolin has quit IRC06:52
*** rcernin has quit IRC07:06
*** lmiccini has joined #openstack-infra07:09
*** AJaeger has quit IRC07:12
openstackgerritJan Kubovy proposed zuul/zuul master: Scheduler test app manager  https://review.opendev.org/70881207:12
openstackgerritJan Kubovy proposed zuul/zuul master: Use scheduler manager consistently in tests  https://review.opendev.org/70954207:12
openstackgerritJan Kubovy proposed zuul/zuul master: Refactor executor_client in tests  https://review.opendev.org/70967207:12
openstackgerritJan Kubovy proposed zuul/zuul master: Refactor merge_client in tests  https://review.opendev.org/70967607:12
openstackgerritJan Kubovy proposed zuul/zuul master: Refactor nodepool in tests  https://review.opendev.org/70970307:12
openstackgerritJan Kubovy proposed zuul/zuul master: Refactor zookeeper in tests  https://review.opendev.org/70970907:12
openstackgerritJan Kubovy proposed zuul/zuul master: Consolidate scheduler pause/exit as hibernation  https://review.opendev.org/70972307:12
openstackgerritJan Kubovy proposed zuul/zuul master: Refactor `self.event_queues` in tests  https://review.opendev.org/70999007:12
openstackgerritJan Kubovy proposed zuul/zuul master: Scheduler's pause/resume functionality  https://review.opendev.org/70973507:12
openstackgerritJan Kubovy proposed zuul/zuul master: WIP: Store unparsed branch config in Zookeeper  https://review.opendev.org/70571607:12
*** AJaeger has joined #openstack-infra07:16
*** dpawlik has joined #openstack-infra07:23
*** matt_kosut has quit IRC07:38
*** ykarel is now known as ykarel|lunch07:39
*** pgaxatte has joined #openstack-infra07:40
*** rpittau|afk is now known as rpttau07:42
*** rpttau is now known as rpittau07:42
*** hashar has joined #openstack-infra07:46
*** tetsuro has joined #openstack-infra08:00
*** slaweq has joined #openstack-infra08:01
*** matt_kosut has joined #openstack-infra08:04
*** tkajinam has quit IRC08:07
*** tesseract has joined #openstack-infra08:12
*** tosky has joined #openstack-infra08:15
*** admcleod has joined #openstack-infra08:18
*** iurygregory has joined #openstack-infra08:20
*** jcapitao has joined #openstack-infra08:26
*** amoralej|off is now known as amoralej08:26
*** tetsuro has quit IRC08:29
*** tetsuro has joined #openstack-infra08:31
*** jpena|off is now known as jpena08:31
*** ricolin_ has joined #openstack-infra08:33
*** dtantsur|afk is now known as dtantsur08:35
openstackgerritYumengBao proposed openstack/project-config master: Add rss link for cyborg-specs  https://review.opendev.org/71187508:37
openstackgerritJan Kubovy proposed zuul/zuul master: Use scheduler manager consistently in tests  https://review.opendev.org/70954208:43
openstackgerritJan Kubovy proposed zuul/zuul master: Refactor executor_client in tests  https://review.opendev.org/70967208:43
*** ralonsoh has joined #openstack-infra08:52
openstackgerritJan Kubovy proposed zuul/zuul master: Refactor merge_client in tests  https://review.opendev.org/70967608:54
openstackgerritLiang Fang proposed openstack/project-config master: New repo: devstack-plugin-open-cas  https://review.opendev.org/71187808:55
*** rpittau is now known as rpittau|bbl08:56
*** ykarel|lunch is now known as ykarel08:59
*** ricolin_ has quit IRC09:00
*** ociuhandu has joined #openstack-infra09:01
*** ricolin_ has joined #openstack-infra09:02
openstackgerritJan Kubovy proposed zuul/zuul master: Refactor nodepool in tests  https://review.opendev.org/70970309:08
*** yolanda has quit IRC09:11
*** yolanda has joined #openstack-infra09:11
*** pkopec has joined #openstack-infra09:23
*** ricolin_ has quit IRC09:23
*** derekh has joined #openstack-infra09:37
*** ijw has joined #openstack-infra09:42
*** apetrich has joined #openstack-infra09:46
*** ijw has quit IRC09:46
*** roman_g has joined #openstack-infra09:46
*** ociuhandu has quit IRC09:48
*** gfidente has joined #openstack-infra09:51
*** happyhemant has joined #openstack-infra09:53
*** owalsh^ is now known as owalsh10:00
*** gshippey has joined #openstack-infra10:01
openstackgerritJan Kubovy proposed zuul/zuul master: Refactor zookeeper in tests  https://review.opendev.org/70970910:03
*** xek_ has joined #openstack-infra10:05
*** auristor has quit IRC10:08
*** zbr|pto is now known as zbr10:20
openstackgerritJan Kubovy proposed zuul/zuul master: Consolidate scheduler pause/exit as hibernation  https://review.opendev.org/70972310:22
openstackgerritJan Kubovy proposed zuul/zuul master: Refactor `self.event_queues` in tests  https://review.opendev.org/70999010:26
openstackgerritJan Kubovy proposed zuul/zuul master: Scheduler's pause/resume functionality  https://review.opendev.org/70973510:26
openstackgerritJan Kubovy proposed zuul/zuul master: WIP: Store unparsed branch config in Zookeeper  https://review.opendev.org/70571610:26
*** sshnaidm|afk is now known as sshnaidm10:27
*** auristor has joined #openstack-infra10:36
*** yboaron has joined #openstack-infra10:51
*** dchen has quit IRC10:58
openstackgerritDonny Davis proposed openstack/project-config master: Setting OpenEdge Provider to 10 test nodes  https://review.opendev.org/71190311:11
donnydlooks to me like we are able to produce successful builds http://logstash.openstack.org/#/dashboard/file/logstash.json?query=node_provider:%5C%22openedge-us-east%5C%22%20AND%20message:%5C%22Upload%20logs%20to%20swift%5C%22%20&from=12h11:12
donnydI would like to turn Openedge up to 10 test nodes and monitor11:13
*** ricolin_ has joined #openstack-infra11:13
openstackgerritLiang Fang proposed openstack/project-config master: New repo: devstack-plugin-open-cas  https://review.opendev.org/71187811:17
*** ociuhandu has joined #openstack-infra11:17
*** ociuhandu has quit IRC11:19
*** ociuhandu has joined #openstack-infra11:20
*** tetsuro has quit IRC11:23
*** ykarel is now known as ykarel|afk11:26
fricklerdonnyd: approved, ping me if something goes wrong11:27
donnydthank you frickler11:28
donnydI will monitor closely to ensure nothing blows up11:28
*** yamamoto has quit IRC11:29
*** AJaeger has quit IRC11:33
*** sshnaidm has quit IRC11:36
donnydalso I got the numa settings from sean-k-mooney so everything is back to how it was11:36
openstackgerritMerged openstack/project-config master: Setting OpenEdge Provider to 10 test nodes  https://review.opendev.org/71190311:37
*** nicolasbock has joined #openstack-infra11:37
*** sshnaidm has joined #openstack-infra11:40
*** ociuhandu_ has joined #openstack-infra11:41
*** rosmaita has joined #openstack-infra11:43
*** lbragstad_ has joined #openstack-infra11:43
*** ociuhandu has quit IRC11:44
*** AJaeger has joined #openstack-infra11:47
*** jpena is now known as jpena|lunch11:48
*** jcapitao is now known as jcapitao_lunch11:51
*** lbragstad_ has quit IRC11:51
*** sshnaidm has quit IRC11:53
*** sshnaidm has joined #openstack-infra11:54
*** weshay is now known as weshay|ruck11:54
*** rlandy has joined #openstack-infra11:58
*** yamamoto has joined #openstack-infra12:01
*** yamamoto has quit IRC12:06
*** yamamoto has joined #openstack-infra12:07
*** jamesmcarthur has joined #openstack-infra12:10
*** jamesmcarthur has quit IRC12:14
*** ykarel|afk is now known as ykarel12:17
*** tetsuro has joined #openstack-infra12:18
*** jamesmcarthur has joined #openstack-infra12:20
openstackgerritMohammed Naser proposed opendev/lodgeit master: Upload container images  https://review.opendev.org/71185412:23
*** yamamoto has quit IRC12:29
*** jpena|lunch is now known as jpena12:31
openstackgerritCédric Jeanneret (Tengu) proposed openstack/project-config master: Add new Validation Framework projects  https://review.opendev.org/71191012:32
*** jamesmcarthur has quit IRC12:36
*** takamatsu has quit IRC12:37
*** Goneri has joined #openstack-infra12:40
*** rh-jelabarre has joined #openstack-infra12:41
*** rpittau|bbl is now known as rpittau12:44
*** jamesmcarthur has joined #openstack-infra12:47
*** AJaeger has quit IRC12:50
openstackgerritDonny Davis proposed openstack/project-config master: Bumping OpenEdge test node commit to 20  https://review.opendev.org/71191412:53
donnydTesting at 10 nodes looks good to me - no failures from a launch / connection perspective12:54
*** lbragstad has joined #openstack-infra12:54
donnydI would like to bump up the test node commit to 20 test nodes and then leave it there for a while12:54
*** yamamoto has joined #openstack-infra12:54
*** jamesmcarthur has quit IRC12:56
donnydThere have been zero launch errors since I fixed the networking issue yesterday12:56
*** jamesmcarthur has joined #openstack-infra12:57
*** rh-jelabarre has quit IRC12:58
*** ricolin_ has quit IRC13:00
*** jamesmcarthur has quit IRC13:02
*** zxiiro has joined #openstack-infra13:04
*** lbragstad has quit IRC13:06
*** sshnaidm has quit IRC13:08
*** sshnaidm has joined #openstack-infra13:09
*** tetsuro has quit IRC13:09
*** sshnaidm has quit IRC13:10
*** ricolin_ has joined #openstack-infra13:10
openstackgerritCédric Jeanneret (Tengu) proposed openstack/project-config master: Add new Validation Framework projects  https://review.opendev.org/71191013:11
*** jamesmcarthur has joined #openstack-infra13:13
*** jcapitao_lunch is now known as jcapitao13:14
*** sshnaidm has joined #openstack-infra13:18
*** sshnaidm has quit IRC13:18
*** ociuhandu_ has quit IRC13:19
*** sshnaidm has joined #openstack-infra13:19
*** ociuhandu has joined #openstack-infra13:19
*** bdodd has joined #openstack-infra13:21
*** hashar has quit IRC13:29
*** cdearborn has joined #openstack-infra13:29
*** ociuhandu has quit IRC13:31
*** ociuhandu has joined #openstack-infra13:31
*** jamesmcarthur has quit IRC13:32
*** jamesmcarthur has joined #openstack-infra13:32
*** ricolin_ has quit IRC13:33
*** ykarel is now known as ykarel|afk13:33
*** amoralej is now known as amoralej|lunch13:34
*** rh-jelabarre has joined #openstack-infra13:37
*** apetrich has quit IRC13:37
openstackgerritMerged openstack/project-config master: Bumping OpenEdge test node commit to 20  https://review.opendev.org/71191413:37
*** yamamoto has quit IRC13:37
*** jamesmcarthur has quit IRC13:38
openstackgerritAlbin Vass proposed zuul/zuul-jobs master: Make revoke-sudo more general.  https://review.opendev.org/70626213:43
*** yamamoto has joined #openstack-infra13:44
openstackgerritBenjamin Schanzel proposed zuul/zuul-jobs master: Kubernetes Node Support for Mirroring Git Repos  https://review.opendev.org/71192013:44
openstackgerritAlbin Vass proposed zuul/zuul-jobs master: Adds variable to toggle whether to revoke sudo  https://review.opendev.org/70624813:45
*** AJaeger has joined #openstack-infra13:46
*** eharney has joined #openstack-infra13:47
*** AJaeger has quit IRC13:47
*** AJaeger has joined #openstack-infra13:49
Tenguhello there! quick question: I apparently would need a new wiki namespace on wiki.openstack.org for a new "project" (linked to Rhttps://review.opendev.org/#/c/711910/) - I'm currently editing the governance thing as Andreas notified, and am a bit strugling for the "url" part..13:50
Tenguunless.... hm. I might just drop it under "TripleO" for now13:51
Tenguweshay|ruck: any thoughts? -^^  or do you want to discuss it during the mtg tomorrow?13:52
AJaegerTengu: you set it up with notifications etc. as part of tripleo. If that is not correct, you can set it up outside of openstack namespace13:52
AJaegerTengu: but if you use openstack/ as prefix, I need a governance change.13:53
TenguAJaeger: errr.... my brain just froze with your first sentence13:53
*** ociuhandu has quit IRC13:53
AJaegerTengu: what did I do wrong?13:53
TenguAJaeger: and I agree with the governance change - no problem with that. Just a bit lost as to "what's the best thing to do"13:53
AJaegerTengu: discuss with weshay|ruck and tripleo team first ;)13:54
TenguAJaeger: yes, that's the main idea - I have a point during the meeting tomorrow :)13:54
*** ociuhandu has joined #openstack-infra13:54
TenguI just pushed the change request today in order to ensure everything is ready :).13:54
AJaegerThen WIP it for now ;)13:55
AJaegerthis all can wait from my side...13:55
*** dave-mccowan has joined #openstack-infra13:55
Tengunp - I should have -w it before, sorry13:55
*** lbragstad has joined #openstack-infra13:57
*** ociuhandu has quit IRC13:58
*** ScottMC has joined #openstack-infra14:01
*** jamesmcarthur has joined #openstack-infra14:03
*** adriancz has joined #openstack-infra14:03
openstackgerritAlbin Vass proposed zuul/zuul-jobs master: Control log archive and user preservation with vars  https://review.opendev.org/70138114:04
*** ykarel|afk is now known as ykarel14:07
openstackgerritAlbin Vass proposed zuul/zuul master: Fix minor spelling error  https://review.opendev.org/71192614:08
*** jamesmcarthur has quit IRC14:09
*** sshnaidm has quit IRC14:15
*** amoralej|lunch is now known as amoralej14:15
*** yamamoto has quit IRC14:16
*** armax has joined #openstack-infra14:20
*** artom has joined #openstack-infra14:21
*** Lucas_Gray has joined #openstack-infra14:22
*** sshnaidm has joined #openstack-infra14:28
mnasermay i have eyes on https://review.opendev.org/#/c/711854/ please14:34
*** rh-jelabarre has quit IRC14:34
*** rh-jelabarre has joined #openstack-infra14:34
*** sshnaidm has quit IRC14:35
*** jamesmcarthur has joined #openstack-infra14:39
*** jamesmcarthur has quit IRC14:44
*** lpetrut has joined #openstack-infra14:49
*** sshnaidm has joined #openstack-infra14:50
*** sshnaidm_ has joined #openstack-infra14:52
*** sshnaidm has quit IRC14:54
*** sshnaidm_ is now known as sshnaidm14:56
*** beekneemech is now known as bnemec15:04
*** ykarel is now known as ykarel|away15:05
*** jamesmcarthur has joined #openstack-infra15:18
*** jamesmcarthur has quit IRC15:19
*** jamesmcarthur_ has joined #openstack-infra15:19
*** rh-jelabarre has quit IRC15:29
*** rh-jelabarre has joined #openstack-infra15:33
*** mattw4 has joined #openstack-infra15:34
openstackgerritMerged opendev/lodgeit master: Upload container images  https://review.opendev.org/71185415:40
mordredmnaser. noonedeadpunk: woot!15:44
*** apetrich has joined #openstack-infra15:45
*** jamesmcarthur_ has quit IRC15:47
clarkbdoes anyone have python3.5 handy? I think we can remove our workaround for importlib-resources to fix virtualenv and tox now that importlib-resources 1.3.x have released15:50
clarkbI 've tested that python2.7 is working (and it does work)15:50
clarkbI'm also going to recheck my zuul-jobs DNM chnage that uses base-test as a parent job15:50
clarkbthat should run without the workaround15:50
clarkbhttps://review.opendev.org/#/c/680178/4 is that change15:51
clarkbdonnyd: we can safely delete this grafana dashboard right? http://grafana.openstack.org/d/3Bwpi5SZk/nodepool-fortnebula?orgId=115:53
clarkb(that doesn't delete the data from graphite, just the easy access dashboard)15:54
mnasermordred: yay thanks!15:54
mordredclarkb: I can have one convenient pretty quickly15:55
*** Lucas_Gray has quit IRC15:56
clarkbmordred: ya I've started a xenial docker container and am just making sure I've got it close enough to test node python before testing15:56
mordredah - test node python I definitely won't have15:56
clarkbI remembered that its actually python3.5 on xenial that matters not just any python3.5 because python3.5 and python3.6 apparently didn't fork properly15:56
mordredyeah- I've got pyenv python3.5 - so that's not xenial python3.5 at all15:57
*** ociuhandu has joined #openstack-infra15:57
*** jamesmcarthur has joined #openstack-infra15:58
*** Lucas_Gray has joined #openstack-infra15:59
fungisame, my py35 is via make altinstall from a cpython checkout of the latest 3.5.x tag15:59
fungibuilt against current state of libraries on debian/unstable16:00
fungiso bears little resemblance to whatever xenial is shipping packaged16:00
clarkbok on a xenial container I've used python3 -m venv to createa  virtualenv, then in that virtualenv I've installed virtualenv to latest using -U. THis gets my importlib-resources 1.3.1. That pip install -U virtualenv reports distlib failed to install but returns 0 anyway and running venv/bin/virtualenv doesthisevenwork succeeds16:00
mordredfungi: yup16:00
clarkbI think that means we can safely clean up our workaround if we confirm our images have all built within the last couple days16:01
mordredclarkb: woot16:01
fungisgtm16:01
donnydYea we can purge the old FN stuff16:01
*** diablo_rojo has joined #openstack-infra16:02
*** lpetrut has quit IRC16:05
*** Lucas_Gray has quit IRC16:06
*** Lucas_Gray has joined #openstack-infra16:08
clarkbinfra-root looking at image builds really quickly we have lots of old images sticking around. My hunch is that those are related to leaked volumes in vexxhost16:09
clarkbI'm in a meeting now, but will look closer afterwards. My hunch is that we'll need to clean all that up and nesure things are all new enough before we remove the workaround16:09
*** matt_kosut has quit IRC16:14
*** lmiccini has quit IRC16:25
*** pgaxatte has quit IRC16:28
openstackgerritAlbin Vass proposed zuul/zuul-jobs master: Control log archive and user preservation with vars  https://review.opendev.org/70138116:29
*** yboaron has quit IRC16:30
openstackgerritAlbin Vass proposed zuul/zuul-jobs master: Make revoke-sudo more general.  https://review.opendev.org/70626216:31
*** jamesmcarthur has quit IRC16:35
*** jamesmcarthur has joined #openstack-infra16:37
*** matt_kosut has joined #openstack-infra16:44
*** matt_kosut has quit IRC16:48
clarkbroman_g: to followup on the airship testing, are those nodes being allocated reliably now?16:48
*** imacdonn has joined #openstack-infra16:49
*** nicolasbock has quit IRC16:53
*** apetrich has quit IRC16:54
*** ociuhandu has quit IRC16:54
*** ociuhandu has joined #openstack-infra16:55
*** nicolasbock has joined #openstack-infra16:55
clarkbfungi: donnyd http://paste.openstack.org/show/790485/ I think that represents the bulk of our leaked images beacuse those aren't in a deleting state (if they were in a deleting state we'd remove them from local disk whichi s full on nb02)16:56
clarkbI think that will require zk surgery, which I can look into in a bit16:56
clarkb(vexxhost does have at least one leaked image but I think we deletedt he local copy due to the deleting state change there)16:56
clarkbbut I think non of those fn resources exist anymore so we can delte them from the zk db directly16:57
*** ijw has joined #openstack-infra16:57
clarkbthen that will result in necessary state updates to image state (and deletions)16:57
fungiahh, yeah, i think something's no longer deleting images when we set the images list to an empty list16:58
fungibecause over the weekend when we added the openedge environment with an empty images list, nodepool happily uploaded all our images to it anyway16:58
clarkbneat16:58
fungiShrews: ^ any suggestions for troubleshooting that?16:59
clarkbok back in a few to poke at zk unless someone else would like to16:59
*** ijw has quit IRC16:59
*** ijw has joined #openstack-infra17:00
*** chandankumar is now known as raukadah17:00
*** jamesmcarthur has quit IRC17:00
*** ociuhandu has quit IRC17:01
*** ociuhandu has joined #openstack-infra17:03
*** ijw_ has joined #openstack-infra17:05
*** AJaeger has quit IRC17:05
*** AJaeger has joined #openstack-infra17:05
Shrewsreading sb17:05
donnydI could probably being the FN endpoint back online, but would take a day or so17:06
Shrewsfungi: can i get more background here?17:07
*** ijw has quit IRC17:08
Shrewsdid we not properly decommission FN or something?17:08
donnydShrews: FN was left up for a week with the image list empty17:08
Shrewsdonnyd: that's a confusing statement to me because http://paste.openstack.org/show/790485/ shows the image list not empty... so.... ?  I'm clearly missing context here17:09
donnydWe moved fort nebula to open edge17:09
ShrewsAnd during this move, was FN decommissioned as outlined in https://zuul-ci.org/docs/nodepool/operation.html#removing-a-provider ?17:10
ShrewsOr has something else led up to this?17:10
fungiShrews: we set diskimages: [] for fortnebula in https://review.opendev.org/70925717:11
*** jamesmcarthur has joined #openstack-infra17:11
fungiapparently images did not get cleaned up after that merged17:11
Shrewsfungi: was it disabled in the launcher first? we have bright red boxes on that op page warning us to do that   :)17:12
Shrewsif nodes remain that use those images, they won't get cleaned up17:12
fungiwe set max-servers: 017:12
fungibut maybe there were leaked nodes?17:12
clarkbthere are only two nodes remaining per my paste17:12
clarkbshouldnt prevent all deletes like that17:13
fungiShrews: but then we merged https://review.opendev.org/711760 to replace fortnebula with openedge17:14
fungiand nodepool immediately uploaded images to it even though diskimages: [] was in there for the new environment17:14
Shrewsclarkb: correct, shouldn't prevent all deletes. i'm just trying to understand the sequence of things rn17:14
Shrewsdid we let all FN nodes get used up after setting max-servers to 0?17:15
clarkbthe issue us that change only updated the launcher17:15
clarkbnot nodepool.yaml for the builders17:15
openstackgerritSorin Sbarnea proposed zuul/zuul-jobs master: Tests bindep role on all-platforms  https://review.opendev.org/70870417:15
clarkbnot sure if openedge bringup had similar problem17:16
fungioh...17:16
fungiyep, thanks Shrews17:16
clarkbbut I think we just need o rm zk nodes17:16
fungithat's what did it17:16
fungiwe set the launcher to diskimages: [] but the builders still had the old diskimages list configured17:17
* fungi sighs17:17
fungiand then we renamed the environment for it, explaining why it immediately uploaded17:17
ShrewsWe need to build a way into nodepool CLI to help cleanup here. We shouldn't ever require manual ZK cleanup (but looks like we may need to do that here).17:18
Shrewsclarkb: i think just deleting the zk nodes won't trigger disk cleanup though. that may also need to be done manually17:20
Shrewslemme look at the code for a sec....17:20
clarkbiirc it scans for images that either are deleting or have no record then deletes them17:21
clarkbremoving the zk entries should cause local disk cleanups17:21
*** ccamacho has quit IRC17:22
fungii believe it did last time i deleted znodes for old images17:23
Shrewsdo we still have FN defined on the builders then?17:23
Shrewsi think if we only delete the *upload* records (not the build records), then it should clean up disk17:24
clarkbShrews: correct17:24
*** rpittau is now known as rpittau|afk17:27
Shrewsclarkb: i have to leave in 10 min to meet with my tax guy. can you handle the deletes? should be paths like /nodepool/images/centos-7/builds/0000121707/providers/fortnebula-regionone/images/000000000117:28
*** dtantsur is now known as dtantsur|afk17:30
Shrewsi'll see about some sort of "force-delete-upload-records" option for the CLI17:31
*** apetrich has joined #openstack-infra17:31
clarkbya I can do the deletes17:33
*** jpena is now known as jpena|off17:33
clarkbbasically delete the things in my paste17:33
clarkbthat should then clean up builds17:33
*** ociuhandu_ has joined #openstack-infra17:35
*** evrardjp has quit IRC17:35
*** evrardjp has joined #openstack-infra17:35
*** ociuhandu has quit IRC17:38
*** ociuhandu_ has quit IRC17:39
*** apetrich has quit IRC17:41
*** jcapitao is now known as jcapitao_off17:41
clarkbnodes have been removed. now doing the image uploads17:42
fungii can also do the znode cleanup (i feel responsible for helping make that mess in the first place), but won't be caught up on other stuff to where i can start in on it for a couple more hours17:43
clarkbno worries, its straight forward once we've agreed that is the course of action17:46
clarkbafter these are done I'll check on nb01 and nb02 to see that they've freed the appropriate disk space and are able to build iamges again17:47
clarkbthen we wait for image updates and can clean up the base job17:47
clarkbfwiw rmr fortnebula-regionone where it shows up under the /nodepool/images tree as well as rmr for /nodepool/nodes/$nodeid seems to be the ticket17:47
clarkbcan also get $node to see more about it to confirm that you want to remoev it17:48
fungirmr is the recursive remove?17:48
clarkbyes17:49
fungii think i missed that and manually recursed last time i did it17:49
donnydanything I can do to be helpful?17:52
clarkbdonnyd: nope I think this was all on our end17:52
clarkbwe now have 280GB free on nb0217:52
fungithat was fast!17:52
clarkb312GB on nb0117:52
clarkbI think it may be worthwhile stopping them, rebooting, and cleaning out the dib tmp dir stuff since that tends to bloat17:53
clarkbI'll go ahead and do that now17:53
clarkboh ya we've got tons of dib processes on nb02 at least17:53
clarkbnb02 is rebooting now. It iwll come up with nodepool-builder disabled, I'll clear out the tmp dib dirs, then reenable nodepool-builder and reboot again17:55
clarkbthen repeat on nb0117:55
clarkbthese reboots are always so slow17:57
* fungi remembers rebooting physical servers which took several hours to complete their power-on selftests17:59
fungithis doesn't seem slow at all17:59
*** ccamacho has joined #openstack-infra18:00
*** derekh has quit IRC18:00
*** jamesmcarthur has quit IRC18:00
clarkbat about 5 minutes now. I wonder if it is fscking (fwiw this seems to always happen on these servers, sometimes I wonder if it is the stop side that is slow since ssh is immediately killed it appaers to be fast but could be trying to gracefully stop stuff after that takes time)18:01
*** jcapitao_off has quit IRC18:07
*** Lucas_Gray has quit IRC18:08
*** jamesmcarthur has joined #openstack-infra18:08
*** andrewbonney has quit IRC18:10
mordredfungi: I remember the good old days of being terrified to reboot a server because there was a chance it would fail POST or just simply not boot back up properly for some reason18:14
clarkbstill not responding and there is nothing on the console. I think that is fsck behavior with ubuntu?18:15
mordredwhich might lead to a couple of days of manual reconstruction18:15
clarkboh it just started doing a boot splash18:15
mordredclarkb: yeah - sometimes it just goes to the bad place18:15
mordred\o/18:15
fungiclarkb: you can `sudo touch /fastboot` before rebooting to skip forced timed fsck of filesystems18:15
mordredclarkb: you know - there is a grub/kernel option that disables the quiet boot thing18:15
clarkbwell the fsck is probably a good idea. disablibg quiet boot seems like a good idea too18:16
fungimordred: keyboard error, press f1 to continue18:16
clarkbmostly its the lack of info that is annoying more so than the server doing what it needs to check its disks are sane18:16
clarkbanyway I will wait patiently since it seems to be doing something (likely fsck)18:16
mordredyeah. these are servers - please print lots of lines of text to console18:16
*** jamesmcarthur has quit IRC18:16
fungii'm being told i need to take advantage of the unusual warm snap here to go for a brief walk. bbiab18:17
*** jamesmcarthur has joined #openstack-infra18:17
clarkbfungi: I'll be taking advantage of the sun here in about an hour :)18:17
clarkbhighly recommend18:17
fungigood call18:17
clarkbI went on a bike ride on saturday and regretted it when the skies decided hail was appropriate18:18
clarkbwas small hail but it got so cold out of nowhere18:18
clarkbanyone know how to give ctrl + alt + f1 in the rax console? Is that what the check marks are for ctrl and alt to have it capture those key presses?18:21
clarkbhrm I think I figured it out. those check boxes seem to be actual inputs so checking them then f1 gives you the keypresses you want18:22
clarkbit appears its stuck on unmounting dib stuff18:22
clarkbwhich isn't unsurprising. I think it actually needs a forceful reboot18:22
clarkbany objections to trying that?18:23
clarkbbasically the lat thing in the log is reached target shutdown with a bunch of unmount failures above it18:23
*** sean-k-mooney has joined #openstack-infra18:24
clarkband now I have ssh access18:25
mordredwoot18:25
mordredclarkb: is the dib unmounting thing something we should try to dig in to?18:26
mordredI seem to remember that coming up before18:26
clarkbmordred: I believe it is a side effect of running out of disk on the server18:26
clarkbmordred: basically when that happens dib starts to fail hard because so much of what it does relies on successful writes18:26
clarkbI'm not sure its worth digging into beyond making dib run out of disk less (which we've been pushing on by cleaning up old images and having it remove the disk files once all upload states are deleting and not actually deleted)18:27
clarkbthe problem this time was we removed a cloud without properly cleaning it up so we basically had a second set of images for all images hanging around18:27
clarkbhowever, its possible something else is causing the leaking18:27
mordredclarkb: oh right18:29
*** amoralej is now known as amoralej|off18:31
clarkbthats said cleaning up dib_tmp is freeing a lot of disk space so those may be leaking which then puts pressure on things even if clouds are all happy18:31
clarkb(I had thought that the cleanups were failing due to running out of space, but this rm is deleting way more data than I would expect if that were the case)18:33
mordredclarkb: :(18:34
clarkbthere is now 467GB free on nb02's volume and i'm still waiting for rm to finihs18:38
clarkbI would've expected an image or two's worth of cleanup if it was running out of disk that caused things to spiral out of control18:38
clarkbnot ~10 images worth18:38
clarkber I guess its "just" 200GB18:38
clarkbwhich is less than 10 images18:38
clarkbbut still18:38
AJaegerclarkb, mordred: do you want to +1 the OpenDev governance change? (the revert-revert) - https://review.opendev.org/#/c/710020/18:39
AJaegermight be good to give some additional +1 to avoid further questions and delays18:40
AJaegerinfra-root ^18:40
clarkbya I'll take a look18:41
*** ralonsoh has quit IRC18:42
*** jamesmcarthur has quit IRC18:44
*** ociuhandu has joined #openstack-infra18:44
mordredinfra-root: running out to store - back in a few18:49
*** jamesmcarthur has joined #openstack-infra18:56
*** eck` has joined #openstack-infra18:58
clarkbok nb02 is all done now and running the builder again19:02
clarkbgoing to look at nb01 next19:02
clarkbnb01 is in a much happier state. I thas leaked ~4 builds looks like not many like nb0219:05
clarkbit could be that there is an underlying bug which is made worse by the disk filling19:05
clarkbok and nb01 is done now19:07
clarkbinfra-root related to images and keeping on top of them, should we be deleting fedora-29 since the upstream packages have been retired?19:07
clarkbmaybe that is a question for ianw19:07
clarkbI'm making a list of images leaked in vexxhost now19:08
*** ociuhandu has quit IRC19:10
AJaegerclarkb: fedora-latest nodeset still points to fedora-2919:10
*** ociuhandu has joined #openstack-infra19:11
*** eharney has quit IRC19:11
openstackgerritAndreas Jaeger proposed opendev/base-jobs master: Switch nodeset fedora-latest to fedora 30  https://review.opendev.org/71196919:13
AJaegerclarkb, ianw, ^19:13
clarkbAJaeger: hrm, ok it was my undertanding that jobs were failing because package mirrors are not working as they retired it upstream of us (then mirrors picked up on that), but maybe it is still working in some capacity?19:13
AJaegerclarkb: no idea19:14
AJaegerclarkb: just noting that we have jobs configured...19:15
clarkb++19:15
clarkbhttp://paste.openstack.org/show/790494/ is a survey of remaining leaks and images that are not building currently19:15
openstackgerritAndreas Jaeger proposed opendev/glean master: Switch to Fedora 30 jobs  https://review.opendev.org/71197019:16
*** ociuhandu has quit IRC19:16
*** gfidente is now known as gfidente|afk19:16
AJaegerclarkb: once those two are merged and one for x/tobiko, we can merge diskimage-builder and retire fedora-2919:17
clarkbthe buster images are broken due to a change I made a while back that got back burnered. I'll look into fixing that first19:18
fungiwalk concluded. back and catching up while lunching on leftovers from the weekend19:19
fungiweather was nice19:19
fungino hail ;)19:19
openstackgerritAndreas Jaeger proposed openstack/diskimage-builder master: Remove Fedora 29 job  https://review.opendev.org/71197219:21
AJaegerianw, clarkb: pushed all changes for fedora 29 (left project-config out), topic is fedora-2919:21
openstackgerritClark Boylan proposed openstack/project-config master: Fix debian-buster partition config  https://review.opendev.org/71197319:24
clarkbinfra-root ^ that should fix the debian buster image builds I thik19:24
*** gyee has joined #openstack-infra19:24
clarkbI need to sort out my time in the sun as well as lunch now19:24
clarkbwill be back to keep digging into nodepool image statuses19:25
openstackgerritAndreas Jaeger proposed zuul/zuul-jobs master: Tests bindep role on all-platforms  https://review.opendev.org/70870419:26
fungithanks, reviewing19:27
*** tesseract has quit IRC19:33
*** dave-mccowan has quit IRC19:38
openstackgerritMerged openstack/project-config master: Fix debian-buster partition config  https://review.opendev.org/71197319:40
*** lbragstad_ has joined #openstack-infra19:54
*** lbragstad has quit IRC19:57
clarkb"ERROR: Cannot uninstall 'six'. It is a distutils installed project and thus we cannot accurately determine which files belong to it which would lead to only a partial uninstall." is the opensuse-15 error19:59
fungiyay20:00
fungialso you should be out enjoying the afternoon20:00
clarkbI should but I decided I wasn't quite ready yet :)20:01
clarkbI think our plan to cleanup python on test images will fix opensuse-1520:01
clarkbI'm not really seeing a good shorter term answer20:01
fungiyeah, any idea what's dragging in six in the first place?20:02
clarkbfungi: its the python2-pip install done before switching over to source in pip-and-virtualenv20:03
clarkbtumbleweed has the same error20:03
fungiahh, okay20:03
clarkbthat means that newer packaging (that will trickle down into leap) isn't fixed either20:03
clarkbAJaeger: ^ suse may want to consider fixing that too20:03
clarkbAJaeger: basically setuptools should be used instead of distutils so that pip is happier20:03
AJaegerdirk, cmurphy, evrardjp, can either of you followup, please? ^20:05
* AJaeger is too far away from python packaging...20:05
clarkbpython2-six is the package looks like20:06
*** jamesmcarthur has quit IRC20:06
clarkbbut any others in the same boat should eventually be updated20:06
dirkclarkb: I will take a look20:08
AJaegerthanks, dirk !20:10
*** dave-mccowan has joined #openstack-infra20:12
dirkthis issue was re-added in20:13
dirkSat Aug 18 09:08:38 UTC 2018 - Matěj Cepl <mcepl@suse.com>20:13
dirkSat Aug 18 09:08:38 UTC 2018 - Matěj Cepl <mcepl@suse.com>20:13
dirk- Break the cycilical dependency on python-setuptools.20:13
dirkso setuptools needs six to build, and six setuptools20:13
*** jamesmcarthur has joined #openstack-infra20:14
mordredsigh20:14
fungii guess suse can't have manual uploads of non-redistributed binary packages to break dependency cycles like in debian20:15
mordredclarkb: I agree - the new python plan should fix this issue20:16
*** jamesmcarthur has quit IRC20:17
fungiyep, not preinstalling python2-pip should do the trick, as long as it's the only thing dragging in six as a dependency20:17
mordredclarkb: the only thing I could think of to work around the issue on opensuse for now is to go the other way in the short term - uninstall python2pip and python2-six install pip with get-pip20:17
mordredit's not ideal as it's a little more spaghetti code in the element with a special case for suse - but the end result should be mostly ok until we can get it sorted for real20:17
*** jamesmcarthur has joined #openstack-infra20:18
fungiand, again, assuming that's the only package depending on it in those images20:18
mordredyah20:19
dirkfungi: well, we can with tricks. but cycles are still frowned upon as they we automatically rebuild full cycles20:19
clarkbya though if we get the images built dealing with this on the distro is a job/software prolem20:19
*** cdearborn has quit IRC20:19
fungidirk: yep, obviously getting rid of cyclic dependencies is the preferred solution. just isn't always possible20:19
clarkbfwiw I've confiemd that the image delete failures in vexxhost are due to leaked volume for boot from volume nodes20:20
clarkbI'm now cleaning up some volumes20:20
*** AJaeger has quit IRC20:21
clarkbI have deleted 3 clearly leaked volumes and 3 volumes stuck in a creating state since new years eve. This cleaned up 3 out of 5 images20:26
clarkbtrying to see where the other two are stuck now20:26
clarkbThere are no held nodes in sjc1 currently so not that20:27
*** ccamacho has quit IRC20:29
clarkb`openstack --os-cloud openstackjenkins-vexxhost --os-region sjc1 server remove volume 485a607e-29dc-4b1e-b2db-a65027757202 89b630af-581e-488f-8fb0-72751bf74652` would make one of the leaked volumes deletable if that command would work20:30
clarkbmordred: ^ did you ever sort out if we could do that more forceful detachment once the server no longer exists20:30
clarkbthat command results in error becaues the server is gone20:31
fungiso i think the solution was that there is a direct cinder api call we can make to "detach" the volume from its perspective even though nova no longer knows about the server instance to which it's supposedly still attached20:33
fungiand i want to say there was work to add that as a fallback in openstacksdk but now i don't remember20:34
*** xek_ has quit IRC20:36
jrosserplease could i get a hold on openstack-ansible-deploy-aio_metal-debian-buster on review 71182120:36
fungijrosser: what are you troubleshooting in that job? just so i can be more specific in the autohold comment20:37
jrosser"symlink to zuul provided repos"20:37
mordredwe had a script to do it20:37
mordredI thought I pushed that up somewhere no?20:38
clarkbinfra-root mnaser http://paste.openstack.org/show/790497/ is what I've found20:38
clarkbmordred: oh maybe20:38
clarkb(sorry was still in investigative mode)20:39
mordredclarkb: oh - that's the other half of the story20:39
clarkbI think the other three volumes need cloud intervention because they are attached to nodes from new years eve that refuse to delete20:39
fungijrosser: added. for future reference we also need to know the repository name, though i inferred it from the change details20:39
mordredclarkb: in tools/clean-leaked-volumes.py20:39
jrosserfungi: ooh ok, i'll bear that in mind for next time - thanks for adding it20:40
fungiyw!20:40
mordredin system-config - the c.block_storage.delete call is what you're looking for20:40
mordredclarkb: you need to delete the attachement the volume has20:40
clarkbmordred: thanks. Though I'm actually about to pop out for a bike ride in the sun now. I can figure out running that if others are busy20:40
clarkbmordred: ya then I can delete the volume20:40
mordredclarkb: yah20:40
*** sshnaidm is now known as sshnaidm|afk20:40
clarkbthe other three in that paste are gonna bestuck I think because the instances still exist20:40
mordredclarkb: do we need to delete the instances too?20:41
clarkbmordred: I've tried, they won't delete :)20:41
mordredwe can still delete the volumes :)20:41
clarkbmordred: we can detach even if the instance is running?20:41
clarkbthat seems dangerous but I guess its fine20:41
mordredwith taht second command - it's basically "sudo hey cinder detach this"20:41
fungisudo cinder make me a sandwich20:41
clarkbok not safe in the general case but since those server instances are test nodes that nodepool doesn't know about that means we don't care about them and can do our best to cleanup20:42
clarkbwe should still eventually clean up those instances though (and that requires cloud intervention I think)20:42
mordredunless we can find more information about why they won't delete - yeah, I think we need mnaser to delete them20:43
clarkbmordred: they are timestamped around the same time I found volumes stuck in a creating state (those did delete)20:43
clarkbmordred: my guess is ceph or something in the cloud was unhappy new years eve20:43
mordredjoy20:43
clarkband these instances were reported back to nodepool as failures, but actually "completed" enough to get uuids and sit around and fial to delete20:44
ianwhey sorry yesterday was a holiday here, back today20:45
ianwinfra-root: speaking of fedora, there's a stack in https://review.opendev.org/#/q/status:open+topic:nodepool-legacy that deploys a builder from containers; reviews welcome.  i want to move the fedora builds to that initially20:48
ianw(mordred has already looked at a few, thanks)20:49
clarkbianw: I noticed an nb01.opendev.org which I assume is where ^ will be deployed20:49
clarkb?20:49
ianwclarkb: yep, i brought that up quite a while ago when i was still thinking of a more traditional deployment20:49
clarkbcool, I'll try to revie after my bike ride. I think the more immediate fires are all out at this point20:49
ianwit hasn't been ansiblised yet20:50
clarkband if someone else wants to try and clean up the volumes/intsances in http://paste.openstack.org/show/790497/ feel free20:50
clarkband really popping out now before I lose all this nice sunlight20:51
dirkmordred: clarkb: so should I push for a fix on suse side or not?20:52
clarkbdirk: I think we'll fix/workaround this issue on our side eventually, but the suse packaging should not use distutils anywhere for this reason20:53
clarkbit should probably be fixed in the distro too if possible20:53
*** trident has quit IRC20:57
dirkclarkb: submitted to tumbleweed via bsc#116613920:57
dirkfor leap we'd have to do a SLE update (it is inherited from there)20:57
*** jamesmcarthur has quit IRC20:58
*** trident has joined #openstack-infra20:58
*** jamesmcarthur has joined #openstack-infra20:58
*** zxiiro has quit IRC21:01
*** trident has quit IRC21:04
*** trident has joined #openstack-infra21:05
*** imacdonn has quit IRC21:17
jrosserfungi: i think my hold is ready now, this is my key http://jrosser.woaf.net/openstack.pub, also....  is this expected? http://jrosser.woaf.net/openstack.pub21:20
jrosseroops this http://paste.openstack.org/show/790498/21:20
*** rh-jelabarre has quit IRC21:21
fungijrosser: ansible is problematic in that if there's no room for it to write a temporary script it will pretend the host is unreachable21:22
fungiso we have a fallback to do a raw exec over ssh in order to check whether the rootfs is full21:22
fungithat looks to me like the zuul executor couldn't reach the node at all at that point, or at least that its sshd wasn't responding21:23
fungitime to find out if i can21:23
fungijrosser: ssh root@192.237.253.11521:28
fungijrosser: also someone held a centos-8 node for you two weeks ago, didn't mention in the comment what you were debugging. can we release that one?21:29
openstackgerritTristan Cacqueray proposed zuul/nodepool master: Add mypy to linter test  https://review.opendev.org/71175021:29
jrosserfungi: yes that one can be released21:30
fungithanks, deleting it now21:30
*** sean-k-mooney has quit IRC21:32
mnaserclarkb: appreciate your thoughts on .. https://review.opendev.org/#/c/711861/121:34
*** eharney has joined #openstack-infra21:35
mnaser(i figured that there wouldn't be interest in maintaining the helm charts in infra and that way we can iterate/move forwards faster rather than blocking on folks who have to review things they dont need)21:35
fungiif we were going to use those to deploy paste.opendev.org then it might make sense as part of the opendev repository namespace, but i don't know to what extent we're relying on helm currently nor what future plans we might have to do so21:36
*** rcernin has joined #openstack-infra21:36
mordredI don't believe we have any plans to use helm - so I think mnaser just maintaining them whever makes sense is a fine idea21:37
fungiyeah, seems fine to me too21:37
mnaseryeah my train of thought is.. will this be used by infra =21:38
mnaser=> opendev/ else vexxhost/21:38
fungimy thoughts exactly21:38
mordred++21:38
mnaserbtw -- https://review.opendev.org/#/c/710020/ should be ready to land once we hit the "deadline" of being open for long enough :)21:39
mnaser(re opendev)21:39
fungiwe've been open since the very beginning! ;)21:41
mnaserfungi: oh i meant that the change is open for long enough to merge :p21:41
fungimy monday evening jokes fall a little flat21:42
mnaserit's okay, it's been a long day21:42
donnydHas anyone noticed any issues with Open Edge today?21:42
donnydit seems to be aside from the image thing with nodepool earlier all is well21:43
fungidonnyd: i have not heard/seen any, no21:43
donnydI have only seen one error to launch in the last 12 hours21:44
fungisounds ultra-stable to me then21:44
donnydyea that is not too bad21:45
donnydif there have been no complaints then I am pretty happy with its first day back into operation21:45
*** jamesmcarthur has quit IRC21:47
*** nicolasbock has quit IRC21:55
*** slaweq has quit IRC21:57
*** dpawlik has quit IRC22:00
*** yboaron has joined #openstack-infra22:01
*** ociuhandu has joined #openstack-infra22:03
*** ociuhandu has quit IRC22:07
*** slaweq has joined #openstack-infra22:09
*** zigo has quit IRC22:13
*** slaweq has quit IRC22:14
*** bdodd has quit IRC22:18
*** zigo has joined #openstack-infra22:19
jrosserfungi: we can release the hold on openstack-ansible / openstack-ansible-deploy-aio_metal-debian-buster / 71182122:19
jrosseri've figured out whats going on - the extra disks on a rax node get mounted over the top of a symlink i set up too early in the job22:20
*** bdodd has joined #openstack-infra22:21
fungid'oh, yep that'd do it22:21
fungideleted!22:22
fungiin rackspace the rootfs is small so some jobs mount the ephemeral disk they provide at /opt22:22
jrossertotal fluke that the buster job that failed last time also landed on a rax node for the held job :)22:23
fungior somewhere similar22:23
fungiwell, if it hadn't failed you could have just kept rechecking. autoholds don't trigger on successful builds22:23
clarkbya I don't think we'd have much input for helm charts22:24
clarkbmaybe one day but our current setup is somewhere between that and puppet using ansible to drive docker compose22:25
*** diablo_rojo has quit IRC22:26
clarkbmnaser: fungi I've approved the change22:29
clarkbmnaser: did you see my notes about about servers and volumes that have leaked in vexxhost? I'm about to try manually removing attachments and deleting volumes but the servers will remain leaked I think22:29
ianwclarkb: i was just having a look at those volumes you mentioned to try and help out -- they seem to think they're attached?22:32
*** rkukura has joined #openstack-infra22:32
clarkbianw: yup. One of them is "attached" to a server that does not exist. We can super safely remove that attachment then delete that volume. For the other three I think those servers still exist but refuse to delete for whatever reason. We can remove the attachment and delete the volumes but the servers will probably be even more unhappy after that22:33
clarkbhwoever those servers were created on new years eve and are nodepool nodes so I don't think we care too much22:33
*** rkukura has quit IRC22:34
clarkbI'm hacking up system-config/clean-leaked-bfv.py now to detach those 422:34
mordredclarkb: let me know if that doesn't work22:35
mordred(also I really do need to put that into sdk as some actual api calls)22:36
clarkbmordred: mostly just trying to understand what an attachmend id is22:36
openstackgerritTristan Cacqueray proposed zuul/zuul master: Implement zookeeper-auth  https://review.opendev.org/61915622:36
clarkbbut otherwise I think I should be able to get it running soon enough22:36
clarkb(the script won't work as is because it only removes volumes that don't have servers that exist in server list and we have 3 of those)22:36
*** slaweq has joined #openstack-infra22:37
openstackgerritTristan Cacqueray proposed zuul/nodepool master: Implement zookeeper-auth  https://review.opendev.org/61915522:38
*** rkukura has joined #openstack-infra22:39
*** bradm has joined #openstack-infra22:40
clarkbmordred: ianw http://paste.openstack.org/show/790499/ how does that look?22:40
openstackgerritMerged openstack/project-config master: Add vexxhost/lodgeit-helm  https://review.opendev.org/71186122:40
clarkbI'm going to try that now22:41
ianwclarkb: lgtm if it works and the zombie servers don't hang on to the reference :)22:42
*** slaweq has quit IRC22:42
clarkbianw: I think we are about to find out if they do or not :)22:42
mordredclarkb: wait22:42
mordredoh - nevermind. yes22:42
mordredthat looks good22:42
mordred(I had a quick panic because my brain skipped over volumes_to_detach22:43
clarkball 4 detachments returned 200 and volume list shows them detached22:44
clarkbI'm going to try and delete them now22:44
clarkbthe three that were attached showed deleting and now show available22:45
clarkbso ya I think those three volumes and their servers will need cloud admin intervention22:46
*** yboaron has quit IRC22:46
clarkbmnaser: noonedeadpunk the volumes and servers on lines 7-9 of http://paste.openstack.org/show/790497/ are sad and can/should be deleted22:46
clarkbdebian buster appears to have built successfully too22:47
clarkbits a (northern hemisphere) spring cleaning day!22:48
mordredwoot!22:48
*** slaweq has joined #openstack-infra22:54
*** tkajinam has joined #openstack-infra22:55
*** slaweq has quit IRC22:59
openstackgerritTristan Cacqueray proposed zuul/zuul master: Implement zookeeper-auth  https://review.opendev.org/61915623:00
*** mattw4 has quit IRC23:00
*** gshippey has quit IRC23:01
*** rlandy is now known as rlandy|bbl23:05
*** diablo_rojo has joined #openstack-infra23:12
*** pkopec has quit IRC23:15
*** lbragstad_ has quit IRC23:17
*** jamesmcarthur has joined #openstack-infra23:22
*** jamesmcarthur has quit IRC23:23
*** gyee has quit IRC23:24
*** jamesmcarthur has joined #openstack-infra23:24
*** jamesmcarthur has quit IRC23:25
*** jamesmcarthur has joined #openstack-infra23:25
*** dchen has joined #openstack-infra23:25
clarkbianw I reviewed the nb01.opendev.org stack. I left comments on much of them about some tweaks we can do that might be more maintainable long term. Let me know what you think (the -1 was left because its an actual error aiui but the others are +2 because we can refine as we go)23:26
ianwclarkb: thanks, yeah just thinking about the nodepool.yaml copy.  i think with remote_src: yes it would work23:27
ianwi.e. not copy itself constantly if it didnt' change by testing checksum23:27
clarkbianw: ++ (on a followon change I note where we can simplify the container mount config if we do that copy)23:27
ianwoh, thanks, a missing git add on the last bit23:27
ianwi'm still a bit unsure on the overall lifecycle management of this container ... but think that will come as we test it23:28
ianwi.e. upgrading on releases but not killing things in ways that leave crap around locally and on remote clouds23:28
*** tosky has quit IRC23:30
clarkbianw: having the state in zk means we do a really good job of keeping leaks out of clouds now23:31
clarkbianw: the exceptions there are when clouds themslves fail to allow us to clean up23:32
clarkbthe local stuff I think will be a learning experience. Maybe we can have a container init script that cleans out /opt/dib_tmp?23:32
*** larainema has quit IRC23:33
*** apetrich has joined #openstack-infra23:33
clarkbI guess we'll update the image whenever changes land to nodepool23:39
*** jamesmcarthur has quit IRC23:40
*** jamesmcarthur has joined #openstack-infra23:40
ianwclarkb: yeah, i'm actually thinking the tmps should be volumes too23:41
ianwthe nesting will be interesting too, if depending on how we keep proceeding with image generation23:42
clarkbianw: you mean proper docker volumes rather than bind mounts?23:43
ianwyeah, something like that; attached just for the container lifespan23:43
clarkbhuh I wonder how those will work with mounts leaking23:44
clarkbwe can find out :)23:44
fungiwell, and image leaks when i forget to clear out the diskimages list on the builder config rather than just the launcher config before removing a provider23:45
*** jamesmcarthur has quit IRC23:46
ianwgoing to merge those two base ones to minimise rebase/merge changes23:49

Generated by irclog2html.py 2.15.3 by Marius Gedminas - find it at mg.pov.lt!