Friday, 2019-07-12

fungiand now it's tomorrow00:00
clarkbnow that you point that out I want to upgrade my router /me does this o/00:00
fungiopenbsd has seriously upped their game on upgrades and patching00:00
fungihttps://man.openbsd.org/syspatch00:01
clarkbthis is pfsense so freebsd00:01
clarkblooks like I get newer unbound00:01
fungibinary patching the kernel... gives me vax/vms shivers (in a good way!)00:01
corvusi rechecked 670402 (a zuul-jobs change) and it has a buildset that looks sane00:02
fungisuccess!00:02
clarkbfungi: does it do that for live kernel updates?00:02
fungiit's the on-disk kernel so you still need to reboot00:02
fungimaybe someday we'll have usable systems built on the hurd00:02
corvusoof, it looks like 670413 is hitting a "our linter isn't smart enough" error00:03
corvusi will convert that into a more boring form of yaml for now00:03
openstackgerritJames E. Blair proposed openstack/project-config master: Add required projects to zuul tenant  https://review.opendev.org/67041300:04
*** dychen has joined #openstack-infra00:14
*** dchen has quit IRC00:17
*** calbers has quit IRC00:17
*** calbers has joined #openstack-infra00:17
openstackgerritMerged zuul/zuul master: Switch to opendev release/docs jobs  https://review.opendev.org/67038800:22
*** weifan has joined #openstack-infra00:25
*** weifan has quit IRC00:25
*** betherly has joined #openstack-infra00:26
*** dychen has quit IRC00:26
*** dchen has joined #openstack-infra00:28
openstackgerritMerged openstack/project-config master: Add required projects to zuul tenant  https://review.opendev.org/67041300:28
*** gyee has quit IRC00:28
*** betherly has quit IRC00:30
*** dmsimard4 is now known as dmsimard00:32
*** rcernin has quit IRC00:34
*** rcernin has joined #openstack-infra00:35
*** betherly has joined #openstack-infra00:46
*** dychen has joined #openstack-infra00:47
*** panda has quit IRC00:48
*** dklyle has joined #openstack-infra00:49
*** panda has joined #openstack-infra00:50
*** dchen has quit IRC00:50
*** betherly has quit IRC00:51
openstackgerritJames E. Blair proposed zuul/nodepool master: Switch to zuul tenant jobs for docs/release  https://review.opendev.org/67042200:53
corvusthat should take care of the last remaining alarm bell00:53
*** irclogbot_2 has joined #openstack-infra00:55
*** irclogbot_2 has quit IRC01:00
*** betherly has joined #openstack-infra01:06
*** diablo_rojo has quit IRC01:07
*** betherly has quit IRC01:11
*** imacdonn has quit IRC01:14
*** imacdonn has joined #openstack-infra01:15
*** tdasilva has quit IRC01:15
*** lseki has quit IRC01:23
*** irclogbot_0 has joined #openstack-infra01:25
*** betherly has joined #openstack-infra01:27
openstackgerritMerged zuul/nodepool master: Switch to zuul tenant jobs for docs/release  https://review.opendev.org/67042201:32
*** betherly has quit IRC01:32
*** irclogbot_0 has quit IRC01:34
*** igordc has quit IRC01:35
corvusno alarm bells \o/01:37
fungidisentanglemnent concluded01:38
*** betherly has joined #openstack-infra01:48
*** betherly has quit IRC01:53
openstackgerritFilippo Inzaghi proposed openstack/os-loganalyze master: Change openstack-dev to openstack-discuss  https://review.opendev.org/62236301:56
openstackgerritFilippo Inzaghi proposed opendev/python-storyboardclient master: fix tox python3 overrides  https://review.opendev.org/57434701:57
*** apetrich has quit IRC01:58
*** ijw has quit IRC01:58
openstackgerritFilippo Inzaghi proposed opendev/bindep master: Change openstack-dev to openstack-discuss  https://review.opendev.org/62232502:01
*** lei-zh has joined #openstack-infra02:19
*** yamamoto has joined #openstack-infra02:22
*** irclogbot_1 has joined #openstack-infra02:25
*** lei-zh has quit IRC02:26
*** irclogbot_1 has quit IRC02:30
*** betherly has joined #openstack-infra02:39
*** betherly has quit IRC02:44
*** altlogbot_2 has joined #openstack-infra02:47
*** yamamoto has quit IRC02:48
*** altlogbot_2 has quit IRC02:52
*** betherly has joined #openstack-infra03:00
*** bhavikdbavishi has joined #openstack-infra03:00
*** yamamoto has joined #openstack-infra03:01
*** bhavikdbavishi has quit IRC03:02
*** betherly has quit IRC03:05
*** michael-beaver has quit IRC03:08
*** diablo_rojo has joined #openstack-infra03:09
*** dychen has quit IRC03:09
*** dchen has joined #openstack-infra03:10
*** rlandy has quit IRC03:16
*** betherly has joined #openstack-infra03:20
*** irclogbot_3 has joined #openstack-infra03:21
*** betherly has quit IRC03:25
*** irclogbot_3 has quit IRC03:26
*** betherly has joined #openstack-infra03:42
*** psachin has joined #openstack-infra03:42
*** psachin has quit IRC03:43
*** psachin has joined #openstack-infra03:44
*** betherly has quit IRC03:46
*** irclogbot_3 has joined #openstack-infra03:51
*** whoami-rajat has joined #openstack-infra03:55
*** irclogbot_3 has quit IRC03:56
*** betherly has joined #openstack-infra04:02
*** ykarel|away has joined #openstack-infra04:02
*** udesale has joined #openstack-infra04:05
*** dklyle has quit IRC04:06
*** betherly has quit IRC04:07
openstackgerritMerged zuul/zuul-jobs master: Normalize test jobs yaml  https://review.opendev.org/67019804:31
*** betherly has joined #openstack-infra04:33
*** factor has quit IRC04:35
*** betherly has quit IRC04:38
openstackgerritMerged zuul/zuul-jobs master: Add add-authorized-keys test job  https://review.opendev.org/67019904:45
*** toabctl has quit IRC04:53
*** rcernin has quit IRC04:54
*** toabctl has joined #openstack-infra04:55
*** irclogbot_3 has joined #openstack-infra04:59
*** ykarel|away has quit IRC04:59
*** bhavikdbavishi has joined #openstack-infra05:01
*** yamamoto has quit IRC05:11
*** kjackal has joined #openstack-infra05:12
*** pcaruana has joined #openstack-infra05:13
*** betherly has joined #openstack-infra05:14
*** jistr has quit IRC05:15
*** JpMaxMan has quit IRC05:16
*** irclogbot_3 has quit IRC05:16
*** JpMaxMan has joined #openstack-infra05:17
*** yamamoto has joined #openstack-infra05:18
*** jistr has joined #openstack-infra05:18
*** betherly has quit IRC05:19
*** ykarel|away has joined #openstack-infra05:24
*** ykarel|away is now known as ykarel05:25
*** aedc has quit IRC05:47
*** kjackal has quit IRC05:54
*** jbadiapa has quit IRC05:55
*** yamamoto has quit IRC06:02
*** betherly has joined #openstack-infra06:06
*** yamamoto has joined #openstack-infra06:07
*** kjackal has joined #openstack-infra06:09
*** altlogbot_0 has joined #openstack-infra06:09
*** betherly has quit IRC06:11
*** altlogbot_0 has quit IRC06:14
*** jtomasek has joined #openstack-infra06:25
*** Goneri has joined #openstack-infra06:25
*** rkukura_ has joined #openstack-infra06:29
*** rkukura has quit IRC06:30
*** rkukura_ is now known as rkukura06:30
openstackgerritMerged zuul/zuul-jobs master: Advance ansible-lint cap to test with 4  https://review.opendev.org/66769506:31
*** witek has joined #openstack-infra06:34
*** irclogbot_0 has joined #openstack-infra06:35
*** betherly has joined #openstack-infra06:37
*** yamamoto has quit IRC06:39
*** rkukura has quit IRC06:39
openstackgerritAndreas Jaeger proposed openstack/project-config master: Remove unused docs jobs from dashboard  https://review.opendev.org/67045206:39
*** irclogbot_0 has quit IRC06:40
*** betherly has quit IRC06:42
*** yamamoto has joined #openstack-infra06:43
*** rkukura has joined #openstack-infra06:44
*** yamamoto has quit IRC06:47
*** aedc has joined #openstack-infra06:58
*** ginopc has joined #openstack-infra07:02
*** gtema has joined #openstack-infra07:07
*** Goneri has quit IRC07:17
*** Goneri has joined #openstack-infra07:18
*** rpittau|afk is now known as rpittau07:18
*** tosky has joined #openstack-infra07:19
openstackgerritTobias Henkel proposed zuul/zuul master: Evaluate CODEOWNERS settings during canMerge check  https://review.opendev.org/64455707:21
*** jbadiapa has joined #openstack-infra07:22
*** pgaxatte has joined #openstack-infra07:23
*** dchen has quit IRC07:23
*** jbadiapa has quit IRC07:27
*** iurygregory has joined #openstack-infra07:30
*** xek has joined #openstack-infra07:36
*** slaweq has joined #openstack-infra07:41
*** lucasagomes has joined #openstack-infra07:43
*** aedc has quit IRC07:48
*** ralonsoh has joined #openstack-infra07:50
*** aedc has joined #openstack-infra07:54
*** slittle1 has quit IRC07:54
*** yamamoto has joined #openstack-infra07:57
*** ccamacho has joined #openstack-infra07:59
*** ykarel is now known as ykarel|lunch07:59
*** altlogbot_2 has joined #openstack-infra08:01
*** altlogbot_2 has quit IRC08:04
*** yolanda has quit IRC08:08
*** slittle1 has joined #openstack-infra08:08
*** yolanda has joined #openstack-infra08:09
openstackgerritJan Kubovy proposed zuul/zuul master: Overriding max. starting builds.  https://review.opendev.org/67046108:09
*** altlogbot_3 has joined #openstack-infra08:11
*** slittle1 has quit IRC08:12
*** altlogbot_3 has quit IRC08:16
*** altlogbot_0 has joined #openstack-infra08:17
*** tkajinam has quit IRC08:19
*** altlogbot_0 has quit IRC08:22
*** altlogbot_2 has joined #openstack-infra08:23
*** pkopec has joined #openstack-infra08:27
*** altlogbot_2 has quit IRC08:29
*** Fidde has joined #openstack-infra08:31
*** derekh has joined #openstack-infra08:32
*** rascasoft has quit IRC08:33
*** rascasoft has joined #openstack-infra08:34
*** Lucas_Gray has joined #openstack-infra08:40
openstackgerritTobias Henkel proposed zuul/zuul master: Evaluate CODEOWNERS settings during canMerge check  https://review.opendev.org/64455708:42
openstackgerritMerged opendev/irc-meetings master: Update networking OVN meeting  https://review.opendev.org/67037208:44
*** dosaboy has quit IRC08:47
*** aedc has quit IRC08:48
*** aluria has quit IRC08:48
*** ykarel|lunch is now known as ykarel08:48
*** altlogbot_3 has joined #openstack-infra08:53
*** rascasoft has quit IRC08:55
*** altlogbot_3 has quit IRC08:58
*** rascasoft has joined #openstack-infra08:58
*** irclogbot_3 has joined #openstack-infra08:59
*** dosaboy has joined #openstack-infra09:00
*** betherly has joined #openstack-infra09:00
*** dosaboy has quit IRC09:03
openstackgerritTobias Henkel proposed zuul/zuul master: Evaluate CODEOWNERS settings during canMerge check  https://review.opendev.org/64455709:03
*** aluria has joined #openstack-infra09:03
*** dosaboy has joined #openstack-infra09:04
*** irclogbot_3 has quit IRC09:04
*** betherly has quit IRC09:05
*** Lucas_Gray has quit IRC09:10
*** Lucas_Gray has joined #openstack-infra09:12
*** Goneri has quit IRC09:14
*** kjackal has quit IRC09:18
*** factor has joined #openstack-infra09:22
*** diablo_rojo has quit IRC09:25
*** Goneri has joined #openstack-infra09:27
*** gtema has quit IRC09:34
*** gtema has joined #openstack-infra09:34
*** Goneri has quit IRC09:39
*** gtema_ has joined #openstack-infra09:41
*** gtema has quit IRC09:43
*** yamamoto has quit IRC09:49
*** yamamoto has joined #openstack-infra09:51
*** yamamoto has quit IRC09:51
*** yamamoto has joined #openstack-infra09:52
*** yamamoto has quit IRC09:57
openstackgerritStephen Finucane proposed openstack/project-config master: Add shared 'oslo', 'oslo-independent' ACL files  https://review.opendev.org/67027009:58
openstackgerritStephen Finucane proposed openstack/project-config master: Update ACLs for moved doc projects  https://review.opendev.org/67026909:58
openstackgerritStephen Finucane proposed openstack/project-config master: Update gerritbot channels for moved doc projects  https://review.opendev.org/67048309:58
*** gtema_ has quit IRC10:02
*** gtema has joined #openstack-infra10:02
*** gtema has quit IRC10:03
*** gtema has joined #openstack-infra10:03
*** pfallenop has joined #openstack-infra10:09
*** kjackal has joined #openstack-infra10:15
*** pfallenop has quit IRC10:18
*** ociuhandu has joined #openstack-infra10:22
*** gtema has quit IRC10:30
*** gtema has joined #openstack-infra10:30
openstackgerritJan Kubovy proposed zuul/zuul master: Overriding max. starting builds.  https://review.opendev.org/67046110:31
*** irclogbot_3 has joined #openstack-infra10:31
*** yolanda has quit IRC10:31
*** yolanda has joined #openstack-infra10:32
*** Lucas_Gray has quit IRC10:35
openstackgerritTobias Henkel proposed zuul/zuul master: Annotate canMerge check with event id  https://review.opendev.org/67049410:35
*** irclogbot_3 has quit IRC10:38
*** irclogbot_2 has joined #openstack-infra10:41
*** irclogbot_2 has quit IRC10:44
*** dosaboy has quit IRC10:45
*** kjackal has quit IRC10:45
*** gtema has quit IRC10:47
*** gtema has joined #openstack-infra10:47
*** Goneri has joined #openstack-infra10:53
*** aluria has quit IRC10:56
*** yamamoto has joined #openstack-infra11:00
*** dosaboy has joined #openstack-infra11:02
*** altlogbot_2 has joined #openstack-infra11:03
*** yamamoto has quit IRC11:06
*** altlogbot_2 has quit IRC11:08
*** snierodz has quit IRC11:08
*** stephenfin has quit IRC11:08
*** tesseract has joined #openstack-infra11:08
*** stephenfin has joined #openstack-infra11:10
*** aluria has joined #openstack-infra11:11
*** altlogbot_0 has joined #openstack-infra11:12
*** altlogbot_0 has quit IRC11:16
*** kjackal has joined #openstack-infra11:20
openstackgerritTobias Henkel proposed zuul/zuul master: Evaluate CODEOWNERS settings during canMerge check  https://review.opendev.org/64455711:26
*** dosaboy has quit IRC11:27
*** dosaboy has joined #openstack-infra11:28
*** dosaboy has quit IRC11:28
openstackgerritJan Kubovy proposed zuul/zuul master: Overriding max. starting builds.  https://review.opendev.org/67046111:29
*** yamamoto has joined #openstack-infra11:35
*** yamamoto has quit IRC11:35
*** yamamoto has joined #openstack-infra11:36
*** apetrich has joined #openstack-infra11:38
*** gtema has quit IRC11:46
*** gtema has joined #openstack-infra11:46
*** aedc has joined #openstack-infra11:54
*** eharney has joined #openstack-infra11:54
*** udesale has quit IRC11:59
*** udesale has joined #openstack-infra12:00
*** altlogbot_0 has joined #openstack-infra12:07
*** altlogbot_0 has quit IRC12:08
openstackgerritMonty Taylor proposed zuul/zuul master: Use a requests session to simplify auth'd calls  https://review.opendev.org/67051112:16
openstackgerritMonty Taylor proposed zuul/zuul master: Use urllib.parse for manipulating client urls  https://review.opendev.org/67051212:16
*** Goneri has quit IRC12:21
*** goldyfruit has quit IRC12:22
*** electrofelix has joined #openstack-infra12:23
*** derekh has quit IRC12:26
openstackgerritMatthieu Huin proposed zuul/zuul master: Zuul CLI: allow access via REST  https://review.opendev.org/63631512:27
*** aedc has quit IRC12:27
*** Goneri has joined #openstack-infra12:29
*** ekultails has joined #openstack-infra12:30
openstackgerritMonty Taylor proposed zuul/zuul master: Use a requests session to simplify auth'd calls  https://review.opendev.org/67051112:33
openstackgerritSimon Westphahl proposed zuul/nodepool master: Don't pause static pool on single label quota  https://review.opendev.org/66737112:37
*** rlandy has joined #openstack-infra12:37
openstackgerritMatthieu Huin proposed zuul/zuul master: Add Authorization Rules configuration  https://review.opendev.org/63985512:41
mnaserinfra-root: i have cleaned up all stale volumes and also deletd all ERROR state instances in sjc112:41
mnaseri... hope nodepool doesnt get angry the ERROR vms disappeared beneath it12:41
openstackgerritTobias Henkel proposed zuul/zuul master: Evaluate CODEOWNERS settings during canMerge check  https://review.opendev.org/64455712:42
openstackgerritMatthieu Huin proposed zuul/zuul master: Web: plug the authorization engine  https://review.opendev.org/64088412:45
*** markvoelker has quit IRC12:45
openstackgerritMatthieu Huin proposed zuul/zuul master: Zuul Web: add /api/user/authorizations endpoint  https://review.opendev.org/64109912:45
openstackgerritMatthieu Huin proposed zuul/zuul master: authentication config: add optional token_expiry  https://review.opendev.org/64240812:45
*** viks___ has quit IRC12:46
Shrewsmnaser: nodepool is pretty resilient against such things. it anticipates pretty much anything disappearing on it12:51
*** aaronsheffield has joined #openstack-infra12:52
openstackgerritMatthieu Huin proposed zuul/zuul master: Web: plug the authorization engine  https://review.opendev.org/64088412:54
*** mriedem has joined #openstack-infra12:56
mnaserShrews: cool, thanks!12:57
*** tjgresha_nope has joined #openstack-infra12:59
*** ricolin has quit IRC13:00
*** aedc has joined #openstack-infra13:00
*** tjgresha has quit IRC13:02
*** derekh has joined #openstack-infra13:06
*** rfarr_ has quit IRC13:06
*** rfarr has joined #openstack-infra13:06
*** lseki has joined #openstack-infra13:10
*** rfarr_ has joined #openstack-infra13:10
*** aedc has quit IRC13:12
*** rfarr has quit IRC13:13
*** sthussey has joined #openstack-infra13:14
*** rfarr has joined #openstack-infra13:19
*** rfarr_ has quit IRC13:21
*** whoami-rajat has quit IRC13:25
*** whoami-rajat has joined #openstack-infra13:25
fungithanks for cleaning that up mnaser!13:26
AJaegerinfra-root, I just approved change https://review.opendev.org/#/c/670378/ to bring Fort Nebula CI cloud online.13:28
fungiAJaeger: thanks! i was about to do that too. will keep an eye on it13:30
*** sshnaidm|off has quit IRC13:30
openstackgerritMerged openstack/project-config master: Bringing Fort Nebula CI Cloud back online  https://review.opendev.org/67037813:32
donnydyey :) :)13:34
*** sshnaidm has joined #openstack-infra13:35
*** irclogbot_0 has joined #openstack-infra13:35
*** sshnaidm is now known as sshnaidm|off13:36
*** irclogbot_0 has quit IRC13:38
*** goldyfruit has joined #openstack-infra13:49
*** ykarel is now known as ykarel|afk13:54
*** ykarel|afk has quit IRC13:59
donnydlooks like the centos7 image is working great14:02
donnydAJaeger: can you check for job failures ?14:05
*** irclogbot_1 has joined #openstack-infra14:09
donnydIf all looks good I would like to take it up to 50 and then watch it for a while14:10
fungiwe'll probably need a bit of time for some longer-running classes of jobs to finish there and get indexed14:11
donnydthat sounds good to me14:11
*** FlorianFa has quit IRC14:11
fungiand then dig into the causes of failures (because there will assuredly be failures, we just hope they're due to buggy patches being tested)14:13
fungiquerying http://logstash.openstack.org/ for node_provider:fortnebula-regionone build-status:failure turns up log lines from some14:13
*** altlogbot_2 has joined #openstack-infra14:13
*** chandankumar is now known as raukadah14:14
fungihttp://logs.openstack.org/13/670513/2/check/nodejs10-npm-run-test/5b8313c/job-output.txt14:15
fungithere's one14:15
fungilooks like that was a chromium-based browser testset for horizon on ubuntu-bionic14:17
donnydThe error was some keystone issue14:18
donnydfrom what I can see14:18
fungiyeah, almost certainly not a provider-level problem14:18
donnydno really infra related, but I do agree we should give it some time and check to see if any of the infra bits are busted14:18
donnydI am assuming that storage speeds/IOPS will be an issue at scale, as the current backend can only do about 75K IOPS.14:20
donnydAt least in the testing I was able to get done14:20
fungiyeah, we usually end up having to tune our quota size relative to the host aggregate in situations where we're basically in a dedicated environment14:21
donnydSo there will be a maintenance window some time in the next few weeks to swap it out with an all nvme based one14:21
fungiwe quickly become our own noisy neighbor14:21
donnydWell this provider does only CI work, no general purpose stuff... so the backends will be tuned to the workload14:22
donnydusually it goes the other way around14:22
fungihttp://logs.openstack.org/56/670556/1/check/legacy-tempest-dsvm-networking-bgpvpn-bagpipe/528149d/14:22
fungithere's another failure14:22
fungii need to pop out to run a quick errand, but will brb14:22
*** ykarel|afk has joined #openstack-infra14:24
fungioh, that was a success14:24
donnydError when trying to get requirement for VCS system Command "git config --get-regexp remote\..*\.url" failed with error code 1 in /opt/stack/new/networking-bagpipe, falling back to uneditable format,Could not determine repository location of /opt/stack/new/networking-bagpipe14:24
fungithat's probably benign14:25
donnydWell it looks like it failed, but not infra related either14:25
fungimy logstash query above was incorrect14:25
donnydoh14:25
fungishould be build_status:failure14:25
fungi(_ not -)14:25
funginode_provider:fortnebula-regionone AND build_status:FAILURE14:26
*** lpetrut has joined #openstack-infra14:26
fungiokay, errand. brb14:26
*** jcoufal has joined #openstack-infra14:27
openstackgerritMerged openstack/ptgbot master: Display count of attendees in each room on web page  https://review.opendev.org/65850114:31
openstackgerritMerged openstack/ptgbot master: Use a badge to show check-ins in "now" display  https://review.opendev.org/65879514:31
*** liuyulong has joined #openstack-infra14:33
*** dpawlik has quit IRC14:34
*** bnemec is now known as beekneemech14:34
openstackgerritJeff Liu proposed zuul/zuul-operator master: Add Kubernetes Operator Functional Test Job  https://review.opendev.org/66802914:38
openstackgerritJeff Liu proposed zuul/zuul-operator master: [WIP] Verify Operator Pod Running  https://review.opendev.org/67039514:38
AJaegerconfig-core, please review these two small cleanups https://review.opendev.org/670452 and https://review.opendev.org/67034414:40
*** rfarr has quit IRC14:44
*** Goneri has quit IRC14:52
*** TheJulia is now known as needssleep14:52
*** rpittau is now known as elfosardo14:52
*** markvoelker has joined #openstack-infra14:53
*** markvoelker has quit IRC14:56
*** Goneri has joined #openstack-infra14:57
openstackgerritThierry Carrez proposed openstack/ptgbot master: Reset to OrderedDict on new day cleanup  https://review.opendev.org/67057714:57
openstackgerritThierry Carrez proposed openstack/ptgbot master: Clean up stale data presence on a #newday command  https://review.opendev.org/67057814:57
*** ociuhandu_ has joined #openstack-infra14:59
*** diablo_rojo has joined #openstack-infra15:00
*** ociuhandu has quit IRC15:02
*** ociuhandu_ has quit IRC15:03
*** gtema has quit IRC15:08
*** Goneri has quit IRC15:15
*** ykarel|afk is now known as ykarel|away15:16
openstackgerritJeff Liu proposed zuul/zuul-operator master: Remove Operator SDK dependency in Zuul Job  https://review.opendev.org/67058415:17
clarkbas a heads up I have followed up with kevinz about new linaro arm64 cloud region for CI resources15:18
clarkbits early yet so talking requirements and the like. Hopefully that gets us more arm test nodes :)15:19
clarkbnb03 needs the same cleanup as nb01 and nb02 so doing that now15:19
*** piotrowskim has quit IRC15:21
*** Fidde has quit IRC15:26
*** iurygregory is now known as skolt15:27
fungidonnyd: so far only 2 failed builds... 5b8313c872d344b6a423c11627a03f56 which we already looked at and 0f66c903686b4ec7b7ca32773e9a02af which is logged here: http://logs.openstack.org/56/670556/1/check/legacy-networking-bagpipe-dsvm-fullstack/0f66c90/15:27
clarkbfungi: that second one failed due to compile of ovs not working15:28
clarkbunlikely a provider issue15:28
fungii concur15:28
fungi/opt/stack/new/ovs/datapath/linux/nf_conntrack_reasm.c:79:31: error: ‘struct inet_frags’ has no member named ‘rnd’15:28
fungi  net_get_random_once(&nf_frags.rnd, sizeof(nf_frags.rnd));15:29
fungiso far so good15:29
*** Lucas_Gray has joined #openstack-infra15:30
fungijobs began running there about 13:45 so we're coming up on 2 hours with only 2 build failures neither of which look like they could be attributable to that provider15:30
fungiprobably safe to crank it up a notch whenever15:31
openstackgerritMerged openstack/project-config master: Remove unused docs jobs from dashboard  https://review.opendev.org/67045215:31
openstackgerritMerged openstack/project-config master: Update description of some docs jobs  https://review.opendev.org/67034415:31
clarkbshould probably check that donnyd is happy with things on the cloud side but ya I'd say go for it15:32
donnydclarkb: I am good. Want me to push up the next step?15:32
fungiright, i was still going to wait for an all-clear from him15:32
clarkbdonnyd: ya if you want to propose the next bump I think that would be good. We should remember to bump quotas too15:33
clarkb(if the quotas aren't already bumped)15:33
*** jcoufal has quit IRC15:34
openstackgerritDonny Davis proposed openstack/project-config master: Scaling FNCI to 40 instances  https://review.opendev.org/67058715:34
clarkbI'm going to find breakfast and then probably get a bike ride in but +2 on ^15:35
dmelladoAJaeger: ping re: opensuse repos down15:35
*** sdoran has joined #openstack-infra15:36
*** tdasilva has joined #openstack-infra15:36
sdoranHello everyone. 👋15:36
sdoranIs anyone seeing metadata downloads hanging for OpenSUSE 15?15:36
fungisdoran: in zuul jobs? have a link to an example?15:37
fungido the jobs eventually fail due to timeouts i guess?15:37
sdoranNo, it's in Ansible CI, so shippable.15:37
sdoranWe cap jobs at 45 minutes, so they are getting killed.15:37
*** mattw4 has joined #openstack-infra15:37
sdoranBut if I just run `zypper install udev` in a test container, it hangs.15:38
sdoranThis in the URL that it gets stuck on: http://download.opensuse.org/distribution/leap/15.0/repo/non-oss/repodata/773c107fe9e932054ad44f31655f245faefbd3172657429e363acf7917e125f0-primary.xml.gz15:38
fungiahh, well i don't know what opensuse 15 "metadata downloads" are, but if they're part of the zypper package repositories we cache those locally in our node providers15:38
sdoranSome URLs from the page download ok, but others do not.15:38
*** kjackal has quit IRC15:39
fungiyeah, we try to prevent our ci jobs from accessing externally-served distro packages15:39
sdoranI'm starting to think we need to do the same. :)15:39
*** diablo_rojo has quit IRC15:39
fungifor opensuse we rsync update a mirror in afs every 4 hours and then atomically release that afs volume if the rsync succeeds15:39
*** diablo_rojo has joined #openstack-infra15:40
fungiand stick afs client caches with apache frontends in each of our node providers for zuul/nodepool15:40
sdoranThat's nice.15:40
fungiand configure our node test images to look at those for their packages15:40
dmelladohey sdoran15:40
dmelladojust pinged AJaeger15:41
sdoran👍15:41
dmelladobut I'm afraid he might be off as it's a little bit late in EMEA and Friday15:43
dmelladothere's another channel we might try15:44
dmellado#opensuse-buildservice15:44
fungidirk is also around sometimes and knows who to reach out to15:45
*** jcoufal has joined #openstack-infra15:45
*** lucasagomes has quit IRC15:47
sdoran@fungi Thanks for the help and suggestions.15:49
AJaegerdmellado: better talk with cmurphy and dirk about openSUSE repos15:51
dirkdmellado: AJaeger: there is a network problem (download.o.org is down)15:52
dirkI thought the openstack ci is not affected because it uses a mirror?15:52
dmelladoAJaeger: dirk thanks!15:52
dirkwhy is it not using the mirror in this case?15:52
AJaegerthanks, dirk15:53
dmelladoyeah, exactly, it seems that ansible doesn't use a mirror but rather queries it directly15:53
*** lpetrut has quit IRC15:53
dirkah, okay15:53
dirkyeah, well, to be honest the main download site shouldn't be down in the first place15:54
mriedemclarkb: just remembered and posted a patch for this issue i found awhile ago https://review.opendev.org/67059115:56
mriedemfor some odd reason, cirros guest shutdown is not happening within 60 seocnds15:56
mriedem*seconds15:56
mriedemnotes and a guest console log are in the related bug,15:56
mriedembut looks like maybe the guest is held up on shutdown waiting for a response from the metadata api15:57
*** ykarel|away has quit IRC15:57
mriedemthis means that tempest tests that do server stop, shelve, rescue and rebuild could be taking up to 60 seconds just to stop the guest15:57
fungidirk: it's not a problem for our environment. sdoran was just asking because the "shippable" ci system for ansible on github was having trouble on opensuse 15 and so was asking if we were encountering similar issues (we're not afaik because we have our own mirrors)15:58
openstackgerritJeff Liu proposed zuul/zuul-operator master: Remove Operator SDK dependency in Zuul Job  https://review.opendev.org/67058415:58
dmelladothanks in any case dirk!15:59
sdoranYes, thanks for the help.16:00
sdoranJust needed to ask some other folks that I know have a pretty busy CI that is hitting OpenSUSE mirrors.16:00
*** adrianreza_ has joined #openstack-infra16:01
fungiturns out our ci system is busy enough we avoid using official distro mirrors ;)16:01
*** gyee has joined #openstack-infra16:01
openstackgerritMerged openstack/project-config master: Scaling FNCI to 40 instances  https://review.opendev.org/67058716:02
fungiit's much faster and more stable not having to drag packages halfway across the internet on every build16:02
*** ijw has joined #openstack-infra16:04
*** ijw_ has joined #openstack-infra16:05
*** Lucas_Gray has quit IRC16:06
*** pgaxatte has quit IRC16:07
*** ykarel|away has joined #openstack-infra16:09
*** ijw has quit IRC16:09
*** pkopec has quit IRC16:10
*** Lucas_Gray has joined #openstack-infra16:12
donnydclarkb: looks like something is still wrong in fedora2816:22
donnydcan't reach the instance16:22
fungithe increased number of nodes started to be used for jobs as of ~16:15z16:23
*** mriedem has quit IRC16:25
*** aluria has quit IRC16:26
*** kjackal has joined #openstack-infra16:26
donnydfungi: yea, def something wrong with fedora2816:27
fungigot it. for those we're probably just logging boot failures if they come up unreachable16:27
*** elfosardo is now known as rpittau|afk16:27
fungiso they're not impacting any jobs, but they are wasting resources out of the quota16:28
donnydwell I can easily fix that16:30
funginot to mention the i/o bandwidth consumed by the boot/delete churn16:31
donnydthere is almost none on my end...16:31
fungioh, nice16:31
donnydIf you look at time to ready its pretty clear, all the images are cached in ram16:32
fungithat helps16:32
donnydI need to move my image backend to something much faster though... because you can clearly see when they aren't16:33
donnydjust gonna wait for the nvme storage to get here, and I should be able to put all of it on that16:33
donnydtrying to get my in-use to a solid 40 and then i need to watch the heat16:34
donnydits my only crutch in having a no A/C based system16:35
donnydSo something interesting I didn't really expect, when the jobs are running I was thinking power would go up exponentially16:36
donnydfor right now... it doesn't seem to have budged at all16:36
donnydbut maybe its because the jobs are just getting spun up and aren't really doing anything yet16:37
fungifor some years i had a full 7' rack full of inefficient/power-hungry antiques in my home lab, and so ducted a freestanding auxiliary air conditioner through them and exhausted it out a spacer in the window16:39
fungialso aluminum-foiled the windows to cut down on heat coming in from the sun16:40
funginot like a proper crac, but it did the job16:40
*** ginopc has quit IRC16:42
fungilooking at http://zuul.opendev.org/t/openstack/nodes there are some in-use for >15 minutes already16:45
fungii mean in addition to the handful which are attributable to the earlier, lower quota16:46
*** derekh has quit IRC16:48
*** kjackal has quit IRC16:49
*** ijw_ has quit IRC16:51
*** rkukura_ has joined #openstack-infra16:52
*** gtema has joined #openstack-infra16:52
*** rkukura has quit IRC16:54
*** rkukura_ is now known as rkukura16:54
Shrewsfungi: very MacGyver of you16:58
clarkbdonnyd: fungi oh you know what I wonder if that image got rebuilt16:58
clarkbdonnyd: its possible that it didn't?16:58
*** udesale has quit IRC16:59
*** betherly has joined #openstack-infra17:00
*** psachin has quit IRC17:03
*** ijw has joined #openstack-infra17:03
*** jtomasek has quit IRC17:04
*** betherly has quit IRC17:05
*** adriancz has quit IRC17:07
donnydso the difference between no workload and 40% is 300 watts17:07
clarkbya the fedora-28 image is old17:09
clarkbfrom the 8th17:09
* clarkb loosk into why that one isn't building17:09
*** jcoufal has quit IRC17:09
corvus7.5 watts per instance17:09
clarkbhttp://paste.openstack.org/show/754350/ is why fedora-28 isn't updating17:10
clarkbfedora-28 is EOL iirc so odd that a package/service would disappear?17:11
corvusif that holds across providers (probably not, but maybe close), the whole cluster is using 6kW.17:11
clarkbpabelanger: http://paste.openstack.org/show/754350/ any idea why that might be happening?17:12
clarkbpabelanger: for more context that is failed dib build of fedora-28 image17:12
mordredinfra-root: The fine folks at the MOC are going to start giving us some capacity. (/me waves at knikolla) I've submitted some forms on the RH side to get accounts spun up, and will be following up with the appropriate config patches once we've got accounts and whatnot17:13
fungiat that rate we're gonna need a lot more nodes to reach 1.21 gigawatts17:14
corvusmordred, knikolla: \o/ neat!17:14
fungiooh, that's awesome!17:14
*** witek has quit IRC17:14
clarkbcool17:14
*** gtema has quit IRC17:15
pabelangerclarkb: looks like fall out of getting fedora-29 to build properly17:15
clarkbdonnyd: not sure if you've seen http://grafana.openstack.org/d/3Bwpi5SZk/nodepool-fortnebula?orgId=1 but that tries to track things for you from our side too17:16
clarkbpabelanger: well we had a successful build on the 8th though17:16
pabelangerthere was a change to DIB to fix fedora-29, but maybe it broke fedora-28?17:16
clarkbI guess I should look at those logs from the 8th and see if it ran that same code17:16
pabelangeryah17:16
clarkbpabelanger: recently?17:16
pabelangermaybe 2 months ago17:16
pabelangerlet me find change17:16
AJaegerknikolla: Great, thanks!17:16
clarkbpabelanger: thanks17:16
pabelangerclarkb: https://review.opendev.org/657126/ might be related17:17
clarkbhrm all of our build files are newer than the 8th, we rotate them more aggressively than I thought17:17
clarkbs/files/logs/17:17
AJaegermordred: what is MOC?17:17
clarkbpabelanger: oh I bet that coincides with a release of dib17:17
clarkbpabelanger: since we consume it from pypi not from source. Thanks for the pointers I should be able to track it down now probably17:18
pabelangerclarkb: there is a comment about fedora-28 too breaking17:18
pabelangerso, guess our testing didn't work well17:18
clarkbok I think I get what happened17:19
*** ralonsoh has quit IRC17:19
clarkbwe probably didn't care about 28 because its eol17:19
clarkbbut tripleo is still using it so whoops17:19
clarkbI'll get a fix up17:19
corvusAJaeger: https://www.bu.edu/hic/research/highlighted-sponsored-projects/massachusetts-open-cloud/17:21
AJaegercorvus: thanks - ok, remember hearing about them but MOC didn't ring a bell17:21
corvusmaybe https://massopen.cloud/about/ is more relevant now17:22
openstackgerritClark Boylan proposed openstack/diskimage-builder master: Only enable dbus-daemon on fedora-29  https://review.opendev.org/67060617:23
clarkbpabelanger: donnyd fungi ^ I think that will fix fedora-28 builds but we'll have to make a dib release too17:23
donnydIt would seem everything else is humming along quite well. Although I haven't seen a gentoo based build come in yet17:24
donnydheat seems to be under control, and power usage is well within expected ranges17:25
clarkbcool17:26
*** Lucas_Gray has quit IRC17:26
donnydthe whole thing is using about 20 amps / 230 volts or about 4600 watts. I have a few extra pieces of equipment that need to be pulled17:26
donnydmaybe scale up a bit more?17:26
clarkbdonnyd: I think we are happy to scale up as high as you are willing :) also it should quiet way down over the weekend17:27
clarkb(since msot of the jobs are run due to demand)17:28
fungiyeah, weekend load may not present a useful load test unless we also scale down our other providers to place more pressure on it17:28
fungiclarkb: 670606 also needs a new dib release before we can take advantage, right?17:29
*** nicolasbock has joined #openstack-infra17:29
clarkbfungi: yes17:29
fungijust making sure17:30
*** rkukura has quit IRC17:31
openstackgerritMerged zuul/zuul-operator master: Add Kubernetes Operator Functional Test Job  https://review.opendev.org/66802917:37
openstackgerritMerged zuul/zuul-operator master: Remove Operator SDK dependency in Zuul Job  https://review.opendev.org/67058417:37
openstackgerritDonny Davis proposed openstack/project-config master: Moving the workload on FNCI to 60%  https://review.opendev.org/67060917:39
donnydI think we can leave it at 60% for a while. I am hopeful over the weekend I can at least get 60 jobs.17:40
*** igordc has joined #openstack-infra17:43
AJaegerdonnyd: for a few more hours for sure - and probably for our periodic runs starting at 6:00 UTC17:48
* clarkb cleans up the node held to debug the glean issues on centos17:50
clarkbcorvus: do you still need your held nodes that appear to be held for debugging gitea things?17:51
clarkbI can clean them up too if not17:51
openstackgerritJames E. Blair proposed zuul/zuul master: Build layout of non-live items with config updates  https://review.opendev.org/67033517:53
corvusclarkb: nope, sorry, i must have missed the timeout setting17:53
corvuswe should set that in nodepool17:53
*** xek has quit IRC17:53
clarkbcorvus: as a default you mean?17:53
corvusyep.  i'll work up a change17:53
clarkbin any case cleaning those up now17:53
*** xek has joined #openstack-infra17:53
corvusi'm writing the change as penance for forgetting the option17:54
corvushrm, that option doesn't behave the way i hoped.  it's a max, not a default, so we wouldn't be able to override it17:55
clarkbwe might be able to change the default on the zuul side?17:56
corvusi don't think there's a setting for that17:57
Shrewscorvus: change --node-hold-expiration default17:57
Shrews1 line change17:57
corvusShrews: that would change it for everyone17:57
corvusi just want to establish an opendev default of 1 day17:57
corvus(but i want us to be able to set it to 1 week or indefinite if necessary for an individual node)17:58
corvusso i think zuul or nodepool needs to grow a new option for a site-customizable default value17:58
Shrewscant you set max-hold-age in cfg?17:59
corvusShrews: that's a *max*.  if i set that to 1 day, we can not override it17:59
corvusmost autoholds we need for only a few hours, so 1 day would be okay.  sometimes we hold on to them for a week while we work with providers to fix deeper issues17:59
Shrewsoh, theres a max() call somewhere then18:00
corvusShrews: i think there's a dedicated cleanup worker that looks for holds > max18:00
corvusso even if the znode has an expration set to 2d if the max is 1d it will still delete that node18:01
*** betherly has joined #openstack-infra18:01
corvusoh, maybe that one cleans up all holds18:02
corvuseither way:18:02
corvus            max_uptime = min(expiration, self._nodepool.config.max_hold_age)18:02
*** betherly has quit IRC18:06
openstackgerritMerged openstack/project-config master: Moving the workload on FNCI to 60%  https://review.opendev.org/67060918:15
openstackgerritJames E. Blair proposed zuul/nodepool master: Add functional jobs to gate  https://review.opendev.org/67061218:17
*** electrofelix has quit IRC18:18
*** betherly has joined #openstack-infra18:21
openstackgerritBrian Haley proposed openstack/devstack-gate master: Support an IPv6 underlay network  https://review.opendev.org/34304118:23
*** jeremy_houser has joined #openstack-infra18:24
jeremy_houserCan anyone assist me in getting final reviews for https://review.opendev.org/#/c/670159/ ? Only need workflow and maybe one more +218:24
jeremy_houserI am attempting to merge first step of new repo for my existing tempest plugin18:24
clarkbhaleyb: re ^ do we really want to add new features to devstack-gate? we should be adding that to the native zuul roles for multinode networking18:24
clarkbhaleyb: If there is an immediate need I guess its fine, but we are (slowly) trying to get away from depending on devstack-gate for stuff18:24
haleybclarkb: i was just going through my old reviews and re-basing18:25
clarkbhaleyb: oh18:25
haleybis there some other place we do this?  i.e. calculate mtu to send to instances?  that review is only like 3 years old :-o18:25
*** betherly has quit IRC18:26
clarkbjeremy_houser: is there existing code you need to import at the same time? if so you should set an upstream. If not I can approve it18:26
clarkbhaleyb: ya let me find a link for the zuul native stuff18:26
clarkbhaleyb: https://opendev.org/zuul/zuul-jobs/src/branch/master/roles/multi-node-bridge is the ansible role. https://opendev.org/zuul/zuul-jobs/src/branch/master/roles/multi-node-bridge/tasks/common.yaml#L78-L113 is the bit that sets the mtu value18:27
jeremy_houserNo, Ive decided against that method. Id rather commit my code from workstation as its already in an opendev repo and that seemed more like it was for something coming from github18:28
clarkbjeremy_houser: the upstream can be any publicly accessible git repo (does not need to be github)18:28
clarkbjeremy_houser: ok just asking beause we don't like to have to do force pushes for people after the facty18:28
clarkbjeremy_houser: I'll go ahead and approve it then if you are ready18:29
haleybclarkb: hardcoded 50 byte overhead for vxlan :(  i'll add looking at that to my pile18:29
fungii often just git init on a personal webserver, git push into that and then that's a clonable repo which can be miported18:29
jeremy_houserI was just going to commit to the repo as normal, would it not work just fine?18:29
clarkbjeremy_houser: that will work just fine. Just double checking :)18:29
jeremy_houserfantastic, then yes, please approve when ready18:29
fungijeremy_houser: the only reason to import would be if you already had a bunch of commits locally and didn't want to have to test and review them all individually when bootstrapping the project18:30
jeremy_houserah no, that wont be an issue. Thank you for the information.18:30
fungiby default you'll end up with a mostly empty repo which only contains a .gitreview file, so you can clone from that, commit whatever changes you want and git review normally18:32
jeremy_houserthat's how I thought it would work. Fantastic.18:33
*** ijw has quit IRC18:35
openstackgerritMerged openstack/project-config master: New repo for ranger-tempest-plugin  https://review.opendev.org/67015918:37
clarkbjeremy_houser: ^ thats in so now just have to wait for the next ansible + puppet pulse18:41
clarkbusually about 30-45 minutes18:41
*** betherly has joined #openstack-infra18:42
*** rascasoft has quit IRC18:44
*** liuyulong has quit IRC18:44
*** betherly has quit IRC18:46
*** rascasoft has joined #openstack-infra18:47
*** ccamacho has quit IRC18:48
AJaegerdonnyd: quote for your cloud is up to 60 - but seems are queue is shrinking, so you might not get a full load today ;(18:48
*** irclogbot_1 has quit IRC18:49
*** edmondsw_ has quit IRC18:49
jeremy_houserso if I wanted to set up my tempest-plugin to gate my project, would I set that up in the ranger .zuul.yaml or the tempest-plugin .zuul.yaml?18:50
AJaegerjeremy_houser: see how other repos do it ;)18:50
AJaegerjeremy_houser: my suggestion: One job defined in your tempest-plugin repo that gates changes to the plugin, the same job is run in the project as well to gate changes of the code18:50
AJaegerjeremy_houser: keep in mind we have global namespace for jobs, so you define once and can use everywhere, see also https://docs.openstack.org/infra/manual/drivers.html#consistent-naming-for-jobs-with-zuul-v318:51
*** irclogbot_3 has joined #openstack-infra18:52
jeremy_houserI apologize, Ive been doing this for six months but I'm trying to whip my team into modernizing their stuff, so I'm going headfirst into everything18:52
donnydoh booooo AJaeger18:53
donnydthat makes me sad18:53
*** michael-beaver has joined #openstack-infra18:56
fungidonnyd: you could always ask dansmith to rebase a 70-commit-deep series of nova changes ;)18:57
* dansmith pulls the rip cord on his chainsaw18:57
AJaegerdonnyd: 56 in use ;)19:05
donnydOk, I will watch it for a while and see if any issues come up. If you see a patch come in for the last 40% its because it looks good to go on my end19:09
*** slaweq has quit IRC19:17
*** tomaw has quit IRC19:23
*** tomaw has joined #openstack-infra19:27
clarkbhttp://logs.openstack.org/91/670591/1/check/tempest-full/73885b4/ timed out on fn so we may want to hold where we are now and monitor things19:33
*** tesseract has quit IRC19:33
clarkb(there is a very real chance we are our own noisy neighbor leading to that timeout)19:33
*** slaweq has joined #openstack-infra19:34
fungiyeah, most frequent problem in these scenarios is we're hitting an aggregate bottleneck somewhere in the environment (cpu, ram, disk i/o, network...)19:38
fungiand that starts to slow down a significant percentage of the jobs which run there19:38
donnydI will start trying to run down the issue19:38
*** skolt has quit IRC19:39
fungisome of our longer-running jobs like devstack (so that one which timed out) do some performance tracking in the job as well19:39
donnydI see my instance launch time has doubled, so I am thinking it may be a storage issue19:39
donnydany pointers on which direction to turn in the investigation19:40
fungiwhat's the storage backend for the guest filesystems?19:40
donnydNFS19:40
donnydI have the storage scaled down to barebones atm while I measure how much each single node can handle19:41
clarkbif mnaser or logan- are around they may have thoughts (though they are all ceph based iirc)19:42
donnydI was using ceph when we started, but my cluster isn't big enough to get the performance I was looking for19:42
fungiif it's nfs i'd look at your bandwidth utilization across your nfs vlan/interfaces19:43
mnaseryeah nfs is going to be pretty rough19:43
fungisee if there are bottlenecks on the network between it and the guests before looking at bottlenecks between the nfs servers and their disks19:44
mnasermainly because you are doing block storage on top of file storage constructs19:44
donnydhttps://www.irccloud.com/pastebin/ppmF4hPR/19:44
fungiyeah, after network bandwidth between the compute hosts and the nfs servers i'd look at cpu utilization on the nfs servers19:44
donnydGonna need to bring more of the storages back online19:44
mnaserdonnyd: is this cloud for other pruposes or is it mainly for openstack only?19:45
donnydWell the network for each storage server is two 40G links19:45
mnaseropenstack-infra ci workloads rather19:45
*** whoami-rajat has quit IRC19:45
donnydyea, this is all it does19:45
mnaseri'd throw local disks and call it a day personally ;)19:45
donnydits not built to be general purpose19:46
donnydcan't... I have blades19:46
donnydand they suck at the local storages19:46
mnaseri mean you could totally raid-0 two drives only and put the os and /var/lib/nova/instances on it19:46
mnaserinfra won't be sad if a bunch of machines disappeared off earth19:46
fungiiscsi will probably buy you marginally better performance than virtual block on nfs19:47
mnaserhttps://www.youtube.com/watch?v=4JWgmv92fQk19:47
mnaserand yeah, iscsi will probably yield some better results too19:47
donnydWell that was where I started and I have weird issues with the containers and using iscsi for nova19:48
*** bhavikdbavishi has quit IRC19:48
donnydlike it just stops working weird19:48
clarkbya the iscsi stuffin the kernel doesn't namespace19:48
mnaseroh with kolla?19:48
donnydtripleo so yea kolla containers19:48
clarkbapparently there si a userspace iscsi driver in libvirt now but unsure of how that performs19:48
fungior you add the necessary caps to get iscsi working because you're using containers for convenience not for security separation19:49
mnaserCAP_ADMIN all the things19:49
donnydI only have 1 of 6 storage nodes online atm... so I could quite easily spread the load a little better19:49
donnydand +150K on the jeremy clarkson video mnaser19:50
*** slaweq has quit IRC19:51
clarkboh ya distributing the load regardless of the system underneath is likely ot help19:51
donnydthe storage servers are the main heat generators at least in my shabby little DC19:52
donnydHowever there is good news, 1. I can just turn on more and 2. all NVME storage is enroute19:53
clarkbzoom zoom19:53
logan-limestone uses imagebackend (local storage) for the nodepool hvs19:54
fungiyeah, sounds like the blades for fortnebula aren't so robust in the local storage department19:55
clarkbspeaking of fungi logan- I don't see any gaps in the limestone mirror cacti graphs for the last couple weeks ish19:55
logan-dual e5-2650v2, 128gb ram, 2x 512GB samsung 850 pro per HV. i/o seems to be the first bottleneck they hit19:55
clarkbshould we try turning on that cloud again?19:55
fungii'm game, no idea if logan- has had any luck hunting that down19:56
logan-yeah I found stacktraces from the nic driver on all 3 of the control nodes. i don't know if it is a kernel regression or some traffic triggering a bug in the igb driver19:56
logan-i don't think it is bad hardware since it happened on all 3 nodes19:56
clarkblogan-: huh, do you want to reboot to latest kernel versions (and drivers) before we turn it back on again?19:56
logan-i updated kernel and rebooted (igb reports the same version)19:56
clarkb(assuming they aren't already up to date)19:56
clarkbway ahead of me I see19:56
clarkbin that case ya maybe we try it again since cacti makes it look stabler19:57
*** ociuhandu has joined #openstack-infra19:57
logan-I have been running jobs on it and have not been able to repro it.. so I guess yeah we can turn it back on and see if it happens again. in other envs I have had issues with ubuntu's 10gig version of this driver (ixgbe) so I usually dkms it from upstream. thats my next thought if we run into more issues19:57
*** ociuhandu has quit IRC19:58
*** slaweq has joined #openstack-infra19:59
donnydSo it should look more like this20:01
openstackgerritClark Boylan proposed openstack/project-config master: Reenable limestone cloud region in nodepool  https://review.opendev.org/67063020:01
clarkbfungi: logan- ^ fyi20:01
donnydhttps://www.irccloud.com/pastebin/n7V2JxgE/20:01
logan-thanks clarkb. i'll keep an eye on it once it merges20:01
clarkbdonnyd: thats looks much better20:01
donnydyea, so I am thinking I may need to spread the load a little better... ;)20:02
*** betherly has joined #openstack-infra20:02
fungiheh20:03
*** slaweq has quit IRC20:03
clarkblike butter20:04
fungithat's roughly theoretical throughput for a 1gbps network link20:04
*** ykarel|away has quit IRC20:04
openstackgerritCorey Bryant proposed openstack/hacking master: Add Python 3 Train unit tests  https://review.opendev.org/67063220:05
openstackgerritCorey Bryant proposed openstack/os-performance-tools master: Add Python 3 Train unit tests  https://review.opendev.org/67063520:07
*** betherly has quit IRC20:07
openstackgerritCorey Bryant proposed openstack/os-testr master: Add Python 3 Train unit tests  https://review.opendev.org/67063620:07
*** davidsha has joined #openstack-infra20:07
*** ociuhandu has joined #openstack-infra20:08
donnydhttps://www.irccloud.com/pastebin/daa1oXyI/20:08
donnydThe first one was with 4k blocks, thats what I am testing because the DB part seemed to be the slowest in the beginning20:09
donnydso I get pretty much all of a single 10G link in disk performance20:11
clarkbis that against the nfs server currently being used?20:11
donnydWell its got a couple instances on it now, but not heavily loaded at all20:12
clarkbah20:12
openstackgerritMerged openstack/diskimage-builder master: Only enable dbus-daemon on fedora-29  https://review.opendev.org/67060620:13
*** slaweq has joined #openstack-infra20:15
clarkbso maybe monday morning I'll cut a dib release? I am wary of doing it before the weekend since I'm not super familiar with all the chagnes that have gone in since the last one20:15
clarkbI'll start reviewing the git log now20:16
clarkblooks like the big changes are the two I've made recently then johnsom added a flag to select a different ubuntu kernel in ubuntu-minimal and prometheanfire added gnupg2 to debian default package list for apt and there is a fix for rhel8 python stuff20:18
clarkboverall not too scary20:18
clarkb(much better than I was worried about)20:18
johnsomWe would love to see a DIB release....20:18
clarkbjohnsom: us too :) I jsut know I'm not going to be able to write fixes or pin dib to previous relase over the weekend and the rtt on getting images built and in clouds is long :/20:19
clarkbI can do it first thing monday20:19
johnsomYeah, Monday is sooner then RC. grin20:20
*** slaweq has quit IRC20:20
openstackgerritMerged openstack/project-config master: Reenable limestone cloud region in nodepool  https://review.opendev.org/67063020:21
*** betherly has joined #openstack-infra20:23
*** betherly has quit IRC20:27
*** davidsha has quit IRC20:30
*** ociuhandu has quit IRC20:31
*** slaweq has joined #openstack-infra20:31
*** slaweq has quit IRC20:35
*** goldyfruit has quit IRC20:35
donnydpretty sure all these jobs are about to hit their timeouts20:36
openstackgerritJames E. Blair proposed zuul/zuul master: Build layout of non-live items with config updates  https://review.opendev.org/67033520:41
donnydI also randomly get neutron timeouts in the log20:44
donnyd('Connection aborted.', BadStatusLine("''",))20:44
donnydtried bumping up the workers for the neutron server, but doesn't seem to help20:44
openstackgerritJames E. Blair proposed zuul/zuul master: Build layout of non-live items with config updates  https://review.opendev.org/67033520:45
fungiyeah, BadStatusLine(empty) is thieves cant for the requests library. in the common tongue it means it never got an answer before it gave up waiting20:49
clarkbdonnyd: oh one trick I should mention is if the quota is reduced nodepool will honor that so you can actually manage things that way if you want to make changes or reduce load etc (just set max instances quota to a smaller number)20:51
clarkb(I say that but I'm not sure we've every tested it for that use case)20:51
donnydI have enough to disable hypervisors, wait for the load to go to zero and then migrate the backend20:52
donnydbut that is a pretty good tip for when the new storage gets here20:53
donnydbecause it would be faster to scale to zero and then swap it out without code changes20:54
*** jeremy_houser has quit IRC20:56
fungiyeah, basically the max-servers in nodepool these days is a way for us to tell it to use less than its actual in-provider quotas would allow20:58
fungibut nodepool still tracks openstack api reported quotas and tries not to exceed them20:59
*** pcaruana has quit IRC20:59
donnydI have two more storage servers back online, so I am just waiting to swap them in and then see if the result is good21:01
*** betherly has joined #openstack-infra21:04
*** weifan has joined #openstack-infra21:07
*** betherly has quit IRC21:09
openstackgerritJames E. Blair proposed zuul/zuul master: Handle existing broken config in job updates  https://review.opendev.org/67066621:17
*** weifan has quit IRC21:18
*** weifan has joined #openstack-infra21:22
*** dpawlik has joined #openstack-infra21:25
clarkbhttps://blog.cloudflare.com/details-of-the-cloudflare-outage-on-july-2-2019/ is an interesting read21:27
* clarkb looks at zuul change now that I'm through ^21:27
*** ijw has joined #openstack-infra21:31
*** bgmccollum has quit IRC21:31
*** ekultails has quit IRC21:32
*** bgmccollum has joined #openstack-infra21:33
*** dpawlik has quit IRC21:35
openstackgerritJames E. Blair proposed zuul/zuul master: Handle existing broken config in job updates  https://review.opendev.org/67066621:37
*** weifan has quit IRC21:38
*** weifan has joined #openstack-infra21:40
*** weifan has quit IRC21:41
*** weifan has joined #openstack-infra21:42
donnydhttps://www.irccloud.com/pastebin/0uWOPRNm/21:43
donnydSo this is what a node looks like when the storage is loaded21:44
donnydstill some left over.. much better than 876/24621:44
*** weifan has quit IRC21:47
*** ianychoi has quit IRC21:54
*** rlandy has quit IRC21:54
*** guimaluf has quit IRC21:57
donnydI'm seeing quite a lot of errors popping up in logstash, not sure if that is on my end or the job21:58
clarkbdonnyd: http://logs.openstack.org/66/670666/2/check/zuul-tox-remote/6959fc6/job-output.txt#_2019-07-12_21_56_21_38835322:00
clarkbif your changes can result in read only filesystems it could be related22:00
donnydis there any way to tell what host that was running on?22:01
clarkbI think we record that, let me look22:03
logan-yep, it is the 'build_hostid' field in logstash22:03
clarkb2019-07-12 21:38:21,006 DEBUG nodepool.NodeLauncher: [node: 0008885215] Node 0008885215 is running [region: regionOne, az: nova, ip: 2001:470:e045:1:f816:3eff:fe70:5c43 ipv4: , ipv6: 2001:470:e045:1:f816:3eff:fe70:5c43, hostid: d8db5efa427a20886209c6207822af9e0c218fda9b7c05e06ab4a546]22:03
clarkbfound it in nodepool too22:03
logan-yep also stored in http://logs.openstack.org/66/670666/2/check/zuul-tox-remote/6959fc6/zuul-info/inventory.yaml22:04
logan-actually getting a hostname out of that is a little bit interesting though. looking for my notes on that22:04
*** xek has quit IRC22:05
logan-http://paste.openstack.org/raw/754361/22:06
donnydhrm22:08
donnydhttps://www.irccloud.com/pastebin/RbjLJa03/22:08
logan-you'll need to replace the project ID with whatever project ID the nodepool tenant is on22:09
donnydoh... ic22:09
donnydyea, so that makes more sense... It failed on the hypervisor I just swapped out the storage i22:10
donnydyea, so that makes more sense... It failed on the hypervisor I just swapped out the storage in22:10
openstackgerritJames E. Blair proposed zuul/zuul master: Add "supercedes" pipeline option  https://review.opendev.org/67067022:13
donnydnot sure why it did that, but at least I know where to start22:13
*** betherly has joined #openstack-infra22:17
*** betherly has quit IRC22:21
openstackgerritJames E. Blair proposed zuul/zuul master: Add "supercedes" pipeline option  https://review.opendev.org/67067022:24
*** slaweq has joined #openstack-infra22:33
*** slaweq has quit IRC22:38
*** michael-beaver has quit IRC22:57
*** ociuhandu has joined #openstack-infra23:01
*** mattw4 has quit IRC23:01
*** sthussey has quit IRC23:04
*** ociuhandu has quit IRC23:05
*** betherly has joined #openstack-infra23:19
*** tosky has quit IRC23:20
*** nicolasbock has quit IRC23:22
*** weifan has joined #openstack-infra23:23
*** betherly has quit IRC23:24
*** weifan has quit IRC23:28
*** eharney has quit IRC23:32
*** betherly has joined #openstack-infra23:39
*** bobh has joined #openstack-infra23:41
*** betherly has quit IRC23:44

Generated by irclog2html.py 2.15.3 by Marius Gedminas - find it at mg.pov.lt!