Friday, 2018-11-09

*** slaweq has joined #openstack-nova00:11
*** mvkr has joined #openstack-nova00:12
*** slaweq has quit IRC00:16
*** eandersson has joined #openstack-nova00:16
*** gyee has quit IRC00:16
*** hshiina has joined #openstack-nova00:23
*** pcaruana has quit IRC00:25
*** hshiina has quit IRC00:27
*** hshiina has joined #openstack-nova00:28
*** brinzhang has joined #openstack-nova00:37
*** tetsuro has joined #openstack-nova00:40
*** k_mouza has quit IRC00:46
*** Dinesh_Bhor has joined #openstack-nova01:03
*** slaweq has joined #openstack-nova01:11
*** erlon_ has quit IRC01:15
*** slaweq has quit IRC01:16
*** Dinesh_Bhor has quit IRC01:25
*** Dinesh_Bhor has joined #openstack-nova01:31
openstackgerritTakashi NATSUME proposed openstack/nova master: Add description of custom resource classes  https://review.openstack.org/61672101:32
openstackgerritTakashi NATSUME proposed openstack/nova master: Fix a help string in nova-manage  https://review.openstack.org/61672301:50
openstackgerritMerged openstack/nova master: Harden placement init under wsgi  https://review.openstack.org/61003401:53
*** hamzy has joined #openstack-nova02:15
*** hongbin has joined #openstack-nova02:42
*** mrsoul has quit IRC02:46
*** eharney has quit IRC02:54
openstackgerritMerged openstack/nova master: Use SleepFixture instead of mocking _ThreadingEvent.wait  https://review.openstack.org/61572403:09
*** slaweq has joined #openstack-nova03:11
*** slaweq has quit IRC03:15
*** Dinesh_Bhor has quit IRC03:17
openstackgerrit98k proposed openstack/os-traits master: Add python 3.6 unit test job  https://review.openstack.org/61674903:18
*** Dinesh_Bhor has joined #openstack-nova03:20
*** tbachman has quit IRC03:27
*** Dinesh_Bhor has quit IRC04:08
*** janki has joined #openstack-nova04:36
*** Dinesh_Bhor has joined #openstack-nova04:40
*** openstackstatus has quit IRC04:59
*** hongbin has quit IRC05:07
*** slaweq has joined #openstack-nova05:11
*** moshele has joined #openstack-nova05:12
*** slaweq has quit IRC05:16
*** moshele has quit IRC05:18
*** openstack has joined #openstack-nova07:09
*** ChanServ sets mode: +o openstack07:09
*** dpawlik has joined #openstack-nova07:12
openstackgerritTakashi NATSUME proposed openstack/nova master: Fix server query examples  https://review.openstack.org/61683407:13
*** sahid has joined #openstack-nova07:18
openstackgerritTakashi NATSUME proposed openstack/nova master: Remove mox in unit/network/test_neutronv2.py (6)  https://review.openstack.org/57411307:20
openstackgerritTakashi NATSUME proposed openstack/nova master: Remove mox in unit/network/test_neutronv2.py (7)  https://review.openstack.org/57497407:20
openstackgerritTakashi NATSUME proposed openstack/nova master: Remove mox in unit/network/test_neutronv2.py (8)  https://review.openstack.org/57531107:20
*** pcaruana has joined #openstack-nova07:21
*** dpawlik has quit IRC07:28
*** dpawlik has joined #openstack-nova07:29
*** dpawlik has quit IRC07:29
*** dpawlik has joined #openstack-nova07:30
*** slaweq has joined #openstack-nova07:45
*** jangutter has quit IRC07:50
*** jangutter has joined #openstack-nova07:50
openstackgerritBrin Zhang proposed openstack/nova-specs master: Support admin to specify project to create snapshot  https://review.openstack.org/61684307:52
gibimelwitt: I think there is only one patch left from use-nested-allocation-candidates that handles allocations that only use nested RPs and not the root RP https://review.openstack.org/#/c/608298/07:57
gibimelwitt: that happen to be lower prio to me as that does not needed for GPU and Bandwidth work. It would be needed for the NUMA work though07:58
gibimelwitt: so I think we can pull use-nested-allocation-candidates from runway08:01
*** takashin has left #openstack-nova08:03
gibimelwitt: I did pull that from the runway now on the etherpad08:03
melwittgibi: got it, thank you. great work on all of that btw. it is exciting08:05
openstackgerritJeffrey Zhang proposed openstack/nova master: Add feature to flatten the volume from glance image snapshort  https://review.openstack.org/61646108:06
*** tssurya has joined #openstack-nova08:07
gibimelwitt: thanks for looking at the bandwidth patches. It was motivating for me to see them moving forward.08:07
gibimelwitt: this week I finally resumed to work on that series08:08
melwittnp, it is really cool to see it all coming together now08:08
*** trident has quit IRC08:12
*** trident has joined #openstack-nova08:14
*** ralonsoh has joined #openstack-nova08:15
*** helenaAM has joined #openstack-nova08:31
*** sridharg has joined #openstack-nova08:54
*** sridharg has quit IRC08:54
*** sridharg has joined #openstack-nova08:55
*** hshiina has quit IRC09:09
*** tetsuro has quit IRC09:20
*** panda|rover|off is now known as panda|rover09:24
*** Dinesh_Bhor has quit IRC09:34
*** derekh has joined #openstack-nova09:42
*** k_mouza has joined #openstack-nova09:52
*** k_mouza has quit IRC09:53
*** k_mouza has joined #openstack-nova09:54
*** Dinesh_Bhor has joined #openstack-nova09:57
*** Dinesh_Bhor has quit IRC10:01
*** ttsiouts has joined #openstack-nova10:05
*** ttsiouts has quit IRC10:10
*** ttsiouts has joined #openstack-nova10:11
*** ttsiouts has quit IRC10:15
*** ttsiouts has joined #openstack-nova10:20
*** ttsiouts has quit IRC10:21
*** ttsiouts has joined #openstack-nova10:22
*** ttsiouts has quit IRC10:26
*** ttsiouts has joined #openstack-nova10:30
openstackgerritzhouxinyong proposed openstack/nova master: delete unavailable links  https://review.openstack.org/61687010:30
openstackgerritSurya Seetharaman proposed openstack/nova master: Make _instances_cores_ram_count() be smart about cells  https://review.openstack.org/56905510:31
openstackgerritSurya Seetharaman proposed openstack/nova master: Make _instances_cores_ram_count() be smart about cells  https://review.openstack.org/56905510:32
*** sambetts_ is now known as sambetts|afk10:33
*** davidsha has joined #openstack-nova10:36
openstackgerritzhouxinyong proposed openstack/nova master: delete unavailable links  https://review.openstack.org/61687010:52
*** maciejjozefczyk has quit IRC10:59
*** ttsiouts has quit IRC10:59
*** ttsiouts has joined #openstack-nova11:00
*** maciejjozefczyk has joined #openstack-nova11:01
*** maciejjozefczyk has joined #openstack-nova11:01
*** tssurya has quit IRC11:04
*** dpawlik has quit IRC11:08
*** rodolof has joined #openstack-nova11:12
*** k_mouza has quit IRC11:31
*** k_mouza has joined #openstack-nova11:33
*** k_mouza has quit IRC11:52
*** dtantsur is now known as dtantsur|brb11:59
*** k_mouza has joined #openstack-nova12:05
*** brinzhang has quit IRC12:10
*** janki has quit IRC12:11
*** ondrejme has joined #openstack-nova12:13
*** maciejjozefczyk has quit IRC12:15
*** maciejjozefczyk has joined #openstack-nova12:16
*** erlon has joined #openstack-nova12:29
*** panda|rover is now known as panda|rover|lch12:31
*** alexchadin has joined #openstack-nova12:39
*** rodolof has quit IRC12:55
openstackgerritzhouxinyong proposed openstack/nova master: modify the avaliable link  https://review.openstack.org/61690513:08
*** dtantsur|brb is now known as dtantsur13:09
*** k_mouza has quit IRC13:11
*** Dinesh_Bhor has joined #openstack-nova13:18
*** k_mouza has joined #openstack-nova13:23
openstackgerritLee Yarwood proposed openstack/nova master: DNM WIP zuul: Add a lioadm based multiattach job  https://review.openstack.org/61691613:26
*** Dinesh_Bhor has quit IRC13:34
*** sahid has quit IRC13:42
*** sahid has joined #openstack-nova13:47
*** mriedem has joined #openstack-nova14:08
*** tbachman has joined #openstack-nova14:11
openstackgerritBalazs Gibizer proposed openstack/nova master: Calculate port_id rp_uuid mapping for binding  https://review.openstack.org/61623914:18
openstackgerritBalazs Gibizer proposed openstack/nova master: Pass allocations and traits to neturonv2 api  https://review.openstack.org/61624014:18
openstackgerritBalazs Gibizer proposed openstack/nova master: Send RP uuid  in the port binding  https://review.openstack.org/56945914:18
openstackgerritBalazs Gibizer proposed openstack/nova master: Test boot with more ports with bandwidth request  https://review.openstack.org/57331714:18
openstackgerritIvaylo Mitev proposed openstack/nova master: VMware: Attach volumes using adapter type from instance  https://review.openstack.org/61659914:18
openstackgerritMatt Riedemann proposed openstack/nova master: libvirt: change "Ignoring supplied device name" warning to info  https://review.openstack.org/61695214:19
*** davidsha has quit IRC14:32
*** eharney has joined #openstack-nova14:34
*** liuyulong has joined #openstack-nova14:35
*** eharney has quit IRC14:54
*** tssurya has joined #openstack-nova14:57
*** maciejjozefczyk has quit IRC15:00
*** lbragstad has quit IRC15:02
*** lbragstad has joined #openstack-nova15:03
*** maciejjozefczyk has joined #openstack-nova15:07
*** maciejjozefczyk has quit IRC15:07
*** maciejjozefczyk has joined #openstack-nova15:09
*** alexchadin has quit IRC15:10
*** k_mouza has quit IRC15:16
*** hongbin has joined #openstack-nova15:18
*** artom has quit IRC15:21
*** artom has joined #openstack-nova15:21
*** k_mouza has joined #openstack-nova15:24
*** maciejjozefczyk has quit IRC15:33
*** maciejjozefczyk has joined #openstack-nova15:34
*** maciejjozefczyk has quit IRC15:34
*** maciejjozefczyk has joined #openstack-nova15:34
*** maciejjozefczyk has quit IRC15:35
*** maciejjozefczyk has joined #openstack-nova15:35
*** rnoriega has quit IRC15:36
*** maciejjozefczyk has quit IRC15:36
*** rnoriega has joined #openstack-nova15:36
mriedemaspiers: done https://review.openstack.org/#/c/609779/15:43
mriedemi'm really not thrilled at the amount of technical debt we'll be taking on if we add this15:43
mriedembut such is life i suppose15:43
mriedemby technical debt i mean "oh i can create and destroy a sev-enabled vm, but that's all i can do with it"15:43
openstackgerritJack Ding proposed openstack/nova master: Add cache=none option for qemu-img convert  https://review.openstack.org/61669215:45
*** spatel has joined #openstack-nova15:48
spatelsean-k-mooney: Howdy!!!15:50
spatelMorning15:50
spatelhttps://bugs.launchpad.net/nova/+bug/179276315:50
openstackLaunchpad bug 1792763 in OpenStack Compute (nova) "tap TX packet drops during high cpu load " [Undecided,Invalid]15:50
spatelYes this could be resolved or close.. because its design question ( i won't say its BUG )15:51
*** bnemec is now known as beekneemech15:53
sean-k-mooneyya the drops you were seeing were a limitation of linux bridge as you said so this is not something nova can fix15:53
sean-k-mooneyspatel: the kernel can only handel about 1.4mpps on a 3.4GHz cpu and generally its less then that15:54
sean-k-mooneyonce you exceed that level you get drops.15:54
spatelIn my test after 50kpps i was seeing TX drop on tap interface15:55
*** eharney has joined #openstack-nova15:56
spatelI think this is issue of tap interface design, it run on kernel space which is overhead on kernel15:56
spatelvirtio i meant15:56
sean-k-mooneyto get to 1.4 you need kernel ovs with kernel vhost module to acclerate it15:56
sean-k-mooneyspatel: yes so its not something openstack can remedy15:57
spatel++15:57
spatelIn my case i am using linuxbridge (not ovs)15:57
spateldo you think OVS is better in performance compare to linuxbridge ?15:58
sean-k-mooneyyes it is15:58
sean-k-mooneybar multicast tunnelling15:58
sean-k-mooneyif you have a multicast hevey workload use linux bridge as ovs falls back to unicast15:58
sean-k-mooneybut in gereral ovs out performes linux bridge in vm based workloads15:59
spatelwhen you say multicast what is the relation here?15:59
*** jistr is now known as jistr|call16:00
sean-k-mooneylinux bridge support using multicast endpoint for teant networks meaning it can more efficetly handel tenatn traffic with a high proportion of broadcase or multicast traffic16:00
sean-k-mooneyovs does not and has to fall back to a unicast mesh toplogy for vxlan16:01
sean-k-mooneybut for typical workloads ovs will out perfrom linux bridge16:01
dansmithmriedem: I threw a comment in there about using sysmeta to let virt drivers declare some ops as invalid for an instance. is there some reason that's not reasonable?16:04
*** burt has joined #openstack-nova16:04
dansmithpresumably 403 is allowed for pretty much any operation on any microversion, so I would think it'd be not a huge deal, and immediately applies to existing operations in certain situations16:04
*** jangutter has quit IRC16:07
*** Luzi has quit IRC16:07
*** janki has joined #openstack-nova16:09
mriedemdansmith: yeah, replied16:09
mriedemit really goes back to the capabilities thing we've discussed several times before16:09
mriedemi'm mostly concerned about snapshot, because if you can't move the instance, users are at least going to want to be able to snapshot it i'd think before it has to be destroyed and recreated elsewhere because the compute it's on is going away16:11
mriedemof course this is where someone says, "just attach a data volume and rewrite the application to use that"16:12
spatelsean-k-mooney: thanks for clear that point.. :) i have all unicast workload16:12
*** lbragstad has quit IRC16:14
dansmithmriedem: did he say snapshot wasn't supported? I would think it would be16:15
*** lbragstad has joined #openstack-nova16:15
dansmiththe airplane wifi is sucking too hard for me to even open it again16:16
mriedemit wasn't mentioned16:16
mriedemthat's why i asked, because it sure seems like a lot can't be supported16:16
*** jistr|call is now known as jistr16:17
*** imacdonn has quit IRC16:18
mriedemwe got a bug b/c of the limit of tenant ids for the aggregate multitenancy isolation filter, that's resolved with the placement request filter, but doesn't look like the docs for the placement filter mention you can namespace the metadata so you can add as many tenants as you want16:18
dansmithmriedem: the suspend/resume and live migration are about in-memory state, which is why they're hard to support I think16:18
dansmithsnapshot, reboot, cold migrate should all be fine I would think16:18
dansmithbased on my reading and assumptions about how this works16:18
*** cfriesen has joined #openstack-nova16:18
dansmithmriedem: hmm, I was sure I put that in there16:19
mriedemdon't see it, i can push up something for that16:19
dansmithokay16:20
mriedemand i'll probably update the docs for the old filter to mention the limitation (and link to the bug) and say the placement one is a better replacement16:20
dansmithack16:21
dansmithdid I have it in the commit message or something?16:21
dansmithI was sure I wrote words about this16:21
*** etp has quit IRC16:21
*** gyee has joined #openstack-nova16:22
*** tssurya has quit IRC16:22
mriedemhttps://review.openstack.org/#/c/545002/27 "This also allows making this filter advisory but not required, and supports multiple tenants per aggregate, unlike the original filter."16:22
mriedemmaybe that16:22
dansmithnova with the ``filter_tenant_id`` key (optionally suffixed with any string for16:24
dansmith    multiple tenants,16:24
dansmithhttps://review.openstack.org/#/c/557490/8/releasenotes/notes/tenant_aggregate_placement_filter-c2fed8889f43b6e3.yaml16:24
dansmithin the reno not hte docs16:24
dansmiththa's mah bad16:25
mriedemk, i'll copy that16:25
*** etp has joined #openstack-nova16:27
openstackgerritBalazs Gibizer proposed openstack/nova master: Calculate port_id rp_uuid mapping for binding  https://review.openstack.org/61623916:28
openstackgerritBalazs Gibizer proposed openstack/nova master: Pass allocations and traits to neturonv2 api  https://review.openstack.org/61624016:28
openstackgerritBalazs Gibizer proposed openstack/nova master: Send RP uuid  in the port binding  https://review.openstack.org/56945916:28
openstackgerritBalazs Gibizer proposed openstack/nova master: Test boot with more ports with bandwidth request  https://review.openstack.org/57331716:29
openstackgerritMatt Riedemann proposed openstack/nova master: Mention meta key suffix in tenant isolation with placement docs  https://review.openstack.org/61699116:33
mriedemsean-k-mooney: is it just me or is NeutronLinuxBridgeInterfaceDriver completely replaced with os-vif now?16:38
sean-k-mooneyill need to check but probably16:41
mriedemi think it would only be used via the linuxnet_interface_driver config option, but i don't see anything with neutron in nova using the code path that hits that option16:42
mriedemonly the nova-network l3 stuff16:42
sean-k-mooneywe can likely kill it when we kill nova networks16:42
mriedemsure but this is a neutron-specific driver16:42
sean-k-mooneymriedem: ill look into it next week while and see what its actully used for but your right16:43
sean-k-mooney* while ye are at the summit16:44
*** spatel has quit IRC16:44
cfriesenstephenfin: question about your commit https://review.openstack.org/#/c/526329   One of our guys says he ran into a scenario in pike where "image_chunks" itself was None due to things like firewall breakage or server-side problems.  Does that get handled properly currently?16:45
openstackgerritMerged openstack/nova stable/pike: Fix the request context in ServiceFixture  https://review.openstack.org/59983916:46
openstackgerritMerged openstack/nova stable/pike: Add functional test for affinity with multiple cells  https://review.openstack.org/59984016:46
*** k_mouza has quit IRC16:46
openstackgerritMatt Riedemann proposed openstack/nova master: Delete NeutronLinuxBridgeInterfaceDriver  https://review.openstack.org/61699516:54
*** spatel has joined #openstack-nova16:54
*** helenaAM has quit IRC16:59
*** dtantsur is now known as dtantsur|afk17:00
*** hongbin has quit IRC17:02
cfriesenmriedem: regarding the disk sector size issue...are you aware of any 8K sector disks or did you suggest it for future expansion?17:03
stephenfincfriesen: Based on that as-is, it would not17:03
stephenfincfriesen: Though I'd have expected to see an exception raised by the client, more so than anything else17:04
mriedemcfriesen: was just suggesting based on what was noted in the bug report17:06
sean-k-mooneyim not aware of any 8k sector discs but i belive we can also diskcover the sector size by querying the disk via sysfs so we proably dont need to hardcode it17:07
sean-k-mooneythat said 4k and 512 are the most common17:07
cfriesenI thought that 4K was still the highest supported physical sector size (since you'd want to be able to read a whole disk sector into a memory page)17:07
dansmithnot everyone uses 4k pages :)17:08
sean-k-mooneypower pc i think is 16k17:08
dansmithI thought there were some SAN types that used larger sector sizes just because of the network optimization,17:08
dansmitheven if not backed by actual 8k17:08
cfriesenas far as I know the block size can be different from the sector size17:09
dansmithalso, netapp I think uses some super odd sizes, even to the point of having weirdly low-level-formatted drives for them17:09
sean-k-mooneyi know some raid controls can be configured to exposed larger sector sizes but i dont know how common that is anymore17:09
dansmithyep17:09
sean-k-mooneymriedem:  i think one of the things you suggested was just making a config  option correct17:10
sean-k-mooneysomething like directio_sector_sizes=512,409617:11
dansmithah, but it's hidden to the LUN: https://kb.netapp.com/app/answers/answer_view/a_id/1001353/~/how-can-the-bytes%2Fsector-be-changed-in-a-luns-geometry%3F-17:13
cfriesenthe goal here is to figure out if the filesystem supports O_DIRECT.  according to the man page, this should be set to the logical block size of the underlying storage, which can be determined using the ioctl() BLKSSZGET operation or by calling "blockdev --getss"17:14
sean-k-mooneycfriesen:  i think part of the issue is that on older kernel < 2.4 O_DIRECT required alinged access17:16
sean-k-mooneybut on bsd and newer linux kernel O_DIRECT did not reuqire alinged acess17:16
sean-k-mooneycfriesen: im not actully sure of we need to do the alignment check we are doing anymore17:17
openstackgerritMerged openstack/nova stable/queens: fixtures: Track volume attachments within CinderFixtureNewAttachFlow  https://review.openstack.org/61249417:17
cfriesenthe man page says: Under Linux 2.4, transfer sizes, and the alignment of the user buffer17:17
cfriesen       and the file offset must all be multiples of the logical block size17:17
cfriesen       of the filesystem.  Since Linux 2.6.0, alignment to the logical block17:17
cfriesen       size of the underlying storage (typically 512 bytes) suffices.17:17
cfriesenoh, ick. sorry17:17
openstackgerritMerged openstack/nova stable/queens: Add regression test for bug#1784353  https://review.openstack.org/61249517:17
sean-k-mooneycfriesen: ok so we still need aligned acess but the function is ment to determin if directio is possibel what its currently doing is determing if direct 512B aligned acess is possible with is a different thing17:19
cfriesenagreed.  I think switching to 4K would cover 95% of the cases.17:19
cfriesendoing it totally correctly woudl require querying the block size from the OS for that specific device17:19
mriedemsean-k-mooney: i suggested a config option as an option because this sounds very hit or miss17:20
sean-k-mooneywell the backing store for libvirt instance is only going to be on one mountpoint17:21
mriedemthis is definitely not something i've got a lot of experience in though17:21
sean-k-mooneypresumable it will all have the same alignment/sector size so we could jsut have a single valus and defualt it to 512 and they could set it to 4k or 8k if they have something else17:22
sean-k-mooneythe other option is just super over align to like a 64K bondary17:23
sean-k-mooneythat said im sure someone will have a 128K lun now that i have said that17:23
cfriesenso guaranteed setting it to 4K will work for both 4K and 512b disks, so I think 4K should be the default17:24
sean-k-mooneycfriesen: its also the most common sector size on most new disks so ya that should work17:24
dansmithfor years now17:25
cfriesenI'd be okay with a config option if someone has weird hardware17:25
sean-k-mooneycfriesen: so your going to submit a patch :)_17:26
cfriesenthere's already a patch in progress17:26
cfriesenby someone else17:26
mriedemwee https://bugs.launchpad.net/nova/+bug/179868817:27
openstackLaunchpad bug 1798688 in OpenStack Compute (nova) "AllocationUpdateFailed_Remote: Failed to update allocations for consumer. Error: another process changed the consumer after the report client read the consumer state during the claim" [Undecided,Triaged]17:27
mriedemlooks like our scheduler allocation claim races have shot up since nov 417:27
cfriesenmriedem: did you want to get the person to update the 4K patch to add a config option?  or change the hardcoded number to 4K (which is still an improvement) and add the config option later if someone complains?17:28
dansmithmriedem: what does that mean exactly? just the scheduler having to retry the allocation part?17:29
mriedemdansmith: we already retry on PUT allocations in the scheduler17:29
mriedemmaybe the error message changed and we're not retrying properly now?17:29
mriedemi haven't dug in yet17:29
mriedemcfriesen: the patch already just changes 512 to 4k right?17:29
mriedemcfriesen: i believe i just said we might want a 'fixes' release note for it17:30
mriedemas a heads up17:30
melwittcfriesen: this is the method we have to checking for directio, if that's the same thing you mentioned earlier https://github.com/openstack/nova/blob/master/nova/privsep/utils.py#L3417:30
dansmithmriedem: I know we do, I'm wondering if you mean it's just having to retry more lately or if it's failing17:30
mriedemhaven't dug into the logs yet17:30
mriedemhopefully we log if we are retrying17:30
sean-k-mooneymriedem: ya i was under teh impression we had planned at least to retry in this case which is why we have the generation on the resouce providers in the first place17:30
*** markvoelker has quit IRC17:30
dansmithsean-k-mooney: we do retry17:31
mriedemthe scheduler doesn't do anything with generations for this as far as i know17:31
mriedemjust duplicated another bug in triage to this if someone is looking for work https://bugs.launchpad.net/nova/+bug/178333817:31
openstackLaunchpad bug 1783338 in OpenStack Compute (nova) "Unexpected exception in API method: ValueError: year is out of range" [Medium,Confirmed] - Assigned to Ghanshyam Mann (ghanshyammann)17:31
mriedemsomething in the simple tenant usage code17:31
cfriesenmriedem: in irc you were talking about a config option I thought.  but yeah, I'd be cool with just a release note for now.17:32
*** imacdonn has joined #openstack-nova17:32
mriedemgmann: looks like https://bugs.launchpad.net/nova/+bug/1783338 was due to bad data in the db? you're probably traveling, but if you don't plan on handling this we should unassign you https://bugs.launchpad.net/nova/+bug/178333817:32
openstackLaunchpad bug 1783338 in OpenStack Compute (nova) "Unexpected exception in API method: ValueError: year is out of range" [Medium,Confirmed] - Assigned to Ghanshyam Mann (ghanshyammann)17:32
cfriesenmelwitt: yes, that's the one.  there's a patch in review to change the 512 to 4096 in there.  which is good, but maybe not sufficient for exotic hardware17:32
mriedemi'm mriedem17:33
melwittI said something to him earlier17:33
melwittcfriesen: ok, cool. *looks for the patch*17:34
mriedemoh missed that17:34
cfriesenmelwitt: https://review.openstack.org/#/c/61658017:34
melwittprobably because our nicks blend together. maybe I need to be jgwentworth all the time17:34
mriedemdansmith: looks like, from the logs, that we're not retrying17:35
dansmithmriedem: maybe something changed recently then?17:36
*** derekh has quit IRC17:37
melwittcfriesen: ok, so looks like trying to decide which value to use for the check17:39
*** sahid has quit IRC17:40
mriedemdansmith: my guess would be https://review.openstack.org/#/c/583667/17:40
mriedembecause the scheduler logs are saying we're doing a double up allocation17:40
mriedemNov 06 19:48:36.969356 ubuntu-xenial-inap-mtl01-0000379614 nova-scheduler[12154]: DEBUG nova.scheduler.client.report [None req-f266a0ff-2840-413d-9877-4500e61512f5 tempest-ServersNegativeTestJSON-477704048 tempest-ServersNegativeTestJSON-477704048] Doubling-up allocation_request for move operation. {{(pid=13677) _move_operation_alloc_request /opt/stack/nova/nova/scheduler/client/report.py:203}}17:40
mriedembut in this test, we're just unshelving a shelved offloaded server17:41
mriedemso that shouldn't really double up any allocatoins17:41
cfriesenmelwitt: basically, yes.  4096 would work for the vast majority of systems17:42
mriedemNov 06 19:48:36.969659 ubuntu-xenial-inap-mtl01-0000379614 nova-scheduler[12154]: DEBUG nova.scheduler.client.report [None req-f266a0ff-2840-413d-9877-4500e61512f5 tempest-ServersNegativeTestJSON-477704048 tempest-ServersNegativeTestJSON-477704048] New allocation_request containing both source and destination hosts in move operation: {'allocations': {u'3ceb7eab-549c-40ba-a70c-320822c310ab': {'resources': {u'VCPU': 2, u'MEMORY17:42
mriedem: 128}}}} {{(pid=13677) _move_operation_alloc_request /opt/stack/nova/nova/scheduler/client/report.py:234}}17:42
dansmithmriedem: it also touches the code near where we raise retry...17:42
mriedem^ is definitely wrong17:42
mriedemthere is only one provider in that log17:43
dansmithmriedem: it seems to specifically exclude the consumer generation conflict from the case where we retry17:43
dansmithmriedem: do you see "another process changed the consumer" in the log?17:46
mriedemyes17:46
dansmiththen it's hitting that consumer case and not retrying17:46
mriedemthat's why we don't retry17:46
dansmithyeah17:46
mriedemi also don't know why it thinks we're starting with existing allocations for a shelved offloaded server17:47
dansmithand that's causing it to try to double?17:47
mriedemwell, it goes into _move_operation_alloc_request but doesn't actually double anything17:48
mriedemNov 06 19:48:36.969659 ubuntu-xenial-inap-mtl01-0000379614 nova-scheduler[12154]: DEBUG nova.scheduler.client.report [None req-f266a0ff-2840-413d-9877-4500e61512f5 tempest-ServersNegativeTestJSON-477704048 tempest-ServersNegativeTestJSON-477704048] New allocation_request containing both source and destination hosts in move operation: {'allocations': {u'3ceb7eab-549c-40ba-a70c-320822c310ab': {'resources': {u'VCPU': 2, u'MEMORY17:48
mriedem: 128}}}} {{(pid=13677) _move_operation_alloc_request /opt/stack/nova/nova/scheduler/client/report.py:234}}17:48
mriedemthere is only one provider in that body17:48
mriedemthis is the error from placement17:51
mriedemNov 06 19:48:37.013780 ubuntu-xenial-inap-mtl01-0000379614 nova-scheduler[12154]: WARNING nova.scheduler.client.report [None req-f266a0ff-2840-413d-9877-4500e61512f5 tempest-ServersNegativeTestJSON-477704048 tempest-ServersNegativeTestJSON-477704048] Failed to save allocation for 6665f00a-dcf1-4286-b075-d7dcd7c37487. Got HTTP 409: {"errors": [{"status": 409, "request_id": "req-c9ba6cbd-3b6e-4e5d-b550-9588be8a49d2", "code": "p17:51
mriedemment.concurrent_update", "detail": "There was a conflict when trying to complete your request.\n\n consumer generation conflict - expected null but got 1  ", "title": "Conflict"}]}17:51
*** ralonsoh has quit IRC17:51
*** ttsiouts has quit IRC17:51
mriedemidk wtf is going on, but i see 3 different PUT allocations in the placement logs for that consumer17:58
mriedemfirst is probably for the initial scheduler, and then we offload and remove allocations17:58
mriedem2nd is for the unshelve17:59
*** ivve has joined #openstack-nova18:00
mriedemaha18:07
mriedemthe allocatoin delete on unshelve changed with this patch https://review.openstack.org/#/c/591597/18:07
mriedemso we no longer actually delete allocations, we PUT {}18:07
sean-k-mooneyquick question. in what cases does nova update the network info cache?18:07
*** Swami has joined #openstack-nova18:07
sean-k-mooneyi know it does it in respconce to neutron notification. there is also a periodic heal task right18:08
sean-k-mooneyis that it?18:08
mriedemin all cases18:08
mriedemattach vifs18:08
mriedemetc18:08
mriedemdansmith: yeah so there are 3 PUTs for allocations, 1st for initial schedule, 2nd for shelve offload (PUT /allocations with {}) and then the 3rd is scheduling during unshelve18:08
mriedemsince the allocations aren't deleted on shelve offload, the consumer must persist in placement with it's existing consumer generation18:09
mriedemwhich the sheduler in the 3rd PUT doesn't account for18:09
dansmithokay18:09
dansmiththe scheduler assumes that the consumer is gone?18:09
dansmithI'm surprised it would care,18:09
dansmithbecause it doesn't know if we're doing an initial boot or a move right?18:10
mriedemit would know if we're doing a move if the consumer already has allocations elsewhere18:10
dansmithright, but otherwise it doesn't,18:10
dansmithand in this case there are no remaining allocations right?18:11
mriedemwell,18:11
dansmithor are you saying it assumes that if you have no allocations the consumer can't exist?18:11
*** sridharg has quit IRC18:11
mriedemthat's not what the scheduler thinks, because it goes down that _move_operation_alloc_request path18:11
mriedemmaybe the tempest test isn't really waiting for the instance be fully shelved offloaded before it unshelves18:11
mriedemah indeed,18:11
mriedemwe set the instance vm_state to SHELVED_OFFLOADED *before* we remove allocations18:12
dansmithI guess I'm still surprised it cares18:12
mriedemwhich is definitely a race18:12
*** zul has quit IRC18:12
dansmithwhy?18:16
dansmithjust because it signals to tempest that it's done?18:16
dansmithif we didn't,18:17
dansmithand we crashed right between deleting the allocations and makring it as offloaded we'd have lost some information I would think18:17
mriedemwell, this also doesn't seem right18:17
mriedemhttps://review.openstack.org/#/c/591597/8/nova/scheduler/client/report.py@209118:17
mriedem"# removing all resources from the allocation will auto delete the18:17
mriedem        # consumer in placement"18:17
dansmithI guess maybe your point is that it's a race between starting the unshelve and there still being allocations...18:18
mriedemmaybe that is happening, i'm not sure18:18
mriedemcorrect18:18
mriedemb/c i'm seeing this being true during unshelve https://github.com/openstack/nova/blob/e27905f482ba26d2bbf3ae5d948dee37523042d5/nova/scheduler/client/report.py#L182418:18
dansmithnot sure that's better either way18:18
mriedemwhich shouldn't be the case18:18
*** k_mouza has joined #openstack-nova18:19
dansmithit's certainly possible that the move to PUT{} from DELETE didn't bring over some "and delete the consumer" part18:19
dansmithbut again, I'm not sure why it should matter to the scheduler that it exists18:20
dansmithalthough...18:20
dansmithwe have no api for looking at the consumer to get the generation if it already exists, IIRC18:20
dansmithso maybe without seeing an allocation, and not being able to see the consumer directly, we have no alternative?18:20
mriedemright, it's supposed to  come back on the GET /allocations/{consumer} call18:20
dansmithI expect jaypipes to pop in here any second and say "ah hah!"18:20
mriedemjaypipes is busy chefing it up18:21
dansmithhis fave18:21
mriedemand getting ready to sleep for a week while the rest of us are in berlin18:21
dansmithlucky bastard18:21
dansmithI'm going to be landing pretty soon, FYI18:21
mriedemi'm overdue for lunch as well18:22
mriedemso, can't claim_resources in the scheduler still just retry if it hits that consumer generatoin conflict?18:22
dansmithwell,18:22
dansmithnot if it doesn't know what the generation is18:23
dansmiththat's what I was saying.. it might not be able to find out what it is,18:23
dansmithwith no consumer api and no existing allocation to look at18:23
mriedemi would expect GET /allocations/{consumer_uuid} to return the consumer generation18:23
mriedemeven if allocations are {}18:23
dansmithwith no alloc records?18:24
dansmithI dunno18:24
mriedembut i guess i'd have to dig into the placement code18:24
dansmithI would expect that code returns 404 if none come back,18:24
*** k_mouza has quit IRC18:24
mriedemshould probably also log in placement when the consumer is deleted b/c allocations went to 018:24
dansmithbecause it would only get the consume through the join, or afterwards I would expect18:24
mriedemno it doesn't 404, you get {"allocations": {}}18:24
mriedemif there are no allocations for the consumer18:24
dansmithoh?18:24
mriedemyeah it's confusing18:24
dansmiththat seems supremely weird to me, but okay18:25
mriedemwhat does taylor think about all this?18:25
dansmithshe's busy with her own work18:26
jaypipesmriedem: fuck chef. fuck ansible. fuck docker. it's all a bunch of complete assbaggery.18:26
dansmithmy little mobile wifi router lets us share the same crappy airline wifi, so after that came online, I might as well not be sitting next to her18:26
mriedemjaypipes: but salt?!18:26
* jaypipes reads back to see something about consumers.18:26
dansmithjaypipes: you're gonna love it18:27
sean-k-mooneymriedem: i think jaypipes has enough salt in his life right now18:27
mriedemjaypipes: notes are in https://bugs.launchpad.net/nova/+bug/179868818:27
openstackLaunchpad bug 1798688 in OpenStack Compute (nova) "AllocationUpdateFailed_Remote: Failed to update allocations for consumer. Error: another process changed the consumer after the report client read the consumer state during the claim" [Undecided,Triaged]18:27
dansmithjaypipes: question.. should placement have a consumers endpoint?18:27
mriedemlooks like we're racing between shelve offload (allocation removal) and unshelve (put new allocations) and hitting a consumer generation conflict18:27
sean-k-mooneyjaypipes: at least you dont have to use tripplo where we use yaml to drive heat to drive puppet to drive ansible to deploy docker containers ...18:28
sean-k-mooneyor maybe the ansible drives puppet its hard to keep track of18:28
dansmithsean-k-mooney: it drives ... me insane18:29
mriedemso, i see in placement handler code where it handles the "allocations are being removed on PUT" case, and it ensures a consumer exists, but then i don't see where that consumer is deleted18:30
mriedemlike the note in the compute code18:30
mriedemhttps://github.com/openstack/nova/blob/e27905f482ba26d2bbf3ae5d948dee37523042d5/nova/api/openstack/placement/handlers/allocation.py#L40418:32
mriedemoh i guess it should happen here https://github.com/openstack/nova/blob/e27905f482ba26d2bbf3ae5d948dee37523042d5/nova/api/openstack/placement/objects/resource_provider.py#L209918:34
mriedemhttps://github.com/openstack/nova/blob/e27905f482ba26d2bbf3ae5d948dee37523042d5/nova/api/openstack/placement/objects/consumer.py#L70 might be broken18:36
mriedemin the same way that ensure was broken https://github.com/openstack/nova/commit/730936e535e67127c76d4f27649a16d8cf05efc9#diff-fcca11e34c1b5fce52a4ddbc418aa2d518:36
openstackgerritMerged openstack/nova stable/queens: conductor: Recreate volume attachments during a reschedule  https://review.openstack.org/61249618:37
openstackgerritMerged openstack/nova master: Update the description to make it more accuracy  https://review.openstack.org/61536218:37
mriedemi can't really tell where delete_consumers_if_no_allocations is tested though...18:39
mriedemsome gabbit i'm sure18:40
dansmithseems easily unit testable,18:41
dansmithand I definitely can't look at that and tell that it works18:41
dansmithsince it's joining on consume id and asserting that it's none in one case18:41
mriedemyeah i can't do anything with the sql w/o testing it18:42
dansmithtime to pack up, back later18:42
mriedemoh i guess DeleteConsumerIfNoAllocsTestCase18:42
mriedemi'll tweak that after lunch18:43
*** bigdogstl has joined #openstack-nova18:50
openstackgerritMatt Riedemann proposed openstack/nova master: Add debug logs when doubling-up allocations during scheduling  https://review.openstack.org/61701618:53
openstackgerritMatt Riedemann proposed openstack/nova master: Log consumers_to_check when calling delete_consumers_if_no_allocations  https://review.openstack.org/61701718:53
mriedemjaypipes: debug logging needed for this gate bug ^18:53
jaypipesstill reading back, sorry18:54
*** mriedem is now known as mriedem_hangry18:55
jaypipesdansmith, mriedem_hangry: there is a check at the end of the server-side of PUT /allocations that will auto-delete the consumer record if there are no allocations still referring to it.19:04
jaypipesdansmith: and yes, I've said for a long time that we should have a GET /consumers endpoint. There are placement devs that vehemently disagreed with that.19:07
*** bigdogstl has quit IRC19:08
*** bigdogstl has joined #openstack-nova19:12
*** ivve has quit IRC19:12
*** janki has quit IRC19:22
*** zigo has quit IRC19:25
*** bigdogstl has quit IRC19:26
sean-k-mooneymriedem_hangry: would you have any objection to backporting https://review.openstack.org/#/c/591607/9 to newton? im pretty sure we have a customer that is hitting this as they reported instance restarting after a host reboot missing interfaces that show up in nova interface-list19:27
*** bigdogstl has joined #openstack-nova19:30
sean-k-mooneymriedem_hangry: actullly i just realised newton is way older then i remembered and is eol19:34
*** mriedem_hangry is now known as mriedem19:34
mriedemjaypipes: yup found that, and the related functional test19:34
mriedemsean-k-mooney: not to mention that isn't even approved on master19:35
*** bigdogstl has quit IRC19:35
sean-k-mooneymriedem: yes :) im aware. i was more asking do you think this is something that can be backported in general upstream19:35
sean-k-mooneyonce it lands in master19:36
mriedemidk19:36
mriedemit seems to be pretty controversial19:37
mriedemlike my david bowie costume on halloween19:37
sean-k-mooneyim waiting on more logs for the downstream bug to confirm this is actully the issue19:38
sean-k-mooneymriedem: ill try to review and digest these chagnes more on monday19:40
sean-k-mooneymriedem: are you flying out to berlin today/tomorow?19:41
mriedemthe problem the huawei ops team ran into was policy changed on the neutron side which started returning an empty list of ports, which was then saved into the info cache in the nova db,19:41
mriedemand the heal periodic relies on the info cache rather than the source of truth to fix the cache19:41
mriedemtonight19:41
mriedemthere are other ways to simply rebuild the cache if that's what is needed, e.g. https://docs.openstack.org/python-novaclient/latest/cli/nova.html#nova-refresh-network19:41
sean-k-mooneyright. i think we discussed this in the past at some point too it feels familar but i have not reviewd this before19:42
mriedemit came up at the ptg i think19:42
mriedemand the public cloud sig has brought it up before (OVH obviously)19:42
sean-k-mooneymriedem: perhaps. oh so we can force a rebuild via the nova client today?19:42
* sean-k-mooney clicks19:42
mriedemit doesn't rebuild from neutron19:43
mriedemi don't think anyway19:43
mriedemit just sends a network-changed event to the compute19:43
sean-k-mooney oh it rebuilds form the vif table in the nova db?19:43
mriedemyes19:43
mriedemwell,19:43
mriedemfrom the info cache19:43
mriedemiow,19:44
sean-k-mooneyok but if the info cache got currpted when the host rebooted your still stuck19:44
mriedemthe network-changed event and _heal_instance_info_cache periodic do the same thing today19:44
mriedemcorrect19:44
mriedemhence the reason for making the periodic actually "heal" from the source of truth, which is neutron19:44
mriedemand not our potentially corrupted cache19:44
sean-k-mooneyso i suggested they detach the missing interfaces and reattach them as a workaround for now to try and force nova and neutron to resync19:45
sean-k-mooneyi think that would work in this case but its not ideal19:45
sean-k-mooneymriedem: anyway thanks. i was sraching my head for the last day or so trying to parse what was going on from incomplete logs but im 98% sure this is it.19:47
sean-k-mooneymriedem: have a safe trip.19:47
*** bigdogstl has joined #openstack-nova19:53
*** bigdogstl has quit IRC19:57
*** sambetts|afk has quit IRC20:00
*** sambetts_ has joined #openstack-nova20:02
*** efried is now known as fried_rice20:09
*** munimeha1 has joined #openstack-nova20:19
*** erlon has quit IRC20:47
mriedemthanks20:48
*** eharney has quit IRC20:53
dansmithmriedem: so you found a test that confirmed the behavior of that thing?20:58
dansmithmriedem: that deletes the consumer?20:58
*** bigdogstl has joined #openstack-nova20:59
mriedemDeleteConsumerIfNoAllocsTestCase is the functional test that covers that case,21:02
mriedemand it looks like a correct test to me,21:02
mriedemcreates 2 consumers each with 2 allocations on different resource classes,21:03
mriedemclears the allocations for one of them and asserts the consumer is gone21:03
mriedemi think we're just hitting a race with the shelve offloaded status change before we cleanup the allocations21:03
mriedembut i've posted a couple of patches to add debug logs to help determine if that's the case21:03
mriedemhttps://review.openstack.org/61701621:03
dansmithokay I'm not sure how we could race and see no allocations but a consumer and get that generation conflict21:04
dansmithit'd be one thing if we thought the consumer was there and then disappeared out from under us21:05
*** lbragstad has quit IRC21:08
*** bigdogstl has quit IRC21:09
*** bigdogstl has joined #openstack-nova21:13
mriedemduring unshelve the scheduler does see allocations21:17
mriedemand it thinks we're doing a move21:17
dansmithokay I thought you pasted a line showing that there was only one allocation going back to placement21:17
mriedemthere are 3 PUTs for allocations21:18
mriedem1. create the server - initial21:18
mriedem2. shelve offload - wipe the allocations to {} - which should delete the consumer21:18
mriedem3. unshelve - scheduler claims resources with the wrong consumer generation21:18
mriedemand when 3 happens, the scheduler gets allocations for hte consumer and they are there,21:18
dansmith...right21:18
*** bigdogstl has quit IRC21:18
mriedemso it uses the consumer generation (1) from those allocations21:18
mriedemthen i think what happens is,21:19
dansmithoh, so it passes generation=1 instead of generation=0, meaning new consumer?21:19
mriedemplacement recreates the consumer which will have generation null21:19
mriedemyes21:19
dansmithokay I see21:19
dansmithI thought you were seeing consumer generation was null or zero or whatever in the third put, but still getting a conflict21:19
dansmithbut that makes sense now21:19
mriedemNov 06 19:48:37.013780 ubuntu-xenial-inap-mtl01-0000379614 nova-scheduler[12154]: WARNING nova.scheduler.client.report [None req-f266a0ff-2840-413d-9877-4500e61512f5 tempest-ServersNegativeTestJSON-477704048 tempest-ServersNegativeTestJSON-477704048] Failed to save allocation for 6665f00a-dcf1-4286-b075-d7dcd7c37487. Got HTTP 409: {"errors": [{"status": 409, "request_id": "req-c9ba6cbd-3b6e-4e5d-b550-9588be8a49d2", "code": "p21:20
mriedemment.concurrent_update",  "detail": "There was a conflict when trying to complete your  request.\n\n consumer generation conflict - expected null but got 1  ",  "title": "Conflict"}]}21:20
mriedemconsumer generation conflict - expected null but got 121:20
mriedemyup - so new consumer but we're passing a generation of 121:20
mriedemfrom the old, now deleted consumer21:20
dansmithcool21:21
mriedemso,21:21
dansmithI wish there was something better to communicate that, but any time we get "expected null" in that case, we should be able to re-try but as a non-move sort of thing21:21
mriedemwe can paper over this by deleting the allocations before marking the instance as shelved offloaded, but that's whack-a-moley21:21
dansmithyeah21:21
mriedemright we need to retry from claim_resources but i'm not sure what's the best way to do that21:22
dansmithand like I said, I think it's not really any better, it just changes the problem21:22
dansmithyeah21:22
mriedemif we do retry that method, the next get for allocations will see there are none and we should be good21:22
dansmithright21:22
mriedemb/c we'll pass consumer_generation=None21:22
mriedemi think i know what we can do21:22
mriedemif we hit21:23
mriedemif 'consumer generation conflict' in err['detail']:21:23
mriedemwe get the allocs again, and if empty,21:23
mriedemwe retry21:23
mriedemeasy peasy21:23
mriedemit's a double get but meh?21:23
dansmithyeah, it's just that it takes us an extra op,21:23
dansmithwhen "expected null" should be enough21:23
dansmithyeah21:23
mriedemi can parse that out of the message if we want..21:23
dansmithI know we can, it's just icky and unfortunate21:23
mriedemwith a TODO for more granular error codes later21:23
dansmithlike all the other cases in there21:24
mriedemright21:24
mriedemi haven't even started packing yet21:24
mriedemlaura is starting to check in on me every 30 minutes21:24
mriedem"this is what i'm wearing all week! god!"21:24
dansmithhah21:24
mriedemplus my mother in law is here,21:25
mriedemso lots of teenage angst memories coming back right now21:25
mriedemthe coffee and metallica doesn't hel[p21:25
dansmithisn't that a good reason to pack and get out?21:25
mriedemi've just holed up in my office21:26
mriedemi'll crank out a patch for this bug and be off21:26
*** eharney has joined #openstack-nova21:43
openstackgerritJack Ding proposed openstack/nova master: Use virt.images.convert_image for qemu-img convert  https://review.openstack.org/61669221:45
fried_ricemriedem: Sanity check, please. The compute manager has a report client via the scheduler client, that's *not* the same as the report client the resource tracker has.21:52
fried_ricewhich means my current SIGHUP doesn't do shit to the RT's cache21:53
fried_riceI need to make the report client a singleton.21:53
mriedemcorrect21:53
mriedemwe have report clients all over the place21:53
mriedemapi, conductor, scheduler21:53
mriedemetc21:53
fried_ricethat's a scroo21:54
fried_ricemriedem: So - make the report client a singleton (per process), or just diddle the compute manager's reset to hit the rt's reportclient instead.21:54
fried_ricef, without knowing what the various ones in the compute manager are used for, is it really safe to make them a singleton?21:55
mriedemthe latter would be a smaller blast area21:56
fried_riceI think I may have actually done this to myself, by removing that LazyLoader21:57
fried_riceI suspect that guy was incidentally singleton-ing.21:57
* fried_rice looks...21:57
fried_ricenah, that should still have been creating separate instances per scheduler client.21:58
dansmithfried_rice: yeah I thought that was making it a singleton a month ago when I was looking at it22:03
dansmitha lot of stuff in nova used to be lazy loaded because.. um, terrible reasons22:04
dansmithlazy loaded or pluggable22:04
fried_ricedansmith: In this case it was supposedly because of a circular import. Whether that was ever really an issue, it isn't now, so I ripped it out. But having just looked, I still don't think it was making the report client a singleton. Care to confirm?22:04
dansmithI think I confirmed that a month ago when I was looking into a seemingly recent memory leak22:05
dansmithso I think it's fine that it's gone22:05
dansmithI agree that randomly making it a singleton now should be done with care22:06
dansmithbut I don't really know how to convince myself that it's okay once its done, tbh22:06
openstackgerritJack Ding proposed openstack/nova master: Use virt.images.convert_image for qemu-img convert  https://review.openstack.org/61669222:06
fried_ricewell, using the RT's report client fixed the problem I was having. So maybe I pretend singleton was never suggested.22:07
fried_riceoh, f, this is gonna break all over the place. I can't see a reason why the compute manager would possibly want or need to use separate report clients. I'd really like to put 'em together. If not making it a singleton, at least using only one of them from the compute manager.22:10
* dansmith ->plane22:10
fried_riceo/22:10
*** eharney has quit IRC22:15
mriedemfried_rice: the compute manager / RT using the same report client is probably fine,22:15
mriedema lot of that compute manager / RT code was cleaned up way back in ocata i think when jaypipes made the RT a singleton that tracked multiple compute nodes,22:16
fried_ricemriedem: Ima put up an independent patch for that22:16
mriedemwhereas before it was 1 RT per compute node22:16
fried_riceah22:16
mriedemthey are very tightly coupled, like how the compute manager passes the virt driver into the RT22:16
openstackgerritJack Ding proposed openstack/nova master: Use virt.images.convert_image for qemu-img convert  https://review.openstack.org/61669222:16
openstackgerritEric Fried proposed openstack/nova master: SIGHUP n-cpu to refresh provider tree cache  https://review.openstack.org/61564622:18
openstackgerritEric Fried proposed openstack/nova master: Reduce calls to placement from _ensure  https://review.openstack.org/61567722:18
openstackgerritEric Fried proposed openstack/nova master: Consolidate inventory refresh  https://review.openstack.org/61569522:18
openstackgerritEric Fried proposed openstack/nova master: Commonize _update code path  https://review.openstack.org/61570522:18
openstackgerritEric Fried proposed openstack/nova master: Turn off rp association refresh in nova-next  https://review.openstack.org/61603322:18
fried_ricelet's see how that goes22:18
fried_ricemriedem: Oh, my removal of lazyload probably reinstated "lockutils spam" mentioned in nova/compute/api.py@25722:21
openstackgerritMatt Riedemann proposed openstack/nova master: Retry on consumer delete race in claim_resources  https://review.openstack.org/61704022:25
mriedemdansmith: jaypipes: fried_rice: ^ bingo bango22:25
mriedemgibi: you too ^22:25
mriedemthe commit message is longer than the code22:25
mriedemand with that i'm off22:29
*** mriedem has quit IRC22:29
*** bigdogstl has joined #openstack-nova22:51
openstackgerritEric Fried proposed openstack/nova master: Rip the report client out of SchedulerClient  https://review.openstack.org/61704222:52
openstackgerritEric Fried proposed openstack/nova master: Rip the report client out of SchedulerClient  https://review.openstack.org/61704222:54
*** spatel has quit IRC22:56
*** betherly has joined #openstack-nova23:02
*** bigdogstl has quit IRC23:03
*** bigdogstl has joined #openstack-nova23:05
*** betherly has quit IRC23:06
*** bigdogstl has quit IRC23:10
*** bigdogstl has joined #openstack-nova23:11
*** munimeha1 has quit IRC23:12
openstackgerritVlad Gusev proposed openstack/nova stable/pike: libvirt: Reduce calls to qemu-img during update_available_resource  https://review.openstack.org/60403923:17
*** elod has quit IRC23:17
openstackgerritMerged openstack/nova stable/pike: Make scheduler.utils.setup_instance_group query all cells  https://review.openstack.org/59984123:18
openstackgerritVlad Gusev proposed openstack/nova stable/pike: libvirt: Reduce calls to qemu-img during update_available_resource  https://review.openstack.org/60403923:18
*** elod has joined #openstack-nova23:18
openstackgerritMerged openstack/nova master: Add recreate test for bug 1799892  https://review.openstack.org/61330423:20
openstackbug 1799892 in OpenStack Compute (nova) rocky "Placement API crashes with 500s in Rocky upgrade with downed compute nodes" [Medium,Triaged] https://launchpad.net/bugs/179989223:20
*** bigdogstl has quit IRC23:24
aspiersmriedem: thanks for the review! Regarding technical debt, my understanding is that the intention is very much for SUSE/AMD to carry on working to flesh out the functionality after implementation of the MVP described in the initial spec, rather than just to dump some half-baked implementation upstream and then vanish ;-) This would include adding support for attestation, migration etc.23:25
aspiersah, he's gone23:26
*** bigdogstl has joined #openstack-nova23:27
openstackgerritVlad Gusev proposed openstack/nova stable/pike: libvirt: Use os.stat and os.path.getsize for RAW disk inspection  https://review.openstack.org/60754423:27
*** s10 has joined #openstack-nova23:28
*** bigdogstl has quit IRC23:32
*** bigdogstl has joined #openstack-nova23:43
*** bigdogstl has quit IRC23:54
openstackgerritMerged openstack/nova master: Mention meta key suffix in tenant isolation with placement docs  https://review.openstack.org/61699123:56
openstackgerritKen'ichi Ohmichi proposed openstack/nova master: api-ref: Add a description about sort order  https://review.openstack.org/61677323:57
*** slaweq has quit IRC23:57

Generated by irclog2html.py 2.15.3 by Marius Gedminas - find it at mg.pov.lt!