Monday, 2022-05-09

gibisean-k-mooney: when you are up. We can go through the open questions in the PCI tracing spec08:24
ralonsohgibi, qq if you know that. For live-migration, all libvirt should have the same "virsh secret-list" ?09:10
gibiralonsoh: o/ good question. I guess you are looking at a situation with encrypted volumes09:14
ralonsohgibi, yes, I've deployed two nodes with devstack and ceph09:15
gibihm, I dont seem to find any code in nova that would move the secret09:18
gibiI'm not a sure how this works but seems like barbican is involved too09:27
gibiralonsoh: as melwitt took over the ephemeral encryption work from lyarwood I expect that she might know more about the volume encryption case as well09:31
ralonsohgibi, thanks, I'm going to try to deploy this env without encription, if possible09:32
opendevreviewJorhson Deng proposed openstack/nova master: Clear the ignore_hosts before starting evacuate  https://review.opendev.org/c/openstack/nova/+/84108909:42
sean-k-mooneydo you mean the ceph secret09:43
sean-k-mooneyif so we expect the operator to distibute that09:43
ralonsohsean-k-mooney, this is in a devstack installation09:43
sean-k-mooneyyep09:44
ralonsohso one compute is requesting its own secret id09:44
ralonsohand the other compute is asking for other09:45
ralonsohand I don't know it that should match the ceph.conf fsid09:45
sean-k-mooneyhum i could try and deploy this and see. but the devstack roles just copy the keyfiels and config to the compute https://github.com/openstack/devstack/blob/master/roles/sync-controller-ceph-conf-and-keys/tasks/main.yaml09:50
sean-k-mooneythen on the compute you set  REMOTE_CEPH=True09:50
ralonsohsean-k-mooney, yes, that's set09:51
sean-k-mooneyhttps://github.com/openstack/devstack-plugin-ceph/blob/master/devstack/lib/ceph#L834-L841=09:51
sean-k-mooneythe secret config seams to be the same09:51
sean-k-mooneyhttps://github.com/openstack/devstack-plugin-ceph/blob/master/devstack/lib/ceph#L244-L259=09:52
ralonsohsean-k-mooney, yeah, it is. In any case, I'll deploy the second compute node again09:52
sean-k-mooneyack09:53
sean-k-mooneyi think from lookign at that that the secure and keyfiles should be the same on all hosts09:53
ralonsohsean-k-mooney, one question09:53
ralonsohin the first node, the fsid is 37...09:54
ralonsohand the "sudo virsh secret-list " is d5...09:54
ralonsohthis is wrong...09:54
ralonsohpffff09:54
ralonsohsean-k-mooney, sorry, I've re-deployed the first controller10:17
ralonsohand CEPH has changed the /etc/ceph/ceph.cong fsid number10:18
ralonsohthat means the virsh secret and the fsid is different now10:18
sean-k-mooneyyes the it will change every time you stack10:21
ralonsohsean-k-mooney, but what ID should I use now?10:22
sean-k-mooneyralonsoh: on a different topic has anyone raised support virtio-failover in neutron10:23
sean-k-mooneyralonsoh: likely the one for the contoler10:23
ralonsohsorry what?10:23
sean-k-mooneythat basically answers my question :) virtio-failover is like a form of automatic bounding10:23
ralonsohah no that I'm aware10:24
sean-k-mooneyyou can declare one virtio device as a failover device if the primary losses connectivity10:24
ralonsohin any case, I'll check it later10:24
ralonsohnope, that is usually done inside the VM10:24
sean-k-mooneyno worries im 99% sure that its not supproted10:24
ralonsohsean-k-mooney, last q10:24
ralonsohabout this fsid10:25
sean-k-mooneyralonsoh: virtio failover allwos qemu to ask the guest virtio driver to do the failover automaticaly in the guest10:25
sean-k-mooneyralonsoh: sure10:25
ralonsohif "virsh secret-list" is one number10:25
ralonsohand in ceph.conf I have another one10:25
ralonsohthen which one should I use in the second comptue?10:25
sean-k-mooneythe secret uuid is the cinder_ceph_uuid10:28
ralonsohyes, and it's generated by devstack-ceph10:28
ralonsohCEPH_FSID=$(uuidgen)10:28
sean-k-mooneyyep so you coudl doble check the cidner config10:28
sean-k-mooneyand determin which is the correct uuid value10:28
sean-k-mooneyill quickly deploy with ceph and see if we can compare10:29
gibiUggla: responded in the manila spec10:33
sean-k-mooneygibi: im just going to grab coffee but we can chat about the pci spec when ever suits 10:38
sean-k-mooneyill be back in 10 mins10:39
gibiI will grab lunch so I will ping you later10:45
sean-k-mooneycool no rush.10:47
gibisean-k-mooney: OK, I'm back11:33
gibisean-k-mooney: so the first question is simple. Will nova create the custom resource class mentioned in the [pci]device_list or we expect the deployer to pre-create that11:38
gibihttps://review.opendev.org/c/openstack/nova-specs/+/791047/4/specs/zed/approved/pci-device-tracking-in-placement.rst#15611:38
gibiI think in the vgpu case nova creates the custom RC11:39
gibias the config does not have a full RC but just some typename11:40
gibiand nova generates the RC name from it11:40
opendevreviewBalazs Gibizer proposed openstack/nova master: DNM: log number of green(thread|let)s periodically  https://review.opendev.org/c/openstack/nova/+/84104011:48
sean-k-mooneygibi: o/ am yes i think nova should create the custom resouce classes in placement11:50
sean-k-mooneythe reason for this is we want to use CUSTOM_<VENDOR_ID>_<PRODUCT_ID>11:51
sean-k-mooneywhen no RC has been specified in the device list11:51
sean-k-mooneyso i think it would be a better end user experince if those custom resouce classes were created automatically11:52
gibiso resource_class=foobar is OK and nova will create CUSTOM_FOOBAR in placement11:52
sean-k-mooneyah are you asking if nova should normalise and prepend the CUSTOM_11:52
gibiyep, as a follow up :)11:52
gibifollow up question11:53
sean-k-mooneyi think that woudl be workable. i would prefer to encurage them to set CUSTOM_<whatever>11:53
sean-k-mooneybut i think its fine to prepend and normalise automatically11:54
gibia bit more user friendly if we normalize and prepend11:54
gibiso I will go with that11:54
sean-k-mooneyyep just so long as we are smart and only prepend when needed11:54
gibiOK, I can make it smart to avoid double custom11:55
gibiOK11:55
gibinext one is future proofing the RC11:55
gibihttps://review.opendev.org/c/openstack/nova-specs/+/791047/4/specs/zed/approved/pci-device-tracking-in-placement.rst#16911:55
gibiI think we support the case today when pci alias and neutorn sriov is configured in the same deployment11:56
gibialso I think it is possible to consume the same type-VF either from alias or from port11:57
gibiso for this case we need an RC name for the type-VF RP that is known before the scheduling for the sriov case11:57
gibiSRIOV_NET_VF could be use for that11:57
gibias that already exists in os-traits11:58
sean-k-mooneyyes you can consume type-vf via alias11:58
sean-k-mooneybecause VF and sriov have nothing to do with networking11:58
sean-k-mooneyyou can have VFs for gpus or ssds11:58
gibitrue11:58
sean-k-mooneythe physical_network tag in the device list11:59
sean-k-mooneyis what marks it as a nic and is required for neutron consumtion11:59
sean-k-mooneyim not sure if we explictly prevent alaise form consuming device with physical_network set11:59
sean-k-mooneybut we could do that going forward i guess11:59
gibiI think we did not prevent it today11:59
sean-k-mooneyack12:00
sean-k-mooneyso for device with physical_network set12:00
gibiwe can simply say that resource_class will not be applicable if physical_network tag is present, and nova will use standard RC for these devices12:00
sean-k-mooneyi think its fine to mark them as SRIOV_NET_VF or SRIOV_NET_PF12:00
gibicool12:00
sean-k-mooneywe would likely track vdpa devices as SRIOV_NET_VF too12:01
sean-k-mooneyalthough i guess that could be SRIOV_NET_VDPA12:01
sean-k-mooneythat actully might make more sense not that i think of it12:01
gibiyeah. but we don't have to decide it now, I just needed to make sure the that the current spec is future proof12:02
sean-k-mooneyso yes making RC and phsynet mutally exclsive i think is correct12:02
sean-k-mooneyyep12:02
gibiOK12:02
gibinext one12:02
sean-k-mooneyso you descoped the current spec to jsut the alias based passthough case yes12:02
gibiyes12:02
sean-k-mooneycool im fine with that by the way12:02
gibiit is still complex enough 12:02
sean-k-mooneyi want the neutron way to work too but it does not need to be in the initall mvp12:03
gibiI just keep an eye on things not to create a dead end with the current spec12:03
sean-k-mooneyyep12:03
sean-k-mooneyok so back ot your next question :)12:03
gibiOK, the next one is simple. With the current proposal the RP is named <hostname>_pci_0000_84_00_012:03
gibiif we follow the pGPUnaming12:04
gibiPGPU naming12:04
gibiis that OK?12:04
gibiwe could have a full normal PCI address if we want as the RP name charset is not restricted12:04
sean-k-mooney am so i was not planning to use the lable for libvirt12:04
sean-k-mooneythe nodedev name pci_0000_84_00_012:05
sean-k-mooneyis not considerd stable by them12:05
gibiahh, good to know12:05
gibithen I think it is better not to rely on it12:05
sean-k-mooneyso i was thinking it woudl be <hostname>_<pci addres in linux format>12:05
gibiso like in DDDD:BB:AA.FF format?12:05
sean-k-mooneyyes12:05
gibiOK, we can do that12:05
sean-k-mooneyis : allowed12:06
gibiyes it is12:06
gibithe RP name is free text12:06
sean-k-mooneyok we could normalise12:06
sean-k-mooneyif not12:06
gibithe traits and RCs are restricted12:06
sean-k-mooneyif we wanted to have pci_0000_84_00_0 by the way i would prefer that nova generated that12:06
sean-k-mooneyrahter then using the value directly form libvirt12:06
sean-k-mooneybut im oke wiht usei the bdf format above DDDD:BB:AA.FF12:07
gibiack12:08
gibinext one12:08
sean-k-mooneyso one thing related to this12:08
gibigo12:08
sean-k-mooneyeven if you add the device to the device list using devname instead of the adress we would still use the pci adress in the RP name yes12:08
gibiyes12:09
sean-k-mooneyi just want to make sure that detail is hidden form placement12:09
sean-k-mooneycool12:09
gibiif we need the devname for any reason during the scheduling we can add that as a trait 12:09
sean-k-mooneyyes we could. currently its not used in the alias or pci request object12:09
sean-k-mooneyso we shoudl not12:09
sean-k-mooneybut if we did a trait woudl be workable12:10
gibiif not needed then we wont add it :)12:10
sean-k-mooney:)12:11
gibiso the next question is related 12:11
gibiwhat traits nova needs to add automatically?12:11
gibionly the ones mentioned in the device_lsit12:11
gibi?12:11
gibior we want to automate things like adding capabilities as traits12:11
sean-k-mooneyah so yes the capablitiy traits shoudl be added in my view12:12
sean-k-mooneythey were intended to be reported to placement orgianly12:12
gibiOK, that make sense12:12
gibido we have in the code somewhere listed what are the capabilities we parse? or we parse verything?12:13
sean-k-mooneyso if i rememebr correctly you were suggeting allow additivie only traits to be listed in the device_list12:13
gibi*everything12:13
sean-k-mooneygibi: im looking for the code now12:13
sean-k-mooneybut we had code to normalise the capablites and report them to placement that ralonsoh wrote in the past12:13
gibisean-k-mooney: yepp the today spec only supports additive traits12:13
sean-k-mooneyhttps://review.opendev.org/q/topic:bp%252Fenable-sriov-nic-features12:14
sean-k-mooneyhttps://review.opendev.org/c/openstack/nova/+/466051 specficically12:15
sean-k-mooneybut you might be able to reuse some of the other work12:15
sean-k-mooneygibi: the traits have already been added to os-traits https://github.com/openstack/os-traits/blob/master/os_traits/hw/nic/offload.py12:16
gibithanks, this helps12:17
gibiso lets report capabilities as traits12:17
gibiand then I will amend the spec to allow removing traits via device_list12:17
gibito disable capability 12:18
sean-k-mooneyyep. right now this only makes sense for neutron nics really12:18
sean-k-mooneysince we dont really gather capablities for other devices12:18
gibiahh12:18
gibiso no generic PCI caps12:18
sean-k-mooneyalthough remote_mannaged might be the excption 12:18
sean-k-mooneywell for the remote managed deviecs we now have the vpd12:18
sean-k-mooneycapablity12:18
sean-k-mooneythat is not yet a trait so maybe we would want to report that12:19
sean-k-mooneygibi: i think for the inital version we could keep it to jsut operator provided traits12:19
sean-k-mooneyif we want to keep it simple12:19
sean-k-mooneythen in the future we can auto discover device capablities and report them if that makes sense12:20
gibiyeah that make sense, it is easy to add later12:20
gibiso keeping traits just additive now as well12:20
gibiOK12:21
gibinext one12:22
gibihttps://review.opendev.org/c/openstack/nova-specs/+/791047/4/specs/zed/approved/pci-device-tracking-in-placement.rst#28812:22
gibiWhat to do if both ``resource_class`` and ``vendor_id`` and ``product_id`` are provided in the alias?12:22
sean-k-mooney good question. i guess we have too options12:23
sean-k-mooneyfirst is consider that an error12:23
sean-k-mooneysecond is use the resouce_class for placment queries12:23
sean-k-mooneyand the the rest for the pci/numa filter12:23
gibithe only use case I can think of is that deployer use a generic RC but then later want to refine the alias via product id12:23
sean-k-mooneymy secret plan is to eventually remove the need for the alias12:24
sean-k-mooneyso over time it woudl be nice if we coudl move to jsut having the resouce class in the alias12:24
gibido you want to go with flavor extra_spec based resource?12:24
sean-k-mooneyyes and no12:25
sean-k-mooneyi current hate that we allow grouping in the extra_specs12:25
sean-k-mooneyso part of me wants to keep the alisa as we can insulate operators form that12:25
sean-k-mooneybut i do kind of like the idea of have just resouce:CUSTOM_<whatever>=112:26
gibiOK, I get the goal that we want to move deployers to RC based alias in the future and if the deployer still want product id based filtering then the deployer can use traits for that12:26
sean-k-mooneyso i think having the RC take prescidence for the placment query and allowign vendor_id and product_id makes sense12:26
gibiOK12:27
sean-k-mooneyso really what i would like is for operators to take devces with the custom resouce class and use the RC name as the "alias"12:28
gibiso we keep the PCIFilter to keep filtering for vendor / product12:28
gibiat least for now12:28
gibithe RC name as alias make sense12:28
sean-k-mooneyya for now althogh in princial i think you could trun it off if we did this right12:28
sean-k-mooneymy idea is12:29
gibiyes if we let the product id filtering case go and no SRIOV in the deployment then we can turn off the PCIFilter12:29
gibineutron based SRIOV12:30
sean-k-mooneydevice_list = some device -> RC gpu_gold12:30
sean-k-mooneyand then you could jsut ask for RC gpu_goal in the alias12:30
sean-k-mooneyright now the reason i dont wnat to go directly to resouce:gpu_gold=112:31
sean-k-mooneyis that intoduced problems wiht vgpu and generic mdev usage12:31
sean-k-mooneyit shoudl be resovlable12:31
sean-k-mooneyi.e. if we see that the resouce does not match any of the generic_mdev types listed in teh config we know tis a pci passthough request12:31
sean-k-mooneybut i tought that would complicate the spec more then need initally12:32
gibiahh yeah12:33
gibithanks for the background12:33
gibinext12:33
gibiI feel that both you and me want to keep the dependent device handling supported. But as stephenfin said, it is a lot of complexity12:34
gibiso just double checking it that you still think this is needed12:34
gibias per https://review.opendev.org/c/openstack/nova-specs/+/791047/4/specs/zed/approved/pci-device-tracking-in-placement.rst#32012:34
sean-k-mooneyhonestly i would like to be able to remove it. but im concerned by the upgrade impact12:34
sean-k-mooneyi do know we have customer that want to dynically choose if they consime a device a a PF or VF when they boot the workload12:35
sean-k-mooneybut its not very reliable today12:35
sean-k-mooneyas in its easy for vms to consume 1 vf on all the devices12:36
gibiI also think that there is many deployment out there that is was used without knowing it. I mean if somebody whitelisted a PF that had VFs then the VFs become scheduleable automatically12:36
sean-k-mooneybasically meaning you can not allocate PF even though your could have allocate the  vfs differntly12:36
sean-k-mooneyright so i think what stephenfin had in mind was if you whitelist the PF and it has VFs we would only expose the VFs12:37
sean-k-mooneywhere as today unless you use the product_id to filter12:37
sean-k-mooneywe expos both the VFs and PFs12:37
sean-k-mooneyif we maintian the current behavior we obvilly need ot dynamically adjst the reserved value12:37
gibiyes, that is the complexity12:38
sean-k-mooneyto emulated the unclaimable state12:38
gibibut it is solveable12:38
gibiI think I will keep this open for bauzas or other reviews to chime in12:38
sean-k-mooneysure12:38
sean-k-mooneywe decided to reduce flexiblity for cpu pinning12:39
sean-k-mooneywith isolate12:39
sean-k-mooneyand we know that not everyone was happy with that12:39
sean-k-mooneywe can elect to do the same here but we need to be deliberiate about it and comunicate it well if we want to force this change12:40
ralonsohsean-k-mooney, sorry, I was having lunch12:40
ralonsohwhat do you need?12:40
sean-k-mooneyif we can live with the complexity then we proably shoudl keep it12:40
sean-k-mooneyralonsoh: i found it12:40
gibiOK, I will plan with the complexity12:40
ralonsohah perfect12:40
sean-k-mooneyralonsoh: https://review.opendev.org/q/topic:bp%252Fenable-sriov-nic-features12:40
sean-k-mooneyralonsoh: your code for tracking pci device in placment12:40
sean-k-mooneyralonsoh: we are just discussing the spec to enable it. gibi will be taking on that feature12:41
ralonsohperfect12:41
gibiralonsoh: https://review.opendev.org/c/openstack/nova-specs/+/791047/4/specs/zed/approved/pci-device-tracking-in-placement.rst this is the spec if you are interested :)12:41
ralonsohsure12:41
gibisean-k-mooney: so I have one more open question12:42
sean-k-mooneygibi: go for it12:42
gibiupgrade12:42
gibiobviously rolling upgrade is a pain12:42
sean-k-mooneyah well you burn down the datacenter and build a new one in the ashes12:42
sean-k-mooneyobvisouly the least  painful approch12:42
gibiin the PCPU case we did a fallback query12:42
gibito allow schedling to not-yet upgraded computes12:43
sean-k-mooneyyep12:43
sean-k-mooneywe could make the prefilter configurable12:43
gibiI'm not sure how we did the allocation in that case12:43
gibibut12:43
gibiin the PCI case if we select a host based on the fallback query then on that host the scheduler will not allocate PCI devices in placement12:43
sean-k-mooneyright so i would not use a fallback12:44
sean-k-mooneyby default we shoudl reprot the inventores to placemnt12:44
sean-k-mooneyand ahve a prefileter12:44
sean-k-mooneythe prefilter would add the pci device request to the query12:44
sean-k-mooneyand we disable it by default in zed12:44
sean-k-mooneythen enable it by default in AA12:44
sean-k-mooneyso you would rolling upgrade to Zed 12:45
sean-k-mooneythen enable the prefilter once all host are upgraded12:45
sean-k-mooneyyou likely would have to then do a heal-allocation like command to update the allcotiosn of existign instances12:45
gibiahh i see12:46
sean-k-mooneywe could also have a nova-status check12:46
gibiso we report devices but we don't allocate yet12:46
sean-k-mooneyyep12:46
gibithen when every compute is ready we do a heal and then start allocating12:46
sean-k-mooneyyes using the claim in the pci_devices table12:47
gibiyepp we keep using the claim and the pci_device table12:47
gibianyhow12:47
gibias that tracks exact VF PCI addresses12:47
gibiPlacement wont12:47
sean-k-mooneyyep so form the contolers12:47
sean-k-mooneywe will have all the info in the pci_devices table to heal the allocations12:48
sean-k-mooneysince we also have the parent adresses12:48
sean-k-mooneywe can constuct the RP names12:48
gibiyepp12:49
sean-k-mooneywe could consider12:49
sean-k-mooneyif we can activate teh filter based on min compute service version12:49
sean-k-mooneyif we were to do that we would likely need the compute agent to heal the allcoations automaticaly12:50
sean-k-mooneyperhaps on starup or in the upsadate_avaiable_resouces periodic task12:50
sean-k-mooneyim not sure if we want that level of compleixty but we already do reshapes in init_host12:51
sean-k-mooneydo you think that is too much "magic"/complexity 12:52
sean-k-mooneyit would make the operator experince much nicer as it woudl just start working once everythin was upgraded12:52
gibithis reshap will be just RP creation, we won't move things12:52
gibiso calling that code from periodic feels OK12:52
gibiif we have that then it is safe to enable the prefilter automatically12:53
gibiby the compute min version12:53
sean-k-mooneyright so when the compaute-agent start with the new code. the first time it create the inventroeis it would also update the allcoations12:53
sean-k-mooneyand then the prefilter woudl activate once all computes are upgraded12:53
sean-k-mooneybased on min verion check12:54
sean-k-mooneythe proably i see with this would be move operations before the prefilter is enabled12:54
sean-k-mooneyunless we have it continue to heal12:54
sean-k-mooneyuntill the min version reaces the required version 12:54
sean-k-mooneythere would be some inconsitency for a time but the pci_tracker would enforce the corret behaivor with regards to not over subscibing12:55
gibiyeah we have the pci tracker and pci claim as a fallback12:55
gibiso we can move even if the prefilter is disabled12:56
gibijust have to have a way to heal the placement allocation12:56
gibieventually12:56
sean-k-mooneyyep12:56
gibiOK, I think I got my answers12:56
gibithank you for your time12:56
gibiI really appreciate it12:56
sean-k-mooneywhich again can use current and min service version to disable the healing when its not needed12:57
sean-k-mooneyno worries12:57
sean-k-mooneyim excited to see this moving forward12:57
sean-k-mooneywill you summerise this in the spec12:57
sean-k-mooneyperhaps like to the irc logs12:57
sean-k-mooney*link12:57
gibiI will do the summary12:57
gibiand linking to the log12:57
gibithen I will respin the spec12:57
gibiand trim the questions12:58
gibiI'm excited to stat coding up some of these in nova and watch them fail in the func env :)12:58
gibiit will be fun12:58
sean-k-mooneygibi: on a related not you reviewed Uggla spec. there was kind of an open question regarding updating hte AZ when you specify a host did you weigh in on that.12:59
gibiI saw it and I think it was settled, I had no objection. But then I will doulecheck12:59
gibidoublecheck12:59
sean-k-mooneygibi: ack ill try and review it again shortly so12:59
sean-k-mooneygibi: on a more selfish note i could also use your input on something else but its not super urgent https://review.opendev.org/c/openstack/nova/+/841017/1/nova/virt/libvirt/driver.py13:01
sean-k-mooneyi dont think that is 100% correct but i works for vdpa i need to test it with VFs and other vnic-types13:02
* gibi clicks13:02
sean-k-mooneybasically we are curently unpluging neutron interface using _detach_pci_dev for suspend13:02
sean-k-mooneythat does not work for vdpa and im pretty sure it does not work in general13:03
sean-k-mooneyso i need to verify that and file a bug13:03
gibiI never tried suspend with PCI / neutron SRIOV. So I neither confirm now deny that it works13:07
sean-k-mooneyit used to but its been a very long time since i checked it.13:07
sean-k-mooneyso ya i need to test it with differnt backends13:08
gibibut your comments seems valid that if something is an interface then that cannot be detached as a hostdev13:08
sean-k-mooneyi have the ablity to test hardware offloaded ovs and sriov at home and i still have the servers i was usign for vdpa although ill be giving those abck today13:09
sean-k-mooneyso i can see if i can test the differnt combinations13:10
sean-k-mooneyi think self.detach_interface(context, instance, vif) shoudl work however in all cases13:10
sean-k-mooneyi dont really know why we have sepcial handelign for the host dev elements13:10
sean-k-mooneydetach_interface13:11
sean-k-mooneyis ment to be the abstraction here and its what is called when we call detach form the api13:11
gibiyepp detach inteface dynamically use hostdev or interface config object 13:12
sean-k-mooneyso i think i can just factor out the common code form https://github.com/openstack/nova/blob/master/nova/virt/libvirt/driver.py#L9811-L9829=13:12
sean-k-mooneymaking the migrate_data optionall effectivly13:13
gibiyepp13:13
gibithat seems doable13:13
sean-k-mooneybasically i woudl jsut take in a list of vifs13:13
sean-k-mooneyso what im wonderign is it better to adapt the curent fucion as i did in the wip patch13:14
sean-k-mooneyor jsut do the refactor13:15
sean-k-mooneyand call detach_interface13:15
sean-k-mooneyvia  _detach_direct_passthrough_vifs13:15
gibiI would do the refactor and call detach_inteface but I'm biased with the detach_interface code :D13:15
sean-k-mooneywell see i trust the detach_inteface code more13:16
sean-k-mooneyits better tested13:16
sean-k-mooneyok thanks ill try and confirm my sepculation that suspend was broken and file a bug13:17
gibicool13:17
sean-k-mooneyone thing i need to bring up in the team meeting tomorrow is how to track the vdpa work13:18
sean-k-mooneyhttps://review.opendev.org/q/topic:bug%252F1970467 the non WIP patch is the bug fix13:18
sean-k-mooneyfor the move operation that actully work13:18
sean-k-mooneythe next 3 add attach/detach, suspend and hotplug live migration13:19
sean-k-mooneyi feel like the last 3 shoudl be a specless blueprint or maybe a small spec13:19
*** dasm|off is now known as dasm13:27
gibiI'm OK with both direction. If there is no API change then I'm fine with specless but if you have open questions then those are easy to discuss via a spec13:27
sean-k-mooneythere are no api change other then me removing the api block on the operation however i think i should be adding a compute service version bump for live migration13:29
sean-k-mooneyto supprot rolling upgade13:29
sean-k-mooneyi dont have that in the wip code13:29
gibiI think this still can fly as specless13:31
sean-k-mooneyack that is what i was hoping but if other felt differently i just wanted to get the spec up quickly13:32
gibiyeah it is worth to ask13:32
opendevreviewBalazs Gibizer proposed openstack/nova-specs master: PCI device tracking in Placement  https://review.opendev.org/c/openstack/nova-specs/+/79104714:36
gibisean-k-mooney: updated according to our discussion ^^14:36
* Uggla plays with tempest today. Unshelve to host should have a tempest test.16:36
opendevreviewribaudr proposed openstack/python-novaclient master: Microversion 2.91: Support specifying destination host to unshelve  https://review.opendev.org/c/openstack/python-novaclient/+/83165116:55
opendevreviewMerged openstack/nova stable/xena: Test aborting queued live migration  https://review.opendev.org/c/openstack/nova/+/83614517:24
opendevreviewMerged openstack/nova stable/xena: Add functional tests to reproduce bug #1960412  https://review.opendev.org/c/openstack/nova/+/83614617:24
opendevreviewMerged openstack/nova stable/xena: Clean up when queued live migration aborted  https://review.opendev.org/c/openstack/nova/+/83614717:24
opendevreviewMerged openstack/nova stable/yoga: Retry in CellDatabases fixture when global DB state changes  https://review.opendev.org/c/openstack/nova/+/84073419:05
opendevreviewArtom Lifshitz proposed openstack/nova master: Reproduce bug 1952745  https://review.opendev.org/c/openstack/nova/+/84117021:26
artomI'm kinda proud of ^^21:26
opendevreviewArtom Lifshitz proposed openstack/nova master: Reproduce bug 1952745  https://review.opendev.org/c/openstack/nova/+/84117021:36
opendevreviewArtom Lifshitz proposed openstack/nova master: Reproduce bug 1952745  https://review.opendev.org/c/openstack/nova/+/84117021:41
*** dasm is now known as dasm|off22:16
opendevreviewmelanie witt proposed openstack/nova stable/ussuri: Define new functional test tox env for placement gate to run  https://review.opendev.org/c/openstack/nova/+/84077123:55

Generated by irclog2html.py 2.17.3 by Marius Gedminas - find it at https://mg.pov.lt/irclog2html/!