Wednesday, 2023-05-03

opendevreviewAmit Uniyal proposed openstack/nova master: Allow swap resize from non-zero to zero  https://review.opendev.org/c/openstack/nova/+/85733907:52
dvo-plv_Hello, All Could you pelase help me with file naming08:09
dvo-plv_opt/stack/glance/releasenotes/notes/zed-milestone-3-3e38697ae4677a81.yaml  where should I get hash ( zed-milestone-....yaml) in this file ?08:09
sean-k-mooneyits auto geneerated using a tool08:31
sean-k-mooneytox -e venv reno new zed-milestone-308:31
sean-k-mooneyreno is the tool we use to create releasenotes08:32
sean-k-mooneydvo-plv_:^08:32
dvo-plv_thank you08:37
opendevreviewDanylo Vodopianov proposed openstack/nova master: Packed virtqueue support was added.  https://review.opendev.org/c/openstack/nova/+/87607511:33
opendevreviewAmit Uniyal proposed openstack/nova stable/wallaby: fup: Print message logging uncaught nova-manage exceptions  https://review.opendev.org/c/openstack/nova/+/87733411:53
opendevreviewDanylo Vodopianov proposed openstack/nova-specs master: VirtIO PackedRing Configuration support  https://review.opendev.org/c/openstack/nova-specs/+/86837711:54
dvo-plv_gibi: I have resolved your comment at this bp https://review.opendev.org/c/openstack/nova-specs/+/86837712:15
gibidvo-plv_: look good to me12:18
opendevreviewDanylo Vodopianov proposed openstack/nova-specs master: VirtIO PackedRing Configuration support  https://review.opendev.org/c/openstack/nova-specs/+/86837712:23
dvo-plv_I fixed pep8 validation. Should I do something more for further bp activities?12:38
sean-k-mooneyim just waitign for ci to report back im still happy with it so ill readd +2 once its green12:38
sean-k-mooneyoh its already repoted back12:39
sean-k-mooneygibi: care to send it on its way https://review.opendev.org/c/openstack/nova-specs/+/86837712:40
gibisean-k-mooney, dvo-plv_ : done12:43
dvo-plv_Great. So spec file is approved. Then we have to wait for code review  ? https://review.opendev.org/q/topic:bp%252Fvirtio-packedring-configuration-support12:47
opendevreviewMerged openstack/nova-specs master: VirtIO PackedRing Configuration support  https://review.opendev.org/c/openstack/nova-specs/+/86837712:58
sahidquick question regarding host aggregate and AZ13:05
sahidwe have noticed that when we add a compute host, that one is always considered in the AZ nova (the one that is by default configured in default_avalibility_zone)13:06
sahidit seems that if we don't want this behavior to happen, we have to first add the host to the correct aggregate and then start nova-compute service13:07
sahidthe problem is that, it seems not possible to add in an aggregate a host that has not been "registered" by a service like nova-compute, right?13:08
sean-k-mooneysahid: that not how it works13:09
sean-k-mooneyif you dont add a host to an aggreate with az metadta13:09
sahidso during deployement of a new compute host we always have that window where the host is considered in AZ nova (this az does not exist in our deployement) until that we set the host to the correct aggregate13:09
sean-k-mooneyif you dont add a host to an aggreate with az metadta then its is considerd part of default_availability_zone13:10
sahid(a aggregate with ZA metadata, yes)13:10
sahidsean-k-mooney: yes we are in the same page13:10
sean-k-mooneysahid: yes ther is no way to atomically ahve a comptue node com up and register in an az13:11
sean-k-mooneysahid: if you want to prevent ti beign selected you can make the service register in a disabeld state13:11
sahidi think it may be possible to achieve that if we allow to add an host that does not yet exist in a aggregate (that has az metadata), right?13:12
sahidsean-k-mooney: ahh interesting point13:12
sahidwhat do you mean by "selected"?13:12
sahidI don't want that nova az appears13:12
sean-k-mooneyconsiderd as a valid host for scheduling13:12
sahidbecause it's confusing for users13:13
sahidah...13:13
sean-k-mooneythat is not somethign you can do currently13:13
sahidno no, from users perspective is not really good to see this "nova" az appearing and disapearing13:13
sean-k-mooneyhttps://docs.openstack.org/nova/latest/configuration/config.html#DEFAULT.enable_new_services13:13
sean-k-mooneysahid: well many deployment just use the nova az13:14
sean-k-mooneya new feature could be added to hide az perhaps13:14
sean-k-mooneybut its workign as intended13:14
sahidand what about to give ability to add a host that does not yet exists to an aggregate?13:14
sean-k-mooneyi.e. we coudl add a config option to hide the defautl az or add a metadata parmater perhasp13:15
sean-k-mooneysahid: i dont really like that option13:15
dvo-plv_gibi, sean-k-monney: Great. So spec file is approved. Then we have to wait for code review  ? https://review.opendev.org/q/topic:bp%252Fvirtio-packedring-configuration-support13:15
sean-k-mooneyi could maybe see adding a cofnig option to let the compute node know what az to regester in by default13:15
sean-k-mooneydvo-plv_: ya but at this point provided all the tests ectra are in place and we are ok with the code all the admin is done13:16
sahidthis already exist right? https://docs.openstack.org/nova/latest/configuration/config.html#DEFAULT.default_availability_zone13:16
sahidbut this work for deployment with one az13:16
sahidi had feeling that configuring the host to the aggregate, then starting nova-compute on it would have been a not so bad solution13:17
sahidnova-compute at start could have been able to be registered in the right way13:18
dvo-plv_sean-k-monney: Do you mean additional test for packed option from our side, or zuul verification ? 13:18
sahidwe need to find a solution so when we start nova-compute for the first time it is be able to determine is AZ properly13:19
sean-k-mooneysahid: well compute agents are not ment to kwno what az they are in realy13:21
sean-k-mooneysame as they are not ment to really knwo what cell they are in13:21
sean-k-mooneysahid: default_availability_zone is what is makign it show as nova13:22
sean-k-mooneybut that is not used by the compute agent13:22
sean-k-mooneythe nova az does not actully exsit as a host aggreate either13:23
sean-k-mooneyor rather the default_availability_zone does not actully exist in the database13:23
sean-k-mooneythat parmater is just read by the api to use as a default when the compute service is not associated with an az. bauzas  can correct me if that is wrong13:24
bauzason a meeting now13:24
bauzasexplain me quickly the issue, please ?13:24
sean-k-mooneybauzas: sahid  would like a way for compute agents to auto register to an az13:24
sean-k-mooneyso that they dont end up in the default_availability_zone i.e. nova13:25
sean-k-mooneybecause they dont wnat that az to come and go as they are provisioning new servers13:25
sean-k-mooneywe coudl adress that multiple ways but it would be a new feature in any case13:26
bauzasso, automatically adding a compute service to an aggregate ? well, some script can do it13:26
sean-k-mooneyit can yes which is what we expect to be done now13:26
bauzasnothing I see for needing Nova to do it automatically13:26
sean-k-mooneythere is a small window btween the compute node starting and being able to add it at the api13:26
bauzasthey can disable it by default, right?13:27
sean-k-mooneythe az will still show up while the ocmptue is disabled13:27
sean-k-mooneyits the default az they want to hide not nessisarly the compute node13:27
dansmithbut aggregates are in the api database13:27
bauzassure, but that's an operator point 13:27
dansmithauto-registering to an az would require compute->conductor->apidb right?13:27
sean-k-mooneydansmith: yep which is why i said this is not really somthing the compute agent should know about via config13:27
sean-k-mooneydansmith: ya it woudl need an up call13:28
dansmithokay good, I thought you were arguing _for_ doing that13:28
sean-k-mooneyno not really13:28
bauzasagain, any script can do it, I don't see *why* nova should support it 13:28
dansmithit also breaks several things: 1. computes don't know their az and 2. no upcalls13:28
sean-k-mooneyyep13:28
sean-k-mooneyso either we add config drive api behavior ( allow hiding the default az via a api config option)13:29
bauzasplus the fact that we check race conditions in the API, not by the compute service13:29
sahidit's why i was asking to give ability for operator to add in aggregate host that does not yet exist13:29
sean-k-mooneyor we make the default az a reall az/hostaggerte in the db13:29
sean-k-mooneyor we just dont change things13:29
sahidbauzas: it's about an user experience13:29
bauzaseverytime you're adding a host to an aggregate, we synchronously check that AZs are correct13:30
sean-k-mooneysahid: we technially dont have a forign key that would prevent that so we could13:30
sahidthey see nova appearing13:30
sean-k-mooneyif we comine that with the stable uuid feature13:30
sean-k-mooneyit might be workable13:30
sean-k-mooneyie preallcoat the uuid that would be used for the comptue somehow but that is still kind of messy13:30
bauzassahid: I think we explained that by default nova supports one AZ13:30
bauzasthat doesn't mean that nova will *check* hosts 13:31
sean-k-mooneyi think we check that there is a host_mapping today before allowing you to add a host to a host aggreate13:32
bauzasonly the operators know whether the AZ that the user uses is actually a right AZ or just a fake one13:32
dansmithsahid: I assume the goal is so that an operator adding a new compute can tell the compute where it wants to plug in and have that happen when it first registers itself, instead of adding, registering, and then having to go back and add it to an aggregate right?13:32
bauzasas we also loudly say that operators shall NEVER create 'nova' AZs, I don't see the problem13:32
sahidsean-k-mooney: yes today, if you try to do that with an host that does not "exist" you receive an error13:34
sahiddansmith: yes it's mostly that, today we start the service and then add the host to the correct aggregate, during this window, nova appears in the list of AZs13:37
sahidwe want to do the opposite (if possible), by configuring the host aggregate and then starting the service13:38
dansmithsahid: so is it just a "day 0" thing, or do you want compute to be able to change its own az if you change its config and restart?13:38
sahidin that way we avoid that "nova"13:38
sahidjust a day 0 thing13:39
dansmithso, what about something like allowing metadata on a service, which you can configure in nova.conf for the compute, and then have the discover_hosts thing look for a "desired_az" or "desired_aggs" key and do the needful during discovery (which has to happen anyway)?13:40
dansmiththat way no upcall, and it's just "desired" and "metadata" and not "this is my az"13:40
dansmithdoesn't work after day zero13:40
dansmithwe could probably use metadata on a service for other things13:41
dansmithI'd have to think some more.. I'm trying to listen to a meeting right now too13:41
sean-k-mooneydansmith: we could also avoid the upcall by having the nova comptue just call the nova api driectly 13:42
sahidthis should resolve our case for sure and is a bit like what sean-k-mooney mentioned13:42
dansmithsean-k-mooney: it needs admin credentials though which would be unfortunate I think13:42
sean-k-mooneywell at least teh service role13:42
sean-k-mooneybut ya13:42
dansmithyeah, not full admin hopefully, but still13:43
sean-k-mooneywhere where you thinking of stashing the desired az13:43
sean-k-mooneyi dont think we have a db table we coudl use for that in the cell db currenlty13:43
dansmithwe'd need to add a generic metadata blob to service13:43
dansmithnope13:43
sean-k-mooneythats also not partically nice but ok13:43
dansmithnot nice because it requires db changes? or not nice for other reasons?13:44
sean-k-mooneythe db changes13:44
sean-k-mooneyif its just for this13:44
sean-k-mooneyif it was like instance_system_metadta13:44
sean-k-mooneyand we had other usecases i woudl care less13:44
dansmithyeah, well, I mean.. something will require changes.. but I figure we might be able to use it for other things too13:45
sean-k-mooneyso you prefer doign ti via discover host to limit the creds13:45
dansmithwell, we already have a discovery process to do very similar things13:45
dansmithlemme finish listening to this meeting for the next 15 minutes and then I can think a little more13:46
sahidthis should resolve our case for sure and is a bit like what sean-k-mooney mentioned13:51
sahidoops sorry13:51
sahidi've create bug report here, https://bugs.launchpad.net/nova/+bug/2018398 in case that we consider it's as valid and we want to do something to improve users experience :-)13:56
dansmithsahid: s/users/operators/ so we're clear which "users" we're talking about13:57
dansmithbut yeah, I think this would be a nice behavior for operators, as long as we can do it within the other constraints and design points we have13:57
bauzassahid: so thanks for the bug report, now I better understand your concern13:57
bauzaslemme check one thing tho13:58
bauzassahid: so, say your operator defined two AZs : AZ1 and AZ213:59
bauzashostA is in AZ1, hostB in AZ213:59
sahiddansmith: wait, perhaps there is something that we are not doing correctly, but it's really our users which see those AZs in our interfaces13:59
bauzasnow, he gonna add hostC13:59
bauzasonce he registers hostC, then it appears a third AZ in the AZ list which is 'default_availability_zone' ie. 'nova'14:00
bauzasI get the problem14:00
sahidwhen they create an instance they can select the AZ14:00
bauzasnown my question 14:00
sahidbauzas: yes rights14:00
bauzaswhy isn't the default availabilty zone be either AZ1 or AZ2 ?14:00
bauzaswe just use that value to fake AZs14:01
dansmithsahid: I know users see AZs, but users will not see this compute node feature14:01
bauzasbut this is not meaning that instances are automatically scheduled to that ZA14:01
bauzasAZ*14:01
sahiddansmith: ack14:01
bauzasthe default value for sending an instance to an AZ is default_schedule_zone14:01
bauzaswhich is None by default14:02
dansmithbauzas: oh right, I always forget about this difference14:02
dansmiththere are two conf knobs right?14:02
bauzasyes14:03
bauzashttps://docs.openstack.org/nova/latest/configuration/config.html#DEFAULT.default_schedule_zone is for pushing new instances to a specific AZ automatically14:04
bauzashttps://docs.openstack.org/nova/latest/configuration/config.html#DEFAULT.default_availability_zone is just for showing a AZ if none exists14:05
bauzasin a context where some AZs already exist, the solution is just to set the latter option to an existing AZ14:05
dansmithbauzas: but it doesn't actually put the compute there right?14:05
bauzasno14:05
dansmithit's like faked in the api?14:05
sean-k-mooneyyep14:05
bauzas" This option determines the default availability zone for ‘nova-compute’ services, which will be used if the service(s) do not belong to aggregates with availability zone metadata."14:05
sean-k-mooneyits not used by the compute service at all14:06
sean-k-mooneyyou could have discover host use it if its set to a real az14:06
dansmithyeah, so now does that help? I mean, it might help the "don't expose some other AZ for a window after adding the compute" but it doesn't help the mechanical operations of just dealing with the compute itself while adding14:06
bauzasdansmith: there are two different concepts that are hard to understand14:07
dansmithsean-k-mooney: but then discover would add all computes to one az, not the one the compute should be in14:07
bauzasthere is a the AZ API 14:07
sean-k-mooneydansmith: ya its not greate to use that directly 14:07
bauzaswhich is just showing the list of aggregates with a AZ metadata + the default az value14:07
bauzasand then, the second concept is the scheduling thing14:07
dansmithyeah, I understand the default_schedule_zone part14:08
bauzasdansmith: if you read https://bugs.launchpad.net/nova/+bug/2018398 sahid is concerned by the fact 'nova' is presented to endusers when they do nova availability-zone-list14:08
sean-k-mooneywhat happens today if you set default_avaiablity_zone=internal14:08
dansmithbauzas: right14:08
sean-k-mooneyinternal is not show to endusers unless they are admin14:08
bauzasdansmith: I'm just saying to modify this value to an existing AZ value, eg. AZ114:09
dansmithbauzas: although I'm not sure how that happens even, because the compute isn't actually in that az14:09
bauzasdansmith: that option is purely API-based14:09
sean-k-mooneyif  default_avaiablity_zone=internal is accpeted today14:09
dansmithbauzas: right, that helps not expose 'nova' for a minute while you fix it14:09
sean-k-mooneythat woudl hide the compute until the admin added them to a real az14:09
bauzascorrect ^14:09
dansmithbauzas: but that's only one aspect of the problem (IMHO)14:09
bauzasdansmith: well, lemme clarify then14:09
dansmithbauzas: the second aspect is that you add a compute, you wait for it to register, you wait for it to be discovered, then you go and add it to the right az14:09
dansmiththat's stuff that a deployment tool has to do a bunch of polling and looping over, but it really shouldn't14:10
sean-k-mooneydansmith: if we remvoed the exists check in the host aggrate add14:10
dansmithif it could say "this should be in AZ foo" then all that stuff can happen automatically and without a bunch of slow steps in the process14:10
bauzasdansmith: technically you don't add a host to an az, but to an aggregate14:10
sean-k-mooneyand we leveraged the fact we can right a uuid for the comptue to use14:10
dansmithsean-k-mooney: yeah but then people can typo things when they add them, so I don't love that14:11
dansmithbauzas: I know :)14:11
bauzasdansmith: and an operator can add a host to an aggregate as soon as it's registered in the services table14:11
sean-k-mooneyi could see the workflow being the installer generate a uuid for the host, then adds it to the righ host aggrate and then deploys the host14:11
sean-k-mooneydansmith: ya that one of the reasons we dont allow it today14:11
dansmithsean-k-mooney: but you don't add compute nodes to aggregates, you add services right?14:11
sean-k-mooneyah you are right its the service14:11
bauzasdansmith: typos can't append, we query the record based on the service name14:11
sean-k-mooneywhich is why this does nto really work with ironic today14:12
bauzashappen*14:12
dansmithbauzas: sean-k-mooney is suggesting we drop that requirement14:12
sean-k-mooneydansmith: well sahid was previously14:12
bauzasagain, why ?14:12
sean-k-mooneybauzas: to allwo you to pregreister the service to host aggrate14:12
sean-k-mooneybefore its created14:12
bauzascan't a script loop over a service list and do the aggregate add after ?14:13
dansmithI'm just thinking of how customers are sometimes afraid to run stack update on a running cloud, just like I'm afraid to re-run my own shoddy ansible sometimes :)14:13
dansmithand how making the act of adding a new compute more of a compute-focused thing than a deployment-focused thing would likely be a welcome pattern for people14:13
bauzassean-k-mooney: we would soften the requirements for the very little benefit of pre-adding a compute to an aggregate14:13
sean-k-mooneybauzas: you can but if you do that in cron and a vm is booted that request nova in that interval then it will be broken14:14
dansmithbauzas: I agree, I don't like that idea14:14
sean-k-mooneyonce an instance gets schdeuled to the host it cant move to a real az14:14
bauzasI like the idea to only add something schedulable if that thing can actually really schedule my stuff14:14
sean-k-mooneycould we hide the nova az if all compute service in it were disabled14:15
bauzasthe second it got added to an aggregate, it can be scheduled14:15
bauzasagain, why ?14:15
sean-k-mooneythen you can set the config option to only register them as disabled14:15
bauzaswe would put a lot of preconditions and added complexity for a very little gain14:15
bauzasand adding a compute isn't really a daily operation14:15
sean-k-mooneybauzas: sure its little gain but its a ligitmate pain point for there end customer apperenlty14:15
sean-k-mooneywell it depend on your scale14:16
bauzassean-k-mooney: the bug report wasn't complaining about it14:16
dansmithwe already have a task that has to run for a new compute (discover) which can either be automated in scheduler, called via cron, or manually during a deployment.. its purpose is to do things in the api database, so it feels like a very natural thing to have that step also do this, which is placing the compute in the right grouping14:16
bauzassean-k-mooney: the bug report was about to prevent 'nova' to show off 14:16
sean-k-mooneybauzas: sahid said it caused confustion for ther custoemr.14:16
bauzasdansmith: I'm confused, everything is API-driven14:17
dansmithsahid: is the confusion of showing the wrong AZ the only thing you care about?14:17
dansmithbauzas: adding a new compute is not api-driven14:17
dansmithbauzas: the exception to that is needing to put it into the right az14:17
sean-k-mooneybauzas: expanding on ^ you can delete compute services from the api, you can only create them via rpc via the conductor14:18
sean-k-mooneybauzas: we do not have a create endpoint https://docs.openstack.org/api-ref/compute/#compute-services-os-services14:19
bauzasOK I admit the feature gap14:20
sean-k-mooneydansmith: addign a create endpoint woudl also work i guess. when the compute starts it would find the existing record14:20
sean-k-mooneyand ocne the record is created you could perhaps add the servicew to an aggreate14:20
dansmithsean-k-mooney: to do that we'd have to expose the concept of a cell to the API which I'm -3 on :)14:20
bauzasbut I'm very reluctant to make add_host_to_agg() less strict14:20
sean-k-mooneyoh ya we would14:20
dansmithmaking this more manual is also not something I'm in favor of14:21
sean-k-mooneycase this is in the cell db14:21
sean-k-mooneyso ya thats not an option relaly14:21
dansmithand pre-creating hosts makes that more manual, and much more likely to be wrong, given what we know about hostnames :)14:21
sean-k-mooneyvery true14:21
dansmithI'd like to make this more automatic14:21
dansmithwhich may or may not be what sahid really wants, but making it more automatic will *also* solve sahid's concern14:21
bauzaswithout any upcalls for sure14:21
dansmithyep14:21
bauzaswhich is quite a chicken-and-egg problem 14:22
sean-k-mooneythe thing is we would like this to be more atomatic while also not havign upcalls or having the comptue specificly knwo what az its in14:22
sean-k-mooneybecause that shoudl still be changable vai the api14:22
dansmithwell, I don't see "desired_az" as breaking the "knowing what az I'm in" rule14:22
dansmithyeah of course14:22
* sahid sorry i had to join a meeting14:23
sean-k-mooneydesired_az to me woudl be like the initial cpu allocation ratio14:23
dansmithjust like our default allocation ratios - they're used if we have to set a default, otherwise it's what is in placement that matters14:23
dansmithyep14:23
sean-k-mooneyrigth somethign that is used once and only once14:23
dansmithexactly14:23
bauzasfwiw, AZs are already a complicated matter to understand14:23
sahiddansmith: yes our biggest concern was the az showed14:23
bauzasdo we really want to introduce another concept ?14:23
dansmithbauzas: what other concept?14:23
bauzas'I, as compute, will be part of an AZ eventually'14:24
dansmithwell, that's true regardless of what we do here :)14:24
sean-k-mooneythat a concept for the installer/deploer 14:24
bauzasreconciliation, if you prefer14:24
sean-k-mooneynot really for the end users and only partly for the admin14:24
bauzasdansmith: see, sahid's problem is the user-facing AZ list problem14:25
bauzasnot the day-2 operational problem of adding a host to an AZ14:25
dansmithbauzas: I've said several times, I get that.. I think there are two benefits to a solution here, his being one14:25
sean-k-mooneyim still not sure if that can be adress by settign default_avaialblity_zone=internal14:25
bauzassean-k-mooney: internal has a very different meaning today14:26
sean-k-mooneyhttps://docs.openstack.org/nova/latest/configuration/config.html#DEFAULT.internal_service_availability_zone14:26
sean-k-mooneybauzas: it does but its not show to non admins14:26
sean-k-mooneyand sahid does not want this to be show14:26
bauzasinternal skips all computes14:26
sean-k-mooneyi know14:26
bauzasagain, this show problem is resolvable by default_availability_zone14:27
sean-k-mooneybut do we have somethign to prevent you setting https://docs.openstack.org/nova/latest/configuration/config.html#DEFAULT.default_availability_zone to internal14:27
bauzasah, no14:27
bauzasbut I can quickly check14:27
sean-k-mooneyi would be ok with extending the concep of the intenal az to have compute in it if you set default_availability_zone=internal and documenting this means they cannot be used for workloads14:28
sean-k-mooneyi.e. make it a parkign ground for newly added compute serivce that the operator is yet to assign to a real az14:28
sean-k-mooneyi dont think that woudl break any exising usecases14:28
bauzasthis could break the filter 14:29
bauzashttps://github.com/openstack/nova/blob/033af941792a9ae510c8d6b2cc318f062f0e1c66/nova/scheduler/filters/availability_zone_filter.py#L6814:29
sahidbauzas: if user can see that, they can try to deploy instance with this az which is not what we want as-well14:29
sean-k-mooneynot if we reject all api request with az=Config.internal_service_availability_zone14:30
sean-k-mooneysahid: i belive they cannot see it unless they are an admin14:30
sahid(i mean in an ideal case)14:30
bauzassahid: again, speaking of examples14:31
bauzassahid: if you already have AZ1, AZ2 and AZ2 that are meaningful14:31
bauzasall those three are shown by nova az-list14:31
bauzasand now you do care of not showing 'nova' in that list14:31
bauzaswhat I'm saying is that then change 'default_az' to any of the three, and job is done14:32
sahidok, let's use this way14:35
opendevreviewSylvain Bauza proposed openstack/nova master: Fix get_segments_id with subnets without segment_id  https://review.opendev.org/c/openstack/nova/+/88216015:06
dansmitheharney: if you're good with this, it could use a +W as everything it depends on is in the gate now: https://review.opendev.org/c/openstack/cinder-tempest-plugin/+/88176415:12
opendevreviewDan Smith proposed openstack/nova master: DNM: Test new ceph job configuration with nova  https://review.opendev.org/c/openstack/nova/+/88158515:18
dansmithman, the nova jobs sure have gotten big again15:21
dansmithwe're running a lot of jobs15:21
auniyaldansmith, yes its takes around 2 hour for a patch to get verified from tox15:28
auniyalzuul15:28
dansmithauniyal: that's generally a function of the slowest job, not the number of jobs15:32
dansmithgouthamr: the tempest stuff is all landed.. if you can +W the cinder-tempest-change now, it can go in and then the devstack plugin change is all that's left16:50
dansmitheharney: good catch I guess... I don't know if it's okay to just bump the requirement or not16:56
dansmithgmann: https://review.opendev.org/c/openstack/cinder-tempest-plugin/+/881764?tab=comments16:56
dansmith27 was wallaby and 30 was yoga16:57
dansmithso upstream I would think we're good bumping the version but I dunno if anything installs this from master in older versions...16:57
gmanndansmith: give me 2 min, will check17:25
dansmithgmann: ack thanks17:25
* gouthamr is catching up now.. 17:33
gouthamrstable jobs install "master" version of the plugin (and tempest) unless explicitly pinned17:34
dansmithgouthamr: but how far back?17:35
dansmithwe pin tempest at some point I think when the stable jobs get too old to support master tempest17:36
dansmithall the release jobs back to xena are passing17:37
gouthamrah good; the first pin i see is in stable/wallaby: https://github.com/openstack/cinder/blob/stable/wallaby/.zuul.yaml#L291-L29217:37
dansmithand I know they're using master tempest because they were honoring the depends-on when I was working on those patches (and breaking things)17:37
dansmithgouthamr: ack, and since the xena job on c-t-p works, we should be good then right?17:38
gouthamri think so dansmith; tosky and gmann would be experts in this area17:38
dansmithack, well, we'll see what gmann has to say17:38
gouthamrfor correctness, we need the latest version of tempest - and afawct, that's working as expected.. 17:39
dansmithyeah, so I'm not sure what the tempest requirement would be, if we're actually using master tempest17:40
gouthamrwe just hope no-one outside of our gates is pinning tempest for whatever reason but using a newer version of ctp.. (a weirdness i've seen albeit temporarily in some rdo jobs) 17:40
gmanndansmith: gouthamr tempest is pinned till wallaby as stable/xena still not in EM https://review.opendev.org/c/openstack/releases/+/88125417:40
dansmithgmann: right, but xena is good according to this: https://review.opendev.org/c/openstack/cinder-tempest-plugin/+/881764?tab=change-view-tab-header-zuul-results-summary17:41
gmanndansmith: ok, which one is failing17:42
dansmithgmann: nothing is failing17:42
dansmithgmann: eharney was concerned about the tempest version in ctp requirements17:42
gmanndansmith: ohk, checking that only. will reply17:43
dansmithbtw "nothing is failing" feels *amazing* to say :)17:43
gmann:)17:45
gouthamr++18:13
gmanndansmith: replied, agree with Eric to bump tempest version there https://review.opendev.org/c/openstack/cinder-tempest-plugin/+/881764/comments/0be5d072_64ceae3418:16
dansmithgmann: okay I'll update it, but I don't understand why it would time out18:17
dansmithoh is it because with ACTIVE we wait for the instance state to become the wait_until value?18:17
gmannyeah https://github.com/openstack/tempest/blob/27.0.0/tempest/common/compute.py#L23618:17
gmannit will pass SSHABLE as server status to wait for 18:18
dansmithack, okay I was focused on the SSHABLE special cases up earlier, I see18:18
dansmithjust updated, gouthamr ^18:18
gmannack18:20
opendevreviewMerged openstack/nova stable/yoga: Remove mentions of removed scheduler filters  https://review.opendev.org/c/openstack/nova/+/85802518:38
dansmithgouthamr: I think I better go ahead and do the cephadm->py3 job name change in the devstack plugin, otherwise we'll break nova as soon as that lands before we either rename it there or change nova (et al) to point to the new job18:41
gouthamrdansmith: +1; to clear this in my head and plan further cleanup - in your revert patch, you set cephadm options in "devstack-plugin-ceph-tempest-py3-base", correct?18:45
dansmithgouthamr: no I set the cephadm thing only in the cephadm job (which would be renamed to -py3).. probably no reason not to set it on the base one yet18:46
dansmithi figured I would rename the jobs and then you can remove the old one when you remove the non cephadm support from it18:46
dansmithjust leave the distro-based one named -ubuntu, nonvoting18:46
dansmithI can do the removal and refactoring there, I just don't want to change so much, especially in a patch that was supposed to be a revert :)18:47
dansmithI just want to get this landed so we can start getting more runs on the new stuff and not keep refactoring this, blocking that process18:48
dansmithand since I have it working in this form I just hesitate to move too much at once18:48
gouthamrack; this makes sense dansmith.. 18:56
dansmithgouthamr: okay I just pushed that up let me know if that looks okay18:57
* gouthamr looks18:57
opendevreviewDan Smith proposed openstack/nova master: DNM: Test new ceph job configuration with nova  https://review.opendev.org/c/openstack/nova/+/88158518:59
gouthamrdansmith: a couple of comments19:03
dansmithgouthamr: oh gate, right thanks :)19:04
dansmithgouthamr: "gate" has seemed so far off for a long time :D19:04
dansmithgouthamr: I don't know what is needed for rpm19:04
dansmithgouthamr: but since these jobs don't run on there, I think we can defer that19:05
gouthamrdansmith: "devstack-plugin-ceph-cephfs-nfs" is passing; although it doesn't do cephadm, it runs on centos-stream-9 19:06
dansmithright, and those packages are only needed for cephadm19:07
gouthamrdansmith: if we bump that to use cephadm soon, and we hit any issues, we'll get that covered :) 19:07
dansmithack, cool19:07
opendevreviewsean mooney proposed openstack/os-vif master: [WIP] set default qos policy  https://review.opendev.org/c/openstack/os-vif/+/88175120:46
opendevreviewDan Smith proposed openstack/nova master: Have host look for CPU controller of cgroupsv2 location.  https://review.opendev.org/c/openstack/nova/+/87312722:04
dansmithI have seen a lot of port binding related failures today, many of them on the grenade multinode job: https://c68ef44fab0ad498c042-f24a7834eba09db97966c05c7e428413.ssl.cf1.rackcdn.com/881585/8/check/nova-grenade-multinode/cbd5c81/testr_results.html22:37
melwittI saw one yesterday 23:25
dansmithmaybe sean-k-mooney could take a look in the morning23:58

Generated by irclog2html.py 2.17.3 by Marius Gedminas - find it at https://mg.pov.lt/irclog2html/!