Wednesday, 2021-10-13

*** pmannidi|AFK is now known as pmannidi00:24
stevebaker[m]TheJulia: it may be wrong, but I did this https://review.opendev.org/c/openstack/nova/+/81372902:06
TheJuliastevebaker[m]: I was thinking something similar to cut down on the generation loss that NobodyCam was talking about with the resource tracker03:14
TheJuliaI'm *hoping* I can spend some time heads down trying to wrap my head around the startup race condition that exists, and also the on-startup issues since they are basically bolted together03:14
stevebaker[m]TheJulia: you mean this startup race? https://bugzilla.redhat.com/show_bug.cgi?id=201222203:18
TheJuliastevebaker[m]: no, uhh.. the resource tracker looses the compute node one where they are trying to startup all of the compute services at once03:19
stevebaker[m]ah03:19
TheJuliaI've got logs on my work machine, I grok what is happening at a high level, just ahve to figure out a way to either make it more sane or retool things to not be as... built upon nova's preexisting assumptions of a compute service03:20
stevebaker[m]heh03:20
TheJuliastevebaker[m]: so. deleted verb to deleting state (which I don't think nova groks) to cleaning or error states. We likely need to consider deleting as well if nova doesn't explicitly know about it because it is visible for a fraction of a second03:25
TheJuliaunless there are problems which I suspect all sorts of other race conditions could hit on a super busy cluster03:26
TheJuliain theory self.cleanup_deploy won't complete until the node has exited a locked state03:28
TheJuliaI need to walk that code03:30
TheJulianot tonight03:30
* TheJulia goes and gets a shower and crashes03:30
stevebaker[m]I think a node DELETING state works by accident according to the current nova code, if it is short-lived https://opendev.org/openstack/nova/src/branch/master/nova/virt/ironic/driver.py#L1237-L125903:30
arne_wiebalckGood morning, Ironic!06:08
iurygregorygood morning arne_wiebalck and Ironic o/06:10
arne_wiebalckhey iurygregory o/06:10
rpittaugood morning ironic! o/07:11
iurygregorymorning rpittau o/07:17
rpittauhey iurygregory :)07:18
timeunnn07:36
dtantsurgood morning08:59
rpittauhey dtantsur :)08:59
iurygregorymorning dtantsur =)09:00
iurygregorydtantsur, rpittau, arne_wiebalck, TheJulia and anyone else interested in the PTG, if you have some time today to check if the new schedule is good https://paste.opendev.org/show/809919/ 09:20
dtantsurlgtm09:21
opendevreviewDmitry Tantsur proposed openstack/ironic-python-agent bugfix/8.1: Software RAID: Call udev_settle before creation  https://review.opendev.org/c/openstack/ironic-python-agent/+/81375309:26
arne_wiebalckiurygregory: one thing I was wondering if we should have a slot on discussing what we do or how we follow up with the input from the 3 sessions with ops feedback (in addition to acknowledge the points raised :))09:30
iurygregoryarne_wiebalck, ++ that makes sense to me I think we can add to Thursday as first topic09:31
jssfrhm, weher can I subscribe for information about the baremetal SIG operators sessions?09:31
jssfr*where09:31
iurygregoryif required we can also book a specific slot for this discussion09:32
arne_wiebalckjssfr: the sessions are announced on the openstack-discuss mailing list09:32
arne_wiebalckjssfr: https://etherpad.opendev.org/p/bare-metal-sig has the planning09:33
arne_wiebalckjssfr: and we usually tweet about these sessions as well in advance09:33
arne_wiebalckjssfr: if you have other suggestions where/how to announce them, let us know!09:33
jssfrhah, I am even subcsribed to openstack-discuss.... the signal/noise of that one isn't ideal though :/09:34
arne_wiebalckiurygregory: we probably need some triaging on which of the points raised can (realistically) turned into dev objectives09:34
jssfrthanks for the info, I might follow the pad09:34
arne_wiebalckjssfr: I agree, filter for [ironic] in the subject :)09:34
arne_wiebalckjssfr: and/or [baremetal-sig]09:35
arne_wiebalckjssfr: the SIG meetings are also announced/discussed here during the weekly meetings09:35
arne_wiebalckhere == here on IRC09:35
jssfrso I was kind of hoping for a low-noise announce mailing list or maybe even a caldav thing09:36
iurygregoryarne_wiebalck, yup =) I agree, your idea would be to triage before the PTG and we would only discuss the dev objectives? 09:37
jssfrthe weeklies are also outside of my working hours, so I'm generally ignoring IRC durign that time... lots of space to improve my processes here :)09:38
jssfrmight set up a highlight on #endmeeting or somescuh09:39
arne_wiebalckiurygregory: I think so, yes: there was a lot of discussion during the 3 hours, and we do not want all to read the notes during the ptg ... so we may want to distill and bring only a handful to the PTG, like the "main operator areas of interest for improvement"09:40
arne_wiebalckjssfr: IIRC, we tried to have split weekly meetings to include more timezones, but this never took off due to lack of interest/commitment in the end09:40
jssfrI imagine.09:41
iurygregoryarne_wiebalck, yeah I totally agree, since the PTG is next week I'm wondering if we have time to triage the topics before hand09:41
jssfrfor someone like me who'd just listen in, it's also rather pointless to have a second meeting.09:41
iurygregoryideas on how we can do this? =)09:41
arne_wiebalckiurygregory: well, the idea was to gather input from ops for the ptg; if would be a pity if, now that we have it, we are not able to include it in the discussion for the next cycle09:43
iurygregoryarne_wiebalck, agree =)09:43
arne_wiebalckiurygregory: many topics are not new and they have patches/ideas/code already, e.g. console support09:43
arne_wiebalckiurygregory: some are not really in our hands, like integration with nova09:44
iurygregoryyup09:44
arne_wiebalckiurygregory: there were some new requests (like power control fallback to the PDUs), but apart from these it is more to make the dev community aware of what operators struggle with daily09:45
arne_wiebalckiurygregory: and then have this given some priority, ofc :-D09:45
rpittauiurygregory: schedule looks good to me, unfortunately I'll have to skip almost all Friday 22 :/09:45
dtantsurarne_wiebalck, jssfr, maybe we need an RSS field (on ironicbaremetal.org?) with SIG news?09:53
arne_wiebalckdtantsur: yes, that sounds like a good idea to me!09:53
* arne_wiebalck has no idea how to do this ;_D09:54
jssfrI never warmed up with RSS and the likes, but this would be a great opportunity to try it again :)09:54
dtantsur(s/field/feed/, but you got the idea)09:55
dtantsurusually this stuff is generated from some kind of blog or even tags in a blog09:56
dtantsurmaybe we should just post announced on ironicbaremetal.org/blog?09:56
cumiHi, everyone. My name is Cumi. I am an Outreachy applicant for the December cohort. Excited to be contributing to the Replace pecan with flask in OpenStack Ironic project. Looking forward to the experience!09:57
dtantsurHi cumi, welcome! TheJulia will wake up a bit later today (she's in the USA), so if you cannot wait, please reach out by email.09:58
rpittaudtantsur: probably posting the announcements directly is easier to maintain09:58
rpittauwecleom cumi :)09:58
rpittauor welcome even09:59
dtantsurwas about to ask if it's french for welcome :D09:59
cumiAlright. Thank you @dantsur and rpittau09:59
cumi* dtantsur09:59
opendevreviewVerification of a change to openstack/sushy master failed: Add revision for drive resource  https://review.opendev.org/c/openstack/sushy/+/81267610:02
*** eandersson8 is now known as eandersson10:50
opendevreviewVerification of a change to openstack/ironic-lib stable/wallaby failed: json_rpc: list_opts() should return tuple  https://review.opendev.org/c/openstack/ironic-lib/+/81357010:51
opendevreviewVerification of a change to openstack/ironic-lib stable/wallaby failed: json_rpc: setup.cfg is missing entrypoint for json_rpc's list_opts  https://review.opendev.org/c/openstack/ironic-lib/+/81357110:51
*** janders_ is now known as janders11:05
jeevastill struggling with the networking component of ironic11:50
jeevaso i must create the cleaning network in openstack as a network11:50
janderssee you tomorrow Ironic o/12:12
TheJulia good morning13:07
dtantsurmorning TheJulia 13:11
TheJuliaso very very tired13:13
dtantsur:(13:13
TheJuliaOkay, cumi replied to13:33
TheJuliaJust... wow.13:33
iurygregorygood morning TheJulia 13:44
*** pmannidi is now known as pmannidi|AFK14:01
iurygregoryarne_wiebalck, TheJulia the new proposed schedule for the PTG https://paste.opendev.org/show/809976/ with the operators sessions 14:22
iurygregoryif we are ok I will update the etherpad to reflect and send an email to the openstack-discuss14:22
arne_wiebalckiurygregory: thanks14:26
arne_wiebalckiurygregory: I guess "bulk operations" is also ops inspired?14:26
arne_wiebalckiurygregory: if so, we could move this up (or ops feedback down)14:27
iurygregoryarne_wiebalck, yeah that makes sense =)14:28
opendevreviewManuel Schönlaub proposed openstack/sushy master: Add support for NetworkAdapter resources  https://review.opendev.org/c/openstack/sushy/+/81383914:29
dtantsurTheJulia: just created https://storyboard.openstack.org/#!/story/2009294, it's a part of the simplified start-up effort, and on the other hand - an early stepping stone to merging inspector14:53
opendevreviewManuel Schönlaub proposed openstack/sushy master: Add support for NetworkDeviceFunction and NetworkPort resources.  https://review.opendev.org/c/openstack/sushy/+/81385014:53
opendevreviewManuel Schönlaub proposed openstack/sushy master: Add support for additional network resources.  https://review.opendev.org/c/openstack/sushy/+/81385014:53
dtantsursee you tomorrow folks15:00
iurygregorybye dtantsur 15:14
rpittaugood night! o/15:48
iurygregorygn rpittau o/15:50
opendevreviewManuel Schönlaub proposed openstack/sushy master: Add support for NetworkAdapter resources  https://review.opendev.org/c/openstack/sushy/+/81383916:38
opendevreviewManuel Schönlaub proposed openstack/sushy master: Add support for additional network resources.  https://review.opendev.org/c/openstack/sushy/+/81385016:59
TheJuliazer0c00l: you around?19:47
TheJuliaI has a question w/r/t hash ring in nova compute19:47
opendevreviewMerged openstack/sushy master: Add revision for drive resource  https://review.opendev.org/c/openstack/sushy/+/81267621:27
opendevreviewJulia Kreger proposed openstack/ironic master: Test nova-compute fix  https://review.opendev.org/c/openstack/ironic/+/81326421:39
TheJuliaNobodyCam:  https://review.opendev.org/c/openstack/nova/+/813897 could use eyes21:46
Inwerejosic1Hi TheJulia, please I have a question regarding the gerrit account ssh connection. Is it my github ssh that I need to add?21:52
tonybI'm prewtty sure the answer is "no", but is there anyway to add a specific DHCP option per node?22:04
melwittTheJulia: I noticed the WIP change you uploaded to nova and wanted to ask, is there any chance this is the race you are looking to fix? https://bugs.launchpad.net/nova/+bug/185300922:31
TheJuliamelwitt: that is amongst the races which should be addressed by it, I *think*22:36
TheJuliamelwitt: I was going to try and think through that one tomorrow22:37
TheJuliatonyb: it is port binding info so if your a cloud admin yes you can rebind the port, with additional options, but as a user afaik no22:37
TheJuliaInwerejosic1: github ssh ? your ssh public key to gerrit?22:37
melwittTheJulia: ok, the fix for that merged recently and I'm backporting it to train, just fyi if it affects what you're trying to do22:37
TheJuliait shouldn't *but* it is like there are three distinct cases that "bad things" can happen in22:39
TheJuliashouldn't impact it, that is22:39
melwittok22:40
JayFIf you all figure out and solve all the races in nova-compute around the resource tracker, someone should send you dozens of cookies.22:41
TheJuliamelwitt: downstream there is another case artom is aware of which is one of the issues where someone went and upgraded their nova computes22:42
TheJuliamelwitt: their controllers, and to fix things they are having to do db surgery :(22:42
TheJulia#lastthinganyofuswant22:42
tonybTheJulia: Thanks.  I'll find another way22:44
melwitthm ok, I'll try to find which one that is22:44
melwittI hadn't heard about it yet22:44
TheJuliaarne_wiebalck: NobodyCam: the discussion over the last couple minutes with melwitt and I may be of interest22:44
TheJuliamelwitt: they aren't broken, just a visual inconvience I think, but also the instance.host getting tired forever seems "wrong" operationally to me22:44
TheJuliawhich is partly where my patch went to after chatting with artom22:45
Inwerejosic1TheJulia, I git that resolved, I ran <ssh-keygen -t rsa -C "email@somethingmail.com", and it generated my ssh public key which I have used to complete that process22:45
TheJuliaInwerejosic1: intentional git pun?22:46
NobodyCam:)22:46
melwittTheJulia: sorry what is happening to the instance.host?22:47
TheJuliause of puns is always appreciated!22:47
TheJuliamelwitt: so, instance.host, aiui, is the compute node that the instance is located on, which governs the select for the population of node cache with the nova.virt.ironic driver. Operationally, it is just out of date, and I *think* the cache not properly representing reality is where the compute node stuffs starts goign sideways because the constraint is against the host on the string, btu there is a missing piece of that in which 22:49
TheJuliaI've not got wired in my head22:49
-opendevstatus- NOTICE: Both Gerrit and Zuul services are being restarted briefly for minor updates, and should return to service momentarily; all previously running builds will be reenqueued once Zuul is fully started again22:50
TheJuliaugh22:51
Inwerejosic1I meant got that resolve...22:52
melwittok so it's something like they upgrade compute nodes and things rebalance but the instance.host's don't get updated to the [compute service] owner of the instance(s) that changed ownership22:52
jandersgood morning Ironic o/23:00
TheJuliamelwitt: and also just general bouncing of the service creates cases, from what I percieve where instances will never map to compute nodes23:03
melwittnoted23:04
*** pmannidi|AFK is now known as pmannidi23:50

Generated by irclog2html.py 2.17.2 by Marius Gedminas - find it at https://mg.pov.lt/irclog2html/!