Thursday, 2014-04-17

*** zdiN0bot has quit IRC00:03
*** zdiN0bot has joined #openstack-ironic00:05
*** zdiN0bot has quit IRC00:22
*** ilives has joined #openstack-ironic00:22
*** zdiN0bot has joined #openstack-ironic00:26
*** ilives has quit IRC00:27
*** ilives has joined #openstack-ironic00:28
openstackgerritA change was merged to openstack/ironic: Add DiskPartitioner  https://review.openstack.org/8339600:30
*** matsuhashi has joined #openstack-ironic00:31
*** blamar has quit IRC00:34
*** zdiN0bot1 has joined #openstack-ironic00:52
*** zdiN0bot has quit IRC00:54
*** blamar has joined #openstack-ironic00:59
*** newell_ has quit IRC01:03
*** zdiN0bot1 has quit IRC01:06
*** eguz has quit IRC01:17
*** todd_dsm has joined #openstack-ironic01:32
*** epim has quit IRC01:32
*** nosnos has joined #openstack-ironic01:33
*** rloo has quit IRC01:41
*** jgrimm has joined #openstack-ironic01:52
*** ilives has quit IRC02:35
*** ilives has joined #openstack-ironic02:36
*** lnxnut_ has joined #openstack-ironic02:51
*** lnxnut has quit IRC02:53
*** matsuhashi has quit IRC03:14
*** nosnos has quit IRC03:27
*** lnxnut_ has quit IRC03:33
*** todd_dsm has quit IRC03:50
*** todd_dsm has joined #openstack-ironic03:51
*** matsuhashi has joined #openstack-ironic04:20
*** nosnos has joined #openstack-ironic04:23
*** vkozhukalov has joined #openstack-ironic04:32
*** eghobo has joined #openstack-ironic04:32
*** todd_dsm has quit IRC04:34
*** harlowja is now known as harlowja_away04:47
*** ilives has quit IRC04:49
*** todd_dsm has joined #openstack-ironic04:49
*** ilives has joined #openstack-ironic04:50
*** todd_dsm has quit IRC05:00
*** eghobo has quit IRC05:00
*** eghobo has joined #openstack-ironic05:00
*** todd_dsm has joined #openstack-ironic05:00
*** boris-42 has quit IRC05:24
*** boris-42 has joined #openstack-ironic05:43
*** vkozhukalov has quit IRC06:03
*** eghobo has quit IRC06:11
*** pradipta_away is now known as pradipta06:14
*** ifarkas has joined #openstack-ironic06:55
*** ekarlso has quit IRC07:12
*** datajerk has quit IRC07:17
*** ilives has quit IRC07:29
*** ilives has joined #openstack-ironic07:29
*** datajerk has joined #openstack-ironic07:31
*** ifarkas has quit IRC07:32
*** mrda is now known as mrda_away07:34
*** lsmola has quit IRC07:34
*** athomas has joined #openstack-ironic07:40
*** ndipanov has quit IRC07:43
*** ifarkas has joined #openstack-ironic07:44
*** max_lobur has joined #openstack-ironic07:54
*** martyntaylor has quit IRC07:57
*** ndipanov has joined #openstack-ironic07:58
*** vkozhukalov has joined #openstack-ironic07:58
*** lsmola has joined #openstack-ironic07:58
*** martyntaylor has joined #openstack-ironic08:09
*** lucasagomes has joined #openstack-ironic08:11
*** jistr has joined #openstack-ironic08:13
*** martyntaylor has quit IRC08:16
dtantsurMorning Ironic08:21
*** foexle has joined #openstack-ironic08:24
*** dshulyak has quit IRC08:26
*** martyntaylor has joined #openstack-ironic08:31
*** Mikhail_D_ltp has joined #openstack-ironic08:34
*** pradipta is now known as pradipta_away08:43
*** derekh has joined #openstack-ironic08:52
yuriyzmorning dtantsur09:02
Mikhail_D_ltpMorning all! :)09:02
*** romcheg has quit IRC09:03
*** romcheg has joined #openstack-ironic09:03
openstackgerritVladimir Kozhukalov proposed a change to openstack/ironic-python-agent: Added disk utils  https://review.openstack.org/8616309:13
Haomengmorning all:)09:17
*** Mikhail_D_ltp has left #openstack-ironic09:18
dtantsurmorning Haomeng, yuriyz, Mikhail_D_wk :)09:29
lucasagomesmorning dtantsur Haomeng yuriyz Mikhail_D_wk09:33
dtantsurlucasagomes, morning09:33
romchegMorning dtantsur lucasagomes and everyone else09:44
dtantsurromcheg, morning09:45
lucasagomes:) romcheg morning09:45
* dtantsur 's job today is too greet everyone :)09:46
romchegdtantsur, :-P10:00
*** dshulyak has joined #openstack-ironic10:01
*** Mikhail_D_ltp has joined #openstack-ironic10:27
*** lnxnut has joined #openstack-ironic10:44
*** Mikhail_D_ltp has quit IRC10:54
Haomengdtantsur: morning:)11:03
*** dtantsur is now known as dtantsur|launch11:08
*** killer_prince has quit IRC11:13
*** lazy_prince has joined #openstack-ironic11:32
*** lazy_prince is now known as killer_prince11:32
*** lucasagomes is now known as lucas-hungry11:53
openstackgerritVladimir Kozhukalov proposed a change to openstack/ironic-python-agent: Added disk utils  https://review.openstack.org/8616312:10
*** jdob has joined #openstack-ironic12:11
*** lucas-hungry is now known as lucasagomes12:29
*** mgagne has quit IRC12:31
*** mgagne has joined #openstack-ironic12:33
*** mgagne is now known as Guest7821712:33
romcheglucasagomes: I'm looking at https://review.openstack.org/#/c/83572/12:39
lucasagomesromcheg, right12:39
lucasagomesromcheg, that superceed the whitelist one12:39
romchegright12:40
lucasagomesromcheg, it was created before the ff... I'll review it too12:40
romcheglucasagomes: test_ports.py contains a lot of changes not related to the patch12:40
lucasagomesew12:40
romchegshould we let it go or ask to separate them12:40
lucasagomeshmmm, the way I usually think about it is... if ur touching a specific method and u want to do a small cleanup/change that is not really related to the patch fine12:41
*** rloo has joined #openstack-ironic12:42
lucasagomesbut changing totally unrelated stuff in the same patch I don't usually like12:42
romchegmaybe it was commited by mistake :D12:42
lucasagomeshaven't reviewed so idk if it's related or not, if not there's no harm in -1 and asking to split in a new patch12:42
lucasagomesromcheg, possible12:42
*** yuriyz has quit IRC12:44
*** yuriyz has joined #openstack-ironic12:46
rlooromcheg: wrt 83572, did you read in the commit, why they changed test_ports?12:46
*** linggao has joined #openstack-ironic12:46
romchegrloo: it's possible to move that into a separate patch anyway. It will just be a dependency for this one12:48
rlooromcheg: yes, that's probably true. Sorry, it did occur to me but I'm not such a stickler for separate patches so I let it go ;)12:51
romchegrloo: That's just my opinion. If the rest of the team disagrees with me, I don't mind letting it go. That's how discussions work :)12:52
rlooromcheg: ha ha. So, to be fair about it (based on previous reviews etc) you are right. It should probably be put in a separate patch ;)12:52
*** nosnos has quit IRC12:56
rlooromcheg, yuriy, lucasagomes: https://review.openstack.org/#/c/83399/. If two cores +2, but don't approve, does a 3rd core need to review and approve or just approve w/o looking?13:01
*** yuriyz has quit IRC13:01
*** yuriyz|2 has joined #openstack-ironic13:01
rloooops, yuriyz|2 ^^13:02
lucasagomesrloo, I usually review and then approve13:02
lucasagomesI don't approve things I haven't reviewed13:02
lucasagomesalso worth take a look at the comments13:02
lucasagomessometimes people +2 but specifically say that he want to give others the oportunity to review as well13:02
lucasagomeshe/she13:03
rlooi don't see that in this review. just a comment from roman. Just wondering if diff cores have diff ways of working?13:03
rloook, i'll quickly review it since I had looked at it before.13:04
lucasagomesrloo, I c, yeah we probably do have diff ways, but that's good as well13:04
lucasagomesrloo, ack thanks :D13:04
*** overlayer has joined #openstack-ironic13:05
*** matsuhashi has quit IRC13:06
*** ilives has quit IRC13:07
*** ilives has joined #openstack-ironic13:08
NobodyCamgood morning Ironic says the man making coffee :)13:09
*** martyntaylor has left #openstack-ironic13:09
rloog'morning NobodyCam13:10
Mikhail_D_wkNobodyCam: g'morning :)13:10
romchegMorning NobodyCam13:11
lucasagomesmorning NobodyCam13:14
*** matty_dubs|gone is now known as matty_dubs13:14
lucasagomesNobodyCam, now I read a bit more about the blueprint stuff that nova is doing :)13:14
lucasagomesNobodyCam, I liked the idea, but agree with JayF that we might not know _everything_ before actually start coding13:14
lucasagomesrloo, yay thanksssss :D13:15
rloolucasagomes: ;)13:15
*** Madasi has quit IRC13:16
*** jbjohnso has joined #openstack-ironic13:16
NobodyCammorning rloo Mikhail_D_wk lucasagomes :)13:17
*** matsuhashi has joined #openstack-ironic13:17
*** Madasi has joined #openstack-ironic13:18
NobodyCamlucasagomes: yea, I really like the gerrit idea for BP's it was just a couple of the steps in the nova one that seemed odd to me13:19
NobodyCamwhy crate a "bad BP"13:19
lucasagomesNobodyCam, I think by "bad BP" they mean... create a tiny bp like 1 line comment13:20
lucasagomesin launchpad13:20
lucasagomesonce the bp is approved in the specs repo, they will copy the text from specs to the bp in launchpad13:20
lucasagomesNobodyCam, I think it's done because13:21
lucasagomestheorically the bp doesn't need to be approved to u to start working on it13:21
lucasagomes(idk if it should be enforced tho)13:21
lucasagomesNobodyCam, so to keep track of the code while it's being approved, we need an entry in launchpad13:21
lucasagomesthat's my theory for that rule at least13:21
NobodyCamI would rather start in the repo, once it been blessed then one of use (core) members who can actually accept the bp. cuts and pasted it to a actual BP and assigns it a priority13:22
*** jistr has quit IRC13:22
lucasagomesNobodyCam, yeah, but if someone starts working before it's approval13:22
lucasagomesthere's no way to tag that bp13:22
lucasagomesif it's not in launchpad13:22
NobodyCamahh13:23
NobodyCamI can see that13:23
lucasagomesyeah... it's quite sad that we are so integrated with launchpad tho13:24
lucasagomesthat's a horrible system13:24
NobodyCam:)13:26
rloolucasagomes, NobodyCam: I only skimmed some of that. Maybe I'm an optimist, but I interpreted "bad BP" == initial version/submission of a BP. Not necessarily bad. OTOH maybe the author really meant that.13:26
*** jistr has joined #openstack-ironic13:26
lucasagomesrloo, yeah I think ur correct13:26
lucasagomesrloo, bad BP is a bad way to describe it13:26
NobodyCam:)13:28
NobodyCamI can see many BP's that just say "SEE REVIEW #123"13:29
lucasagomeslol13:30
*** newell has joined #openstack-ironic13:30
NobodyCamlucasagomes: the lol comes then there is another patch tha link that BP to itself.13:32
NobodyCams/then/when/13:32
lucasagomeshaha oh boy13:32
NobodyCam:)13:32
lucasagomesyeah it's kinda fucked up the way we track bps now13:32
lucasagomeswe def need to improve it13:32
NobodyCambut I really like the bp review that perfect13:33
NobodyCamthats*13:33
lucasagomes+1 yeah that's much more in control13:34
rloospeaking of blueprints... https://blueprints.launchpad.net/ironic/+spec/nodeless-vendor-passthru13:36
*** jgrimm has quit IRC13:37
rloolucasagomes, NobodyCam. I don't actually know what/if there is a process. Should I be reviewing code associated with a blueprint, if the blueprint doesn't have an approver or some indication that someone thinks it is OK?13:37
lucasagomesrloo, well there's no problem reviewing it I think13:38
rloolucasagomes. There IS a problem. My TIME.13:38
lucasagomesI wouldn't approve without talking to the PTL13:38
lucasagomesrloo, heh13:38
rlooI'm serious.13:38
lucasagomesrloo, you can skip if u want13:38
lucasagomesrloo, sure it's fine to skip those reviews... if it's not approved, it's grand to skip it13:39
lucasagomesI think some people even -1 that13:39
lucasagomesI think I saw a case where people -1 saying the bp wasn't approved yet so they didn't want to merge13:39
rloook, will bring it up with deva then. I think we need to have clear expectations so everyone understands.13:40
lucasagomesrloo, +113:40
lucasagomeswe need to start a wiki or something with those reviewers faq13:40
lucasagomesit's a lot of corner cases13:40
rloowell, this doesn't seem like a corner case. seems like a process or lack of process, problem?13:41
lucasagomesyeah indeed lack of process13:42
lucasagomes:(13:42
lucasagomesI'm about to submit a review for a blueprint that wasn't approved yet13:42
lucasagomesso would be good to bring it up to devananda today13:42
rloook, but you're at the end of the queue buddy :)13:43
lucasagomeslol13:43
lucasagomesyeah13:43
NobodyCamlucasagomes: rloo: if I see a review for a bp that is not approved I tend to want to only +/- 1 it notting in the review "would get +2 if BP was approved"13:47
NobodyCamthat way the dev gets some feedback13:48
rlooNobodyCam, lucasagomes: hmm. I'll keep that in mind, but honestly, I don't plan on reviewing anything if the bp isn't approved. There are lots of other reviews that I could be looking at.13:50
lucasagomesNobodyCam, I c, yeah thats a good idea to use only +1 stuff13:50
NobodyCam:)13:51
openstackgerritLucas Alvares Gomes proposed a change to openstack/ironic: Port to oslo.messaging  https://review.openstack.org/8830713:59
* lucasagomes done sigh13:59
romcheglucasagomes: wow14:02
romcheglucasagomes: How was that? :)14:02
lucasagomesromcheg, heh please help review/test14:02
lucasagomesromcheg, tiring heh14:02
lucasagomesyou can't change one piece and get the whole working14:02
lucasagomesso it's a lot of changes in diff places14:02
romcheglucasagomes: Yup, I saw the same patch in nova14:03
romchegIt was super huge14:03
lucasagomesromcheg, yeah, I hope the Ironic one will be easier to review14:03
lucasagomesromcheg, yeah14:03
lucasagomesromcheg, the nova one had 78 patch-sets haha14:03
lucasagomeshttps://review.openstack.org/#/c/39929/7814:03
romcheglucasagomes: Should we do better (more)? :)14:04
lucasagomesromcheg, but it's better to port now than later, otherwise we are going to get stuck in the rpc stuff of the incubator and will be way more painful to port later14:04
lucasagomesromcheg, lol please no :P14:04
* lucasagomes hopes for the better case, only 1 patch-set!!!! *dreams*14:05
lucasagomesromcheg, damn just saw a debug thing I left14:05
lucasagomesok 2 patch sets14:06
romcheglucasagomes: lol14:06
*** viktors has joined #openstack-ironic14:06
rloolucasagomes: if those config options are removed, how can ironic be backwards compatible?14:07
romcheg22390614:07
lucasagomesrloo, those options came from openstack/common/rpc/* stuff14:08
romchegWhoops!14:08
lucasagomesrloo, now they exist in oslo.messaging14:08
romchegYou didn't see that....14:08
lucasagomesbut the syntax is backward compatible, the rpc.py is a wrap on oslo.messaging14:08
lucasagomesso it needs to understand both syntax the new and old14:08
* romcheg entered a second factor code to the wrong place14:08
openstackgerritLucas Alvares Gomes proposed a change to openstack/ironic: Port to oslo.messaging  https://review.openstack.org/8830714:09
lucasagomesfeck sample is not updated >.<14:10
lucasagomespatchset #314:10
openstackgerritLucas Alvares Gomes proposed a change to openstack/ironic: Port to oslo.messaging  https://review.openstack.org/8830714:11
*** ifarkas_ has joined #openstack-ironic14:12
*** viktors has quit IRC14:13
*** ifarkas has quit IRC14:14
rloolucasagomes: thx. Just wondering if we need to document that ;)14:15
lucasagomesrloo, I c14:15
lucasagomesyeah a good bit of config options were removed now14:16
lucasagomes1 new added14:16
*** viktors has joined #openstack-ironic14:19
viktorslucasagomes: you probably missed to change tools/config/oslo.config.generator.rc file in https://review.openstack.org/8830714:23
viktors14:23
lucasagomesviktors, :(14:24
lucasagomesviktors, thanks I will update that :D14:24
NobodyCamand back.. morning romcheg :)14:24
*** ifarkas_ has quit IRC14:25
viktorslucasagomes: I just found, that you removed the large number conf options in config.sample14:25
lucasagomesviktors, yeah, adding it back :D thank you for pointing it out14:26
*** ifarkas_ has joined #openstack-ironic14:26
viktorslucasagomes: no problem :)14:26
*** matsuhashi has quit IRC14:30
openstackgerritLucas Alvares Gomes proposed a change to openstack/ironic: Port to oslo.messaging  https://review.openstack.org/8830714:30
*** jgrimm has joined #openstack-ironic14:33
NobodyCamhumm.. [-] Error during ClusteredComputeManager.update_available_resource: f34540e9-5cc4-420c-95f2-378fd71c7485 is not a valid node managed by this compute host.14:37
NobodyCamthats a strange error14:37
agordeevgood morning/evening Ironic14:37
NobodyCammorning agordeev14:37
NobodyCambbt..brb14:38
lucasagomesNobodyCam, I'm looking at it now14:39
lucasagomeslemme open a bug about it, I found the cause14:39
lucasagomeslooking at how to solve14:39
lucasagomesNobodyCam, the cause is that u enrolled a powered on machine14:39
ShrewsNobodyCam: quit breaking things14:40
Shrews:)14:40
lucasagomesyeah I will put a fix up soon14:42
agordeevjroll: JayF are you arouund? could you help me? i'm trying to test how IPA works. For now dracut has failed to boot and dropped to emergency shell. Idk how to troubleshoot it (also ironic-conductor logs are pretty silent)14:42
lucasagomeshttps://bugs.launchpad.net/ironic/+bug/130904814:42
agordeevmorning NobodyCam Shrews and everybody else!14:42
NobodyCamlucasagomes: https://review.openstack.org/#/c/85529/15/ironic/conductor/manager.py14:42
NobodyCammornign Shrews14:42
NobodyCam:-p14:42
Shrewsagordeev, NobodyCam, et.al: morning14:42
lucasagomesNobodyCam, yeah... but I think the bug is a bit diff14:43
NobodyCamok..14:43
lucasagomesNobodyCam, but that would actually work around the problem14:43
NobodyCamthat was from the last ci run of that patch14:44
lucasagomesNobodyCam, hmm lemme investigate a bit more, I not complete sure about what node_is_available should do14:44
NobodyCamlucasagomes: sure14:45
NobodyCam:)14:45
agordeevjroll: JayF also it took too long to init http://picpaste.com/pics/dracut_fail-vUpXXQTh.1397745849.png slightly less than 10 minutes. How did you test it?14:45
*** ilives has quit IRC14:55
*** ilives has joined #openstack-ironic14:56
*** blamar has quit IRC14:56
*** foexle has quit IRC14:59
*** hemna_ has joined #openstack-ironic15:00
openstackgerritLucas Alvares Gomes proposed a change to openstack/ironic: Better check for available nodes  https://review.openstack.org/8468715:02
*** blamar has joined #openstack-ironic15:04
devanandamorning, all15:08
NobodyCamgood morning devananda :)15:08
*** blamar has quit IRC15:08
NobodyCamlucasagomes: is ^^ that releated to that error?15:08
lucasagomesNobodyCam, yes, that actually fixes it15:08
lucasagomesNobodyCam, I just dunno why we need node_is_available15:09
lucasagomesdevananda, morning15:09
agordeevdevananda: morning15:09
lucasagomesdevananda, deva about bps, what you feel about? https://blueprints.launchpad.net/ironic/+spec/oslo-messaging15:09
agordeevlucasagomes: morning15:09
lucasagomesdevananda, I just put a patch up doing that ^, can you approve the bp for me please?15:10
openstackgerritLucas Alvares Gomes proposed a change to openstack/ironic: Better check for available nodes  https://review.openstack.org/8468715:10
NobodyCamlol ... lucasagomes: stop that...hehehe15:11
lucasagomesNobodyCam, heh sorry, just tagged the bug in the commit message15:11
NobodyCamhehehe its all good.. just giving you a hard time15:12
NobodyCam:-p15:12
lucasagomesheh15:13
lucasagomesbut I'm wondering whether we should overwirte the node_is_available it's just duplicating the tests15:13
lucasagomesthe default implementation of node_is_available calls get_available_nodes() and check if the node is there15:14
lucasagomesthat's grand, I think that's how it should be15:14
lucasagomeswe shouldn't duplicate the conditions there (although it's more efficient to get() a specific node)15:15
lucasagomesmaybe I can isolate the conditional for both methods15:16
lucasagomes1 sec... NobodyCam will mark that ^ as WIP15:16
devanandalucasagomes: done. i need to go through all the BPs and look at the ones we should approve pre-summit15:17
lucasagomesdevananda, ta much!15:17
lucasagomesdevananda, I did everything in one commit, because I really dunno hw to split that in multiple ones15:17
devanandaheh15:18
NobodyCamya they both are check the exact same things15:18
NobodyCam*checking15:18
lucasagomesNobodyCam, yeah it does15:19
devanandalucasagomes: i haven' tlooked yet -- will that supercede my update https://review.openstack.org/#/c/87328/ ?15:19
JayFagordeev: I don't understand what you're saying or what that screenshot is showing me. More information?15:20
lucasagomesdevananda, ops, it will15:21
lucasagomesdevananda, actually I removed the whole openstack/common/rpc/ module15:21
JayFagordeev: as far as testing the agent, we have a couple of hardware labs we're working it, within the next 2 weeks we're going to work on a way to better test the agent locally, but we're just trying super hard to get a working prototype by Atlanta.15:21
devanandalucasagomes: great15:21
NobodyCamJayF: I cant wait to see your demo..15:22
JayFI can't wait for us to have a demo15:22
JayFlol15:22
*** ifarkas_ has quit IRC15:23
NobodyCam:)15:23
lucasagomes- check-tripleo-ironic-seed-precise http://logs.openstack.org/07/88307/4/check-tripleo/check-tripleo-ironic-seed-precise/989d0b1 : SUCCESS in 36m 31s!!!!!15:23
lucasagomeson the olso messaging!15:23
devanandalucasagomes: \o/15:24
agordeevJayF: hello15:24
NobodyCamlucasagomes: w00t15:26
lucasagomes:D was waiting for that check ^^15:26
agordeevJayF: it looks like agent image is unable to boot properly. Then it just gives an emergency shell that's all.15:26
openstackgerritLucas Alvares Gomes proposed a change to openstack/ironic: Better check for available nodes  https://review.openstack.org/8468715:26
lucasagomesNobodyCam, done ^15:26
* NobodyCam refreshed his window15:26
Shrewsdevananda: possibly a dumb question, but is the intent that preserve_ephemeral will only be supported with pxe?15:27
agordeevJayF: does it require any additional configuration at ironic's side? Maybe i'm just missing something15:27
JayFagordeev: that's strange, we have the agent booting on several machines. Are you building from latest master? (imagebuild/coreos/coreos_oem_inject.py is using 197.0.0)15:27
devanandaShrews: nope. it should be supportable by other deploy drivers, too15:27
JayFagordeev: the agent should just boot to a prompt from the perspective of the terminal15:27
JayFagordeev: the terminal is mostly useless in coreos though, as no password is set on the core user15:27
lucasagomesNobodyCam, argh ops, spotted an error15:27
lucasagomes>.<15:27
agordeevJayF: yup, also i'd applied your latest patch which reduces image size15:27
devanandahey guys -- look :) https://launchpad.net/ironic/icehouse/2014.115:28
Shrewsdevananda: ok. saw that pxe was the only driver with a driver_info attribute for that, so it made me question it's use15:28
devanandaall the things we did in icehouse, in one page15:28
devanandaShrews: pxe is the only deply driver in trunk today, too :)15:28
NobodyCam\o/15:28
JoshNangwoo!15:28
JayFagordeev: if you can give me a full boot output when it fails (serial console is good for this), I'll gladly take a look, but without that information it's hard to troubleshoot15:28
agordeevJayF: nice, so it's impossible to figure out what happened15:28
JayFagordeev: no, not impossible. You can use the terminal from that rescue prompt to get info on why it wouldn't boot from dmesg and journalctl15:29
NobodyCamour first release... awesome work EVERYONE!!!!15:29
agordeevJayF: ok, i'll try to get serial console output15:29
lucasagomesNobodyCam, heh I forgot it was in the nova driver was running tests with testr instead of tox so I didn't see the error15:29
JayFagordeev: just saying if it boots to a prompt, you can't login, but that's a good thing15:29
JayFagordeev: aha, what's the kernel params you're passing?15:30
JayFI bet I just thought of what your problem is15:30
JayFtwo things: how much ram is on that box, and can you send me the ipxe.txt or pxe config file you're booting it with15:31
agordeevJayF: default params just what ironic has15:31
JayFI've never used the built in ironic dhcp/pxe stuff, we've been managing dhcp/pxe outside of ironic15:31
JayFI suspect it's two things:15:31
agordeevJayF: i'd tryed 3G and 4G RAM. It doesn't matter. What is the minimum?15:31
JayF1) You need to have the correct kernel params on the command line for it to boot (I'll gist what we're using in a sec)15:31
JayF2) The default run.sh for coreos is set to remount the tmpfs to be 20G instead of their default of 2G. I'd suggest you lower that number to 1/2 your ram (or comment out the line) in imagebuild/coreos/oem/run.sh15:32
agordeevJayF: http://paste.openstack.org/show/76140/ here is the pxe config from ironic.15:33
JayFhttps://gist.github.com/jayofdoom/c30538662b9661141f0b as I note in there, you probably don't need ipa-advertise-host among other things -- we have multiple interfaces that DHCP in our hardware, so we have to tell the agent which one to advertise on15:34
JayFagordeev: yeah, that append line will have to be different for the agent, and not just because we're not using DIB15:35
*** lynxman has quit IRC15:36
lucasagomesthierry is spamming my mail inbox ;(15:36
openstackgerritLucas Alvares Gomes proposed a change to openstack/ironic: Better check for available nodes  https://review.openstack.org/8468715:37
*** lynxman has joined #openstack-ironic15:37
agordeevJayF: also question. Is that information stated anywhere in docs? I think it's worth to be mentioned somewhere15:38
JayFagordeev: I agree. For now, I think a sample set of pxe configs in the imagebuild/coreos/ directory might be sufficient, with obviously something more significant in the future15:39
NobodyCamOH! --  POWER_MANAGER=nova.virt.baremetal.virtual_power_driver.VirtualPowerManager15:39
JayFright now you still have to apply about a half dozen patches or more to Ironic to even get the agent talking to Ironic at all, we'd like to get something working and merged before we write a ton of docs about how to use it :)15:40
JayFagordeev: https://github.com/openstack/ironic-python-agent/blob/master/imagebuild/coreos/oem/run.sh#L8 this is the place where you'll have to either comment out that line or change the ram amount in order to get it to boot on regular machines15:40
*** lsmola has quit IRC15:41
*** eghobo has joined #openstack-ironic15:41
JayFagordeev: http://coreos.com/docs/running-coreos/bare-metal/booting-with-pxe/ the section "Adding a Custom OEM" here describes what the imagebuilder is doing if you wanted more docs for now about how it works15:42
agordeevJayF: ironic's patches have been already applied :)15:42
agordeevJayF: great!15:42
lucasagomesShrews, all that pxe_root_gb, pxe_swap_gb, pxe_ephemeral_gb, pxe_preverse_ephemeral... will be moved from driver_info to instance_info and should be driver agnostic15:42
JayFagordeev: tl;dr of ^ is that the oem/ dir gets injected into the provided coreos pxe image, and coreos runs /usr/share/oem/run.sh on boot to configure things15:42
JayFagordeev: in our case, that's making the tmpfs larger (optional, and something that we should probably un-upstream), injecting SSH keys, then loading up the agent container and adding the unit file to systemd15:43
Shrewslucasagomes: is there a current review for that?15:43
lucasagomesShrews, not yet, I think one of the rackspace guys are moving that (I think there's a bug or bp about it up)15:43
JayFyeah, there's a BP up for it15:44
JayFand I think it's next up on our todo once some of the outstanding stuff gets reviewed or merged15:44
agordeevyuriyz|2: ^^ note about run.sh15:45
lucasagomesack15:45
lucasagomesJayF, thanks15:45
*** ndipanov has quit IRC15:49
*** dtantsur|launch is now known as dtantsur15:50
openstackgerritA change was merged to openstack/ironic-python-agent: Deprecated extension name attribute  https://review.openstack.org/8755715:50
*** ndipanov has joined #openstack-ironic15:50
Shrewsbiab. change of venue...15:51
*** ndipanov has quit IRC15:58
rloodevananda: when you have a moment, would you please take a look/approve this BP: https://blueprints.launchpad.net/ironic/+spec/nodeless-vendor-passthru15:59
lucasagomestools/config/generate_sample.sh is returning that it's not updated from master16:00
JayFrloo: code already exists for that too... here --> https://review.openstack.org/#/c/81919/16:00
lucasagomesit's causing pep8 gate checks to fail16:01
JayFrloo: that's one of the patches that's required for getting the agent driver merged, so I'm glad you're looking at the bp16:01
rlooJayF: yes, I know. I was about to review it, but the BP hasn't been approved.16:01
lucasagomesdevananda, ^ should we have a patch updating the config?!16:01
JayFrloo: :D Thanks16:01
devanandalucasagomes: hm?16:01
rlooJayF: don't thank too soon. I won't start before the BP is approved. Is that the first review to look at? I know you guys have several.16:01
lucasagomesdevananda, the tools/config/check_uptodate.sh is failing on master16:01
devanandalucasagomes: yikes. that can happen when oslo libs change16:02
lucasagomesdevananda, :( should we make a patch fixing that?16:02
lucasagomesdevananda, it's causing pep8 checks in gate to fail16:02
devanandayes16:03
devanandai'm testing now16:03
lucasagomesdevananda, ack, will prepare one after u confirm16:03
*** viktors has quit IRC16:03
jrollrloo: the first patch we need merged is https://review.openstack.org/#/c/81919/16:04
jrollrloo: then this is the big agent driver patch: https://review.openstack.org/#/c/84795/16:05
JayFDon't we need swift urls before the big agent driver patch?16:05
jrollrloo: after that there's some related things16:05
jrollJayF: I thought that landed?16:05
openstackgerritLucas Alvares Gomes proposed a change to openstack/ironic: Update ironic.conf.sample after oslo changes  https://review.openstack.org/8835216:05
NobodyCamrloo: didn't you have a issue with generate_sample.sh and the keystone conf settings?16:05
rloojroll: ok, so need the BP to be approved first.16:05
jrollJayF: oh, you're right16:05
jrollrloo: ah, yes16:06
jrollthanks16:06
openstackgerritDevananda van der Veen proposed a change to openstack/ironic: Update ironic.conf.sample  https://review.openstack.org/8835316:06
rlooNobodyCam: hmm, I don't remember. Except for the issue that generate_sample.sh doesn't generate a .sample that is quite right for some options.16:06
jrollrloo: devananda: this bp is related to all of the subsequent patches: https://blueprints.launchpad.net/ironic/+spec/agent-driver16:06
devanandaFWIW nova moved away from having a conf.s.ample in tree at all, because of this problem16:06
openstackgerritMikhail Durnosvistov proposed a change to openstack/ironic: Cleanup mock patch without `with` part 2  https://review.openstack.org/7325616:08
lucasagomesdevananda, hmm... I don't know if moving away, but maybe the check shouldn't be mandatory to pass16:08
openstackgerritMikhail Durnosvistov proposed a change to openstack/ironic: Cleanup mock patch without `with` part 3  https://review.openstack.org/8653616:08
openstackgerritMikhail Durnosvistov proposed a change to openstack/ironic: Cleanup mock patch without `with` part 1  https://review.openstack.org/7322316:08
openstackgerritMikhail Durnosvistov proposed a change to openstack/ironic: Get rid of the newline "\"  https://review.openstack.org/6679316:08
NobodyCamhummm.. I like the sameple file, but if its going to cause issues :(16:09
lucasagomesdevananda, then we could have a jenkins job that generate it for us16:09
lucasagomesfrom time to time, like once a week it generates a patch updating the sample file16:09
devanandai prefer the gate check16:09
devanandarequires patch authors & reviewers to know that they're changing the config16:10
devanandabut yea, it causes issues like this when upstream req's change, like oslo libs or keystoneclient16:10
lucasagomesright, then we have to deal with this out of sync problems16:10
lucasagomesyeah16:10
*** rloo has quit IRC16:11
*** romcheg1 has joined #openstack-ironic16:11
devanandarussell_h: so it's a bit backwards at this point, but could you update the nodeless-vendor-passthru BP with some implementation details?16:11
* NobodyCam git clones more coffee16:12
devanandarussell_h: like what the new API (would) look like, how the RPC routing (would be) is done, etc16:12
*** rloo has joined #openstack-ironic16:12
*** todd_dsm has quit IRC16:12
devanandarussell_h: this is a problem for most of our BPs, in my opinion - -they're too light on the proposed implementation details16:13
devanandai get that, often, one can't know exactly how the change will play out, but eg. "the API will randomly select a suitable conductor to send the message to"16:13
devanandavs "the API will braodcast the message to all conductors with that driver"16:13
*** foexle has joined #openstack-ironic16:17
lucasagomesromcheg, re 83572... although I think it can be splitted in another patch, I think the changes in the manager and driver factory stuff caused test_ports to need to be updated16:19
lucasagomesromcheg, so i think I'm ok with that patch16:19
romcheg1lucasagomes: ok, I'm going to change my vote then16:20
lucasagomesromcheg1, ack, but ur not wrong tho... I think yeah it could be in another patch16:20
rloolucasagomes: but to be devil's advocate, test_ports can be updated first, and then the changes in mgr/driver factory will work.16:20
lucasagomesrloo, yup, it could16:21
lucasagomesrloo, but it's hard to distinct what in that patch is a refactor or needed to make it work16:21
lucasagomesrloo, so, I think I'm fine with that change16:21
rlooI'm actually fine with it, BUT I've seen previous reviews where it was suggested that things be separated...16:21
romcheg1Since there are 2 +2s already I'm going to approve it16:22
rloothx romcheg1 ;)16:22
romcheg1my pleasure :)16:22
lucasagomesromcheg1, thanks!16:22
dtantsurlucasagomes, romcheg1, rloo: just to remind you guys: we already have 9 (!) refactoring patches around, and I've no idea how to merge them :)16:22
rloodtantsur: yes, I know :-(16:23
lucasagomesdtantsur, :(16:23
lucasagomesdevananda, romcheg NobodyCam rloo ... what about a review jam to get those merged ^16:23
lucasagomesfocus on that refactoring stuff and get what's possible in16:23
*** Guest78217 is now known as mgagne16:23
*** mgagne has joined #openstack-ironic16:24
rlooto be honest, I don't really think a review jam is necessary. I think if we do one at a time, and by 'do', need two cores to approve fairly quickly, it can be done.16:24
romcheg1Got to rush to my classes. lucasagomes, will give you some feedback on oslo.messaging after that16:24
dtantsurwhat I want to do actually is the following: 1. make sure https://review.openstack.org/#/c/73223/ is fixed (work with Mikhail_D_wk) 2. one morning get to guys with +2 power to quickly review and approve it 3. repeat with the next one16:24
lucasagomesromcheg1, right, ta much for review it!16:24
dtantsur* get 2 guys with +2 power even16:25
*** matty_dubs is now known as matty_dubs|lunch16:25
rloobut I don't know if any of the refactoring reviews are ready. Haven't looked recently but they seemed to need jenkins approval or they depended on some other review.16:25
devanandaI'd like to see comstud's refactoring land, as that's solving some things I feel are significant16:25
jrolldevananda: so maybe it's just an openstack process thing. but, I don't understand the point of updating blueprints with implementation details after patches are up. the implementation details are in the code - I don't understand why someone would go back and look at completed blueprints (and thus why do we need to update those?).16:25
devanandajroll: yea. folks do go back and look at BPs as artefacts of process by which they judge what has changed in the project at release time16:26
rloojroll: from my perspective, I don't want to find out what something is supposed to do when I'm reviewing it. I'd like to see some sort of design first. But I may be a minority in this.16:26
devanandajroll: taht said, to a degree, it's process overhead taht I don't like16:26
jrolldevananda: but do they look at implementation details to judge what has changed?16:26
devanandajroll: and i'd rather we got those details sorted up front (during BP review, before a single line of code is written16:26
devanandajroll: so i'd ilke us to start movnig towards the specs repo style of work that nova, neutron ,and tripleo are adopting16:27
devanandajroll: but not going to impose taht on existing work ...16:27
comstud+1 on specs repo16:27
devanandajroll: and yes, there are point-haired-folks who DO look at the BPs and DONT look at the code16:27
devanandajroll: and they make judgements based on that16:27
jrolldevananda: I know you're busy, but this blueprint has sat for two months with no questions or comments until now when people are ready to land it.16:27
jrolldevananda: sigh16:27
devanandajroll: i know... sorry :(16:28
jrollI understand16:28
jrollI'm just getting frustrated at constantly rebasing things while code is being changed out from under us, sorry if I'm coming off unpleasant :/16:28
devanandajroll: what i'm asking for now is that you add some description of the change to the BP, so that there is an artefact describing it16:29
jrollright16:29
devanandajroll: no worries. your frustration is justified, and fairly directed at me. I can take it :)16:29
jrollheh, thanks :)16:29
jroll(for putting up with me)16:29
devanandajroll: I'd say that the core team wasn't large enough 2 months ago to handle the influx of patches that you guys have added -- but I also believe I said that we propbably wouldn't get to looking at the agent code until Juno opened16:30
devanandajroll: and the core team just expanded and we're trying to catch up now :)16:31
jrolldevananda: I know, that's fair16:31
* jroll will poke russell_h to update the blueprint16:33
openstackgerritDmitry Tantsur proposed a change to openstack/ironic: Implement caching for master images  https://review.openstack.org/8538716:34
*** dshulyak has quit IRC16:35
comstudjroll: What's the conductor method being called in this thread starvation issue? change_node_power_state ?16:36
*** dshulyak has joined #openstack-ironic16:36
jrollcomstud: yes16:37
comstudty16:37
comstudwhat does task.driver.power.validate do? ipmi call?16:38
jrollyes, checks power status16:39
jrollI believe16:39
comstudty16:39
* jroll checks16:39
comstudi assume task.driver.power == ipmitool driver?16:39
comstudyeah, IPMIPower, I'm guessing16:40
jrollyeah, it does an IPMI call to make sure the bmc is there16:40
comstudok16:40
jrollyes, ipmitool16:40
lucasagomescomstud, check if the credentials you have is valid16:40
comstudthanks16:40
*** rloo has quit IRC16:47
JayFagordeev: https://bugs.launchpad.net/ironic/+bug/1309110 you /do/ have to set ipa-advertise-host. I filed a bug to make that default much better.16:48
*** rloo has joined #openstack-ironic16:48
JayFagordeev: would be a very good low hanging fruit for someone wanting to get a start in the agent.16:48
*** jistr has quit IRC16:49
*** zdin0bot has joined #openstack-ironic16:49
lucasagomesdone for today, came to the office so gotta take the train back home16:55
NobodyCamhave a good ride back home16:55
lucasagomeshave a g'night devananda NobodyCam rloo comstud JayF jroll , everyone...16:55
comstudnight!16:55
jrollnight lucas16:55
rloociao lucasagomes!16:55
dtantsurlucasagomes, g'night16:56
JayFnight16:56
comstudjroll: what state is passed to change_node_power_state?16:57
comstudIs this REBOOT, power on, off, or.. ?16:57
jrollI was mostly sending off16:57
comstudok, ty16:58
jrollbut have seen it with on16:58
*** lucasagomes has quit IRC17:00
*** eghobo has quit IRC17:02
*** derekh has quit IRC17:04
*** max_lobur has quit IRC17:04
*** eghobo has joined #openstack-ironic17:07
openstackgerritJay Faulkner proposed a change to openstack/ironic-python-agent: Better documentation and defaults for coreos image  https://review.openstack.org/8836417:07
openstackgerritJay Faulkner proposed a change to openstack/ironic-python-agent: Use docker import/export to make image smaller  https://review.openstack.org/8781917:07
jrollagordeev: I know it's kinda of late for this, but how would y'all feel about writing up a blueprint for partioning in the agent? I'd kind of like to see the end goal - what is planned to be supported, not supported, etc.17:07
*** dwalleck has joined #openstack-ironic17:07
*** Mikhail_D_ltp has joined #openstack-ironic17:07
vkozhukalovGuys, what are you thinking about assigning more people to be cores in IPA in European timezone? I mean some core reviewers romcheg  and yzveryanskyy who are in european tz. I believe it can help to speed up merging patches.17:08
vkozhukalovafaik, lucasgomes is also in Europe17:09
jrollvkozhukalov: my first question is, do you see problems with the speed of reviews?17:09
jrollor do you just mean that a patchset cycle takes a full day because we don't review until you're done for the day?17:10
vkozhukalovactually, I see the problem with merging particular request https://review.openstack.org/#/c/86163/17:10
jrollI'm open to adding some ironic cores to IPA core team if needed, we even talked about merging the two teams at one point17:11
jrollvkozhukalov: right, so it sucks that it takes a full day to do a patchset / review cycle due to timezones17:11
vkozhukalovjroll: exactly17:12
jrollbut we have reviewed each patchset on that review quickly17:12
jrollok17:12
jrollJayF / JoshNang / russell_h wdyt?17:12
JayFI think some of what will help with this is the design summit at Atlanta and better up front understanding about what's being done17:12
*** harlowja_away is now known as harlowja17:12
JayFon that merge request in particular, I think it's awkward to review because it's a middle step without a clear idea of how it's going to integrate in the end17:13
vkozhukalovjroll: yes, i don't think that review is slow, you guys are really rock, I just think that we can improve the situation a bit17:13
JayFi.e. all the existing stuff in the agent assumes that an image with partitions is being given in, how does the agent determine what disks to partition, etc17:13
JayFso for me personally, even though I've already +2'd it, it probably wasn't a good idea to in retrospect because I don't know what the next steps are17:14
JayFI'm hoping that once we have something working, and the base layer merged, we'll all be able to talk about where we're going, and things like that17:15
JayFthe main piece that I think hasn't been well communicated is the scope of the agent, and there might be some conflict about what things belong in the agent vs not (like the LVM/MD conversation we had with devananda the other day)17:15
JayFand until the scope is well understood by all, I'd be nervous about expanding the core group of reviewers17:16
vkozhukalovthere is bp https://blueprints.launchpad.net/ironic/+spec/ironic-python-agent-partition and everyone is welcome to criticize it17:16
jrollheh, I had no idea this existed17:16
vkozhukalovnext step is making paritiotion extension17:17
jrollvkozhukalov: I'd like to see more info on that - what is supported, what is not supported, what the goal is, etc17:17
vkozhukalovjroll: JayF: it is my fault, that you did not know about it17:17
jrollthat's ok17:17
vkozhukalovhttps://etherpad.openstack.org/p/ironic-disk-partitioning some additional info17:18
comstuddevananda, jroll: Ok. So, I think I know what is happening with regards to the thread starvation17:18
*** rloo has quit IRC17:19
JayFvkozhukalov: can you reflect the contents of that etherpad in the blueprint? or at a minimum link it in?17:19
vkozhukalovit is clear for now about lvm, we are not going to support it in agent, but md is still in the scope17:19
*** rloo has joined #openstack-ironic17:19
JayFAFIACT, it's not in scope for the same reason LVM isn't.17:19
comstuddevananda, jroll: What is happening is that we get into a state where all RPC pool threads are blocked waiting on the Conductor worker threads. And in this state, the scheduling list is such that none of the RPC pool threads run until *all* of the Conductor workers complete.17:19
vkozhukalovthere is a link in bp to this etherpad17:19
comstuddevananda, jroll: So nothing is sending RPC responses until all of the worker threads are done... and at a default of 64, this can be a long time.17:21
devanandacomstud: hmm... taht doesn't seem logical17:21
vkozhukalovJayF: I don't remember anyone was against md.17:21
comstuddevananda, jroll: It's how eventlet does scheduling.. which is arguably bad.17:21
jrollside note, all involved with the agent, I've updated the main agent blueprint (and would like to see blueprints for other agent features in general): https://blueprints.launchpad.net/ironic/+spec/utility-ramdisk17:21
devanandacomstud: ah. lovely17:21
devanandacomstud: is taht only a problem if all worker threads get consumed?17:21
JayFHmm I thought the argument was that if you couldn't use it with a Windows image, you shouldn't support it17:21
devanandafwiw guys, i just emailed the list re: an ironic specs repo17:21
comstuddevananda, jroll: Say your worker thread pool is full.  So you have RPC threads blocking on them..17:21
JayFbecause of the strangeness of, should we support /windows/ software raid as well?17:22
devanandaJayF: fwiw, I do not beleive i said that17:22
comstuddevananda, jroll: When a worker finishes... it doesn't immediately schedule the RPC thread17:22
JayFThere's just a whole path of nastiness that supporting linux software raid takes us down17:22
comstuddevananda, jroll: It just adds to end of the list17:22
JayFdevananda: my misunderstanding then17:22
comstuddevananda, jroll: So that means all of the workers have to be scheduled and run first before you get back to the RPC pool thread17:22
*** matty_dubs|lunch is now known as matty_dubs17:22
jrollcomstud: sure... the thread I see getting starved is a heartbeat looping call thing. where does that fit in?17:22
comstudah right, I did't factor that in here17:22
jroll:)17:22
*** rloo has quit IRC17:22
comstudI was focusing on API locking up17:22
vkozhukalovwe agreed that we can store some attributes in image metadata about supporting or not such things like md17:23
jrollcomstud: I mean, that's a problem too :P17:23
comstudof which there's a clear explanation17:23
*** rloo has joined #openstack-ironic17:23
comstudbut17:23
JayFvkozhukalov: that etherpad is not linked in the blueprint here: https://blueprints.launchpad.net/ironic/+spec/ironic-python-agent-partition17:23
comstudthe periodic task would have the same problem here...17:23
JayFOh! I see it now!17:23
JayFvkozhukalov: ^ sorry, missed the "read the full specification"17:23
comstudIt can be behind all of the worker threads and need to wait for them all to complete17:23
comstudbut it certainly should run17:23
*** rloo has quit IRC17:24
jrollright, it does run... about 20x slower than it should :)17:24
comstudyeah17:24
*** rloo has joined #openstack-ironic17:24
comstudif 64 ipmi + DB calls takes 2 minutes...17:24
comstudthe periodic task is going to be... behind17:25
jrollyeah17:25
comstud(64 being the default pool size)17:25
comstuddevananda, jroll: So the solution here is rpc pool size of 4 or something small. :)17:25
comstudfor now17:25
jrollmakes sense17:25
jrollI'll test that this afternoon17:25
jrollpoke me if you don't hear from me about it :P17:26
comstuddevananda: I was telling jroll earlier.. we found this happen with nova-scheduler and DB calls also17:26
devanandacomstud: interesting. is this how you solved it?17:26
comstuddevananda: 100 concurrent builds... the 100 DB calls happen to run before we do any processing on the return from the first one.. (we have an explicit yield after every DB call.. and it happened to cause this problem)17:26
comstudyes17:26
comstudwe lowered down to 417:26
devanandawhy 4?17:26
devanandawhy not 8?17:27
russell_hcomstud: is this just a scheduler bug?17:27
comstudwell17:27
comstudI'm not sure if I would call it a bug17:27
comstudbut it's an eventlet implementation detail17:27
devanandahm17:28
comstuddevananda: well17:28
comstuddevananda: it's a guess.  but..17:28
comstudsay IPMI + DB call takes 1s17:28
comstud8 threads means something could be waiting 8 seconds17:28
comstud4 threads means something could be waiting 4 seconds17:28
devanandacomstud: ipmi can reasonably take *much* longer17:29
comstudassuming those worker threads are all scheduled in a row17:29
comstudwhich they are17:29
devanandalike 30 seconds17:29
comstudok, then we may need to do something to force some better scheduling.17:29
devanandawhich is why we spin off a separate greenthread to handle that, and the RPC response should be sent back from conductor separately17:29
comstudright17:30
comstudwell17:30
devanandabut when we designed that, i dont think any of us were aware of this eventlet scheduling detail17:30
comstudI think ditching the pool would maybe help here.. I was going to model that out.17:30
comstudand see what the scheduling would look like17:30
comstud(ditcht he worker pool)17:30
comstudAnd just call eventlet.spawn()17:30
*** foexle has quit IRC17:30
devanandahm17:30
comstudbecause that would keep the RPC pool from blocking17:30
comstudhowever...17:30
comstudi dunno maybe worker pool just needs to be 2 * rpc pool size17:31
comstudor something17:31
russell_hso that would exacerbate the issue if you did wind up with a ton of threads blocked on DB calls right?17:31
comstudyeah, you don't really want to stack up infinite numbers of threads, probably17:31
devanandacomstud: why would calling eventlet.spawn() vs GreenPool.spawn() be different here?17:31
comstudand you'll run out of resources doing a million fork()/exec for ipmitool17:31
comstuddevananda: The RPC pool thread is blocking on the worker pool... an available thread17:32
russell_his there a reasonable facility for putting db calls into OS threads?17:32
comstuddevananda: so, going back on what I said.. I think maybe things are better if the worker pool is a multiple of the rpc pool17:32
devanandacomstud: wait. it's blocking trying to get a free worker? no... it raises ane xception17:32
comstudlower rpc pool down... and keep worker pool higher17:32
comstuddevananda: huh17:32
devanandacomstud: there's an explicit limit to the worker pool size to avoid infinite resource consumption there17:32
devanandacomstud: see manager._spawn_worker17:33
comstud66         if self._worker_pool.free():17:33
comstudheh17:33
devanandaright17:33
comstudso i missed this17:33
vkozhukalovdevananda: can we place this bp https://blueprints.launchpad.net/ironic/+spec/ironic-python-agent-partition as a dependency into bp tree here  https://blueprints.launchpad.net/ironic/+spec/utility-ramdisk?17:33
comstudWell, what the hell17:33
devanandacomstud: in my tests, it's easy to hit that exception17:33
devanandabut the RPC messages are still not getting returned timely17:33
comstudI think there's another detail here I'm missing... with the RPC response17:34
comstudi'm assuming it's not causing a switch17:34
comstudbut it could be17:34
*** tatyana has joined #openstack-ironic17:34
comstudeither way.. The problem has to be something with the RPC pool threads...17:34
comstudnot running until after all conductor workers finish17:35
*** zdin0bot has quit IRC17:35
comstudSomehow we're getting into a state where the scheduling list in eventlet is in that order17:35
comstudworker threads all in a row -> rpc pool threads all in a row17:35
comstudI think that's the only thing that makes sense here17:35
*** overlayer has quit IRC17:36
comstudbut17:36
devanandavkozhukalov: done. also updated status of that bp17:36
comstudThe fact that you raise on no workers nixes a big part of my hypothesis17:36
comstud:)17:36
devanandaheh17:37
comstudI would try a lower rpc pool size, however... drop it down to like 417:38
comstudAnd see if you get API responses more quickly17:38
comstudand maybe raise # of workers to be a mulitple of rpc pool size17:39
comstud(give it its own conf setting)17:39
*** rloo has quit IRC17:39
*** rloo has joined #openstack-ironic17:40
*** dwalleck has quit IRC17:40
comstudanyway, i gotta run an errand real quick17:40
comstudbbiab17:40
devanandacheers, thanks for the investigation17:41
devanandai'll test those out on my other machine shortly17:41
russell_hdevananda: filled this out some: https://blueprints.launchpad.net/ironic/+spec/nodeless-vendor-passthru17:41
devanandarussell_h: much better, thanks!17:43
devanandarussell_h: oh. ya'll can also mark the implementation as 'started' when you post code for a BP -- i just did it for that one17:44
russell_hthanks17:45
devanandarussell_h: afaik, that field is independent of direction/definitino/etc17:45
devanandaand can bet set by the bp author17:45
openstackgerritDmitry Tantsur proposed a change to openstack/ironic: Implement caching for master images  https://review.openstack.org/8538717:47
*** zdin0bot has joined #openstack-ironic17:49
*** zdin0bot has quit IRC17:49
*** zdin0bot has joined #openstack-ironic17:49
JayFFYI: I'm going to be out and mostly-unreachable from this afternoon through Tuesday morning.17:49
dtantsurGuys, could someone comment on the direction I am going to with https://review.openstack.org/#/c/85387/ ? (It is by no means finished or close to be finished)17:52
dtantsurThanks in advance :)17:52
devanandahuh, it looks like http://docs.openstack.org/developer/ironic/ is suddenly out of date?17:53
devanandaor i've got a bad cache somewhere in the way17:54
NobodyCamhumm restoreing from backup maybe?17:55
openstackgerritA change was merged to openstack/ironic: Update ironic.conf.sample  https://review.openstack.org/8835317:55
openstackgerritA change was merged to openstack/ironic-python-agent: Use docker import/export to make image smaller  https://review.openstack.org/8781917:56
devanandaJoshNang: can you join #openstack-swift?18:01
devanandaJoshNang: we're talking about the tempurl patch18:01
JoshNangdevananda: certainly18:02
*** ilives has quit IRC18:03
*** ilives has joined #openstack-ironic18:04
openstackgerritVladimir Kozhukalov proposed a change to openstack/ironic-python-agent: Added disk utils  https://review.openstack.org/8616318:09
vkozhukalovJayF: changed commit message https://review.openstack.org/#/c/86163/ and added comments  into the whiteboard  https://blueprints.launchpad.net/ironic/+spec/ironic-python-agent-partition18:10
JayFI'll make sure to have a look before I leave today -- but I'm out until Tuesday. I'll trust jroll and other cores to tkae my concerns into account if they want to +2 while I'm gone :)18:11
*** zdin0bot has quit IRC18:15
vkozhukalovJayF: ok, thanks18:15
openstackgerritJay Faulkner proposed a change to openstack/ironic-python-agent: Uniquely identify builds to prevent race  https://review.openstack.org/8838318:16
*** rloo has quit IRC18:17
*** rloo has joined #openstack-ironic18:17
*** rloo has quit IRC18:26
*** rloo has joined #openstack-ironic18:27
*** rloo has quit IRC18:29
*** rloo has joined #openstack-ironic18:29
NobodyCambrb18:29
*** rloo has quit IRC18:31
*** rloo has joined #openstack-ironic18:32
*** tatyana has quit IRC18:33
*** vkozhukalov has quit IRC18:34
* devananda gets brunch18:48
openstackgerritTushar Kalra proposed a change to openstack/ironic: cleanup docstring for drivers.utils.get_node_mac_addresses  https://review.openstack.org/8839518:49
NobodyCamenjoy devananda :)18:49
Shrewsi was so deep in concentration on my work that my laptop battery just died on me. doh!18:53
NobodyCamdoh :-p18:53
* Shrews blames NobodyCam18:53
*** dwalleck has joined #openstack-ironic18:54
NobodyCamheheh :)18:54
*** dwalleck_ has joined #openstack-ironic18:57
*** athomas has quit IRC18:57
*** dwalleck has quit IRC18:59
openstackgerritDavid Shrewsbury proposed a change to openstack/ironic: Fix bypassed reference to node state values  https://review.openstack.org/8840319:05
comstudback19:08
NobodyCamwb19:08
*** hemna_ has quit IRC19:14
rloodevananda: did you do anything to get this updated http://docs.openstack.org/developer/ironic/ ?19:17
* rloo sees a period missing. Shouldn't have +2'd it.19:17
*** ilives has quit IRC19:22
*** rloo has quit IRC19:24
*** rloo has joined #openstack-ironic19:25
*** zdin0bot has joined #openstack-ironic19:25
comstudcould use a 2nd +2 on this: https://review.openstack.org/#/c/87076/19:26
*** hemna_ has joined #openstack-ironic19:27
* NobodyCam looks19:27
comstudty19:28
*** zdin0bot has quit IRC19:30
devanandarloo: i don tknow why the docs page seems to have reverted19:30
*** rloo has quit IRC19:30
devanandarloo: welp. now it's back.19:31
*** rloo has joined #openstack-ironic19:31
*** rloo has quit IRC19:32
*** killer_prince has quit IRC19:32
*** rloo has joined #openstack-ironic19:32
*** rloo has quit IRC19:33
*** rloo has joined #openstack-ironic19:34
*** lynxman has quit IRC19:34
*** lynxman has joined #openstack-ironic19:34
NobodyCamany reason to add a node in maintenance mode to these tests?19:35
comstudno, because the DB call filters it19:35
comstudin this patch19:35
comstudbut19:35
comstudyou'll see in my follow up patch.. I change this19:35
comstudand add a test19:35
comstudI mean.. the code I'm testing here has no checks on 'maintenance' mode at all in the code path.19:37
NobodyCamahh so there is more to come :) (oh and I love the lambda trick)19:37
comstud(other than we pass 'maintenance': False to get_nodeinfo_list)19:37
comstud(which is tested)19:37
*** rloo has quit IRC19:38
*** rloo has joined #openstack-ironic19:38
comstudyeah, in my follow up patch I change code so that we're actually checking maintenance19:38
comstudto try to avoid races... so I add a test there19:38
comstud(that patch is already +2d -- but this one has to land first :)19:39
comstud(er +A'd)19:39
*** killer_prince has joined #openstack-ironic19:40
adam_guh oh19:40
adam_ghttps://bugs.launchpad.net/ironic/+bug/130915619:40
rlooadam_g: uh oh. Easy to fix though.19:41
jrollthat's.... fun19:41
comstudthat's kind of nasty19:41
adam_grloo, yeah, and backport19:41
rlooadam_g: not sure it needs to be backported. The change was made recently.19:42
adam_goh? cool19:42
rlooadam_g: like maybe today. there were two places. let me check.19:42
adam_gis there even going to be a long-lived stable/icehouse branch of ironic?19:42
adam_g828060ff79f3f96f143331e3054f66685e5f562d19:43
rlooadam_g: i was wrong. one of the changes hasn't gone through yet: https://review.openstack.org/#/c/84687/19:43
adam_grloo, oh, nice. will test those now19:44
rlooadam_g: so to fix it, that review needs to be modified to allow for power on and off.19:44
NobodyCamcomstud: LGTM19:44
NobodyCamoh seems like its afternoon walkies time.. brb19:45
rlooadam_g: but I know there was a discussion on IRC about whether it should be power on and off, or just power off. but i didn't pay attention to it. Guess it needs to be on and off though.19:45
comstudNobodyCam: thnx!19:45
rlooadam_g: just looking at https://bugs.launchpad.net/ironic/+bug/1309048. Seems like there could be a problem if node is power on...19:47
devanandarloo: hi! i just realized https://review.openstack.org/#/c/73005/ got dropped19:48
devanandarloo: i'd love to see that re-opened -- i think lucas and I were both almost ready to land it a week ago19:48
rloodevananda: yeah. I didn't think it was as important as some other stuff.19:48
rloodevananda: ok, I'll fix it next week then. (am off in a few hours, til next Tues.)19:49
adam_grloo, yeah, but that may not be an issue. eg, an admin has powered the node on to do some maintenance on it out of band from nova?19:49
devanandarloo: you're probably right, but it's good clean up, both in the driver code and in the API19:49
rlooadam_g: I just skimmed that bug, don't know the details and haven't thought it through. wanted to point it out to you just in case...19:49
devanandaadam_g: not sure I grok that bug yet19:52
devanandagah19:53
* devananda re-reads the definition of nova.virt.driver:node_is_available19:53
devananda: Return whether this compute service manages a particular node.19:53
adam_grloo, powering the node on seems to work fine (with that patch). removed from hypervisors resources until it is powered off again, and n-cpu restarts just fine19:53
NobodyCamdevananda: are you also looking at https://review.openstack.org/#/c/84687/19:54
*** vkozhukalov has joined #openstack-ironic19:55
adam_gdevananda, yeah, just conflicting reports from the driver about what exists and whats available19:55
adam_gs/available/manageable19:55
adam_g84687 seems to fix it here19:55
rlooadam_g: oh, so both had to be in sync wrt power off only?19:57
rlooadam_g: both functions.19:57
devanandaI should have looked at / remembered how nova.compute.manager handles resources19:58
devanandare-reading the code now, hiding a node that is functional but provisioned doesn't seem the way that nova expects it to work19:59
devanandait should be presented as totally-utilized, not hidden19:59
adam_grloo, right, well currently get_available_nodes() lists any node /w a power state (OFF or ON)19:59
*** killer_prince has quit IRC19:59
devanandahttps://github.com/openstack/nova/blob/master/nova/compute/manager.py#L544120:00
adam_gdevananda, yeah, with that patch i'm seeing nova deleting orphan compute nodes for those that are in use20:00
devanandaright20:00
devanandait shouldn't be doing that20:00
devanandathey're not orphans - -they're in use20:00
devanandathat patch is fixing it the wrong way, IMO20:00
adam_gmaybe the kind of availability check we're doing now should be moved into the calculation of available resources20:01
devanandayes20:02
devanandathough i believe there's an open bug that the resourcemanager isn't working right now20:02
*** zdiN0bot has joined #openstack-ironic20:03
openstackgerritA change was merged to openstack/ironic-python-agent: Better documentation and defaults for coreos image  https://review.openstack.org/8836420:03
openstackgerritThomas Bechtold proposed a change to openstack/ironic: Improve error handling for conf sample generation  https://review.openstack.org/8842020:04
comstuddevananda: ya, that's what I was going to say.. show them to compute manager as fully utilized20:06
comstudmaintenance ones that are not provisioned... maybe you can hide those.. although20:07
comstudhm, ya.20:07
*** killer_prince has joined #openstack-ironic20:08
devanandaoh hum20:09
devanandawhat if a provisioned instance is put into maintenance mode in ironic...20:09
devanandanova is suddenly goign to NOT see that resource20:09
devanandabut still see its instance20:09
devananda:(20:09
comstudright20:10
comstudwhich is why i think you still present it.. but fully utilized (as it would be with an instance)20:10
comstudelse you get the bomb on nova-compute start, it appears20:11
comstudnova's ComputeNode table maybe should have a field for maintenance or disabled20:11
comstudinstead of just the service record20:11
comstudactually, maybe it does20:12
devanandaadam_g: should we revert 84214 while waiting for a fix?20:14
*** jistr has joined #openstack-ironic20:15
adam_gdevananda, maybe? im hacking ontop of lucas' patch atm20:15
devanandaadam_g: try http://paste.openstack.org/show/76169/20:17
adam_gdevananda, yeah, i have that already. that will throw resource tracking off though, working on that20:17
devanandak20:18
adam_gdevananda, this is what i had in mind, http://paste.ubuntu.com/7270560/20:24
adam_gnot working just yet, need to run out. back in 2020:24
*** epim has joined #openstack-ironic20:25
devanandaadam_g: this n_list = icli.node.list(maintenance=False, associated=False) doesn't seem to fit20:25
NobodyCamoh20:25
devanandanova doc string says20:26
devanandaReturns nodenames of all nodes managed by the compute service.20:26
devanandaand update_available_resources starts with the list returned by get_available_nodes20:26
devanandaif a node isn't in taht list, it gets deleted from the resourcetracker20:26
adam_gdevananda, yeah--that was just a quick mod to 84687. will continue when i get back, f you haven't beat me to it :)20:27
devanandai actually need to step away for a while as well, probably about an hour20:29
devanandagoign to test comstud's theory first, tho20:31
*** romcheg1 has quit IRC20:32
*** jdob has quit IRC20:34
openstackgerritJosh Gachnang proposed a change to openstack/ironic: Adding a reference driver for the agent  https://review.openstack.org/8479520:37
openstackgerritChris Krelle proposed a change to openstack/ironic: Fix for tripleO undercloud gate tests DO NOT MERGE  https://review.openstack.org/8552920:41
devanandacomstud: setting rpc_thread_pool=4 and conductor.worker_pool=64, even with ~100 nodes and spamming multiple requests for each node, with a 10-second utils.execute(sleep), it's fine20:42
devanandajroll: ^20:42
devanandataht seems to solve your conductor stops heartbeating problem20:43
jrollniiiice20:43
devanandaand my API requests are not blocking either20:43
NobodyCambbiafm20:43
devanandajroll: so to be clear, this works for utils.execute() calling ipmitool. it does not address slow DB queries20:44
devanandabut at least locally, with 100 nodes, i'm not seeing any20:44
jrolldevananda: right20:44
jrollthanks for testing that :)20:44
devanandajroll: i'm also seeing >30 nodes with power operations in progress at the peak of my test20:46
devanandathere's some throttling happening... that number is quite likely an artefact of my test env20:46
openstackgerritTushar Kalra proposed a change to openstack/ironic: Remove hardcoded node id value  https://review.openstack.org/8843320:47
*** eguz has joined #openstack-ironic20:49
*** rloo has quit IRC20:49
*** rloo has joined #openstack-ironic20:49
*** eghobo has quit IRC20:53
devanandajroll: commented on the bug https://bugs.launchpad.net/ironic/+bug/1308680 -- are you guys going to post a fix?20:53
jrolldevananda: yeah, would be happy to20:55
jrolljust assigned it to myself20:56
jrolldevananda: would you be backporting this?20:56
jrollprobably not with a new conf option, right?20:56
linggaoHi ironic, why is ironic.conf.sample  not up to date on the master branch?21:06
linggaoI just git pull origin master, and tox shows that ironic.conf.sample is not up to date.21:07
openstackgerritChris Krelle proposed a change to openstack/ironic: Fix for tripleO undercloud gate tests DO NOT MERGE  https://review.openstack.org/8552921:08
jrolllinggao: that checks against your local copy of ironic.conf.sample21:09
jrollI believe21:10
NobodyCamdoh guess have to close parentheses when you open them :-p21:10
jrolloh no, you're right, linggao, my mistake21:10
devanandalinggao: i fixed that earlier today, i believe https://review.openstack.org/#/c/88353/21:10
jrollthat's not good :|21:10
devanandadid it break twice in one day?21:10
matty_dubsNobodyCam: Heh, I over-did that earlier today!21:10
jrolloh21:11
matty_dubsNobodyCam: Some sort of keyboard issue or something caused me to have " if not (driver_info['virt_type'] == 'ontap_sp')))))))):"21:11
jrolldevananda: I probably need to rebuild my venv, yeah?21:11
NobodyCam:O ieek21:11
devanandayes21:11
jroll:)21:11
jrolllinggao: ^21:11
linggaodevananda, so it is merged already.21:11
devanandajroll, linggao: pip install -U -r requirements.txt -r test-requriements.txt21:11
linggaothanks jroll.21:11
devanandawoops, typo, but you getthe idea21:11
jrolldevananda: but it takes so long :( :P21:11
devanandajroll: that's faster21:12
devanandajust update, dont rebuild the whole thing21:12
jrollI assumed the long part was pip21:12
jrollohhh, right, it won't reinstall everythign21:12
jrollgood call, thanks21:12
*** eguz has quit IRC21:12
* jroll is not sure his brain is at 100% right now21:12
*** eghobo has joined #openstack-ironic21:13
dividehexDoes anyone know if the consistent hash prevents nodes from mapping to conductors not in the same availability zones?21:13
*** jbjohnso has quit IRC21:13
devanandadividehex: ironic conductors don't know anything about nova host aggregates21:13
russell_hdividehex: it doesn't21:13
devanandaso, no21:14
comstuddevananda: cool :)21:14
devanandadividehex: unless your AZs use separate DB's21:14
dividehexok thanks21:14
openstackgerritDevananda van der Veen proposed a change to openstack/ironic: Better handling of missing drivers  https://review.openstack.org/8357221:14
russell_hI think we'll eventually wind up wanting to do something like that, but I don't think it'll be based on host aggregates21:15
devanandarack-aware scheduling :)21:16
russell_hif anything, you might base host aggregates on the same location info Ironic uses21:16
*** matty_dubs is now known as matty_dubs|gone21:16
rloorussell_h: didn't realize you are around. I just commented on your review 81919. Let me know what you think.21:16
linggaodevananda, I did pip install -U -r requirements.txt -r test-requriements.txt, and still got ironic.conf.sample is not up to date.21:16
russell_hdevananda: right, but not just scheduling of instances, but also of conductor mapping21:16
devanandalinggao: sorry, i had a typo -- it should be "pip install -U -r test-requirements.txt"21:16
*** eguz has joined #openstack-ironic21:16
russell_hprobably "scheduling of conductor mapping" makes no sense21:17
devanandarussell_h: mapping based on proximity21:17
russell_hright21:17
russell_hpeople are going to want to do top of rack conductors or whatever21:17
devanandasure21:17
russell_hanyway, that seems like something that can be reasonably added later21:17
devanandabut that creates a SPoF for each rack21:17
devanandaand not having that SPoF is sorta the point ...21:17
russell_hright, you'd still want a fallback21:17
russell_hthats where it gets... more fun21:18
devanandaso a preferred-conductor21:18
russell_hright21:18
devanandaso then we have a different mapping function. hash-based and proximity-preference-based21:18
linggaodevananda, same error.21:18
russell_hit could literally be that every node maps to the conductor at the top of its rack, with its secondary in the next rack over21:18
devanandalinggao: hmm... what is the diff?21:18
jrolllinggao: your connection to pypi must be much better than mine :)21:18
russell_hbut making it suitably generic sounds hard21:18
devanandarussell_h: we need rack discovery and such first, but yes, i agree that would /eventually/ be nice :)21:19
russell_hwhich is why I'm really happy to not worry about it for now21:19
russell_hright21:19
linggaolet me try the diff...21:19
russell_hrloo: let me take a look21:19
russell_hrloo: oohh, thank you!21:19
russell_hI'll work through this21:20
russell_hI didn't realize which review that was at first21:20
rloorussell_h: that's the first review of many, right? :-)21:20
*** eghobo has quit IRC21:21
*** eguz has quit IRC21:21
*** eghobo has joined #openstack-ironic21:22
russell_hrloo: so, regarding your top-level comment, I thought some about that, but I like keeping them separate more21:22
russell_hfundamentally, I view per-node vendor passthru methods as totally different methods than per-driver ones21:22
jrollJoshNang: this maybe needs to be fixed? https://review.openstack.org/#/c/84795/25/etc/ironic/ironic.conf.sample21:22
NobodyCamseems odd that that pxeandSeamicroDriver is the only class that checks if it loaded :-p21:23
russell_hso making them the same actual method on the VendorInterface gets odd21:23
jrollJoshNang: just regen the sample config21:23
JoshNangheh that's what i did.21:23
russell_hrloo: in an ideal world, I would rename "vendor_passthru()" to "node_vendor_passthru()"21:23
JoshNangand gave me that.21:23
russell_hrloo: but thats a whole can of worms I don't want to get into21:23
rloorussell_h: Ok, so a vendor passthru method is specifically either geared towards the node, or not.21:23
jrollJoshNang: pip install -U -r requirements.txt -r test-requirements.txt21:23
russell_hright21:23
jrollJoshNang: those are changes in keystone21:24
russell_hrloo: the best example I have of this is what we're doing in the agent driver, where we need a way to look up what node an agent belongs to21:24
JoshNangah awesome21:24
rloorussell_h: so if method1 is not node-specific, you cannot invoke it via nodes/.../vendor_passthru.21:24
russell_hcorrect21:24
rloorussell_h: and it is/will be obvious to whomever invokes the methods, to know what the method names are and whether to go via /nodes/.. or /drivers/...21:25
russell_hin the agent driver, we pass in a JSON document with MAC addresses and serial #s and whatnot to /drivers/.../vendor_passthru/lookup21:25
russell_hand the agent driver tries to find the node in the database21:25
rloorussell_h: ok. fair enough.21:25
devanandaNobodyCam: yes, we should do the same for pyghmi, and PXE should check the presence of /tftpboot/ and tfptd and such during its __init__ too21:25
NobodyCamso devananda how many +2's does 83572 need. If I add mine that will make 421:25
russell_hand if it doesn't, it'll can optionally add it (not sure if that code is written yet)21:25
NobodyCamdevananda: yes!21:26
russell_hrloo: but it wouldn't make sense to do like /nodes/.../vendor_passthru/lookup21:26
devanandaNobodyCam: lol - that failed to merge, i just rebased it. if it passes jenkins, we should +A it21:26
russell_hthats how I'm thinking about it anyway21:26
NobodyCamlol ack will keep an eye on it and land it21:26
*** Mikhail_D_ltp has quit IRC21:26
jrollrussell_h: wdyt about making both passthru things at /drivers/.../vendor_passthru/method, and adding ?node=uuid if it's for a node?21:27
jrolllucas mentioned this the other day21:27
openstackgerritDevananda van der Veen proposed a change to openstack/ironic: Use DiskPartitioner  https://review.openstack.org/8339921:27
rloorussell_h: but even with that, why at the driver level, do you need to separate into (node_)vendor_passthru() and driver_vendor_passthru(). Unless that makes it easier to check/distinguish the two.21:27
linggaodevananda,  here is the diff. http://paste.openstack.org/show/76174/.21:28
linggaodevananda, something wrong with my keystone level again. It had the same problem but magically healled.21:28
devanandalinggao: yes, the python-keystoneclient library in your environment needed to be updated21:29
*** zdiN0bot has quit IRC21:29
jrolldevananda, linggao: JoshNang's patch that I mentioned above has the same diff :/21:29
devanandalinggao: that wil lsolve your problem21:29
devanandathe ironic.conf.sample on the master branch is correct right now21:29
rloorussell_h. I don't see why you can't change vendor_passthru() to node_vendor_passthru() then. but that could be a future change I guess.21:29
jrolloh hmm21:29
russell_hrloo: right21:29
devanandaso anyone getting that error has an out-of-date venv21:29
JoshNangjust did pip -U... and generate but nothing changed21:29
russell_hrloo: I'd worry about anyone with an out-of-tree driver21:30
JoshNangi just rebuilt my whole vagrant box this morning (whoops)21:30
russell_hI don't know the policy on that, are driver interfaces meant to be stable?21:30
jrolldevananda: what is the "correct" version of python-keystoneclient for this?21:30
rloorussell_h: How could someone have an out-of-tree driver but updated conductor/etc?21:30
russell_hbut in any case, IMO it belongs in a separate change if it should happen at all21:30
jrolldevananda: I wonder if maybe keystone did another release and now our sample is wrong again21:30
devanandajroll: i see 0.8.0 locally21:30
russell_hrloo: drivers just get loaded from your python path21:30
jrolldevananda: yay version ranges in global requirements.21:30
russell_hso they can be versioned entirely independently of Ironic21:30
devanandayea...21:30
russell_hrloo: you just drop some info in your setup.cfg, install it somewhere on your python path, and it'll show up to Ironic as a driver21:31
devanandarussell_h: ironic driver API is meant to be stable at certain points21:31
russell_hstevedore is actually kind of neat21:31
linggaomine is  0.7.121:31
russell_hright, thats sort of what I figured21:31
rloorussell_h: ah yes, that's true. Although I don't think anyone has been worried about that since eg there's some new ManagementInterface coming up.21:31
russell_hwe could do a deprecation thing21:31
jrolllinggao: pip install -U python-keystoneclient==0.8.021:32
devanandagiven that folks are writing out-of-tree drivers, we need to have deprecation plan / cycle for driver API changes21:32
russell_hrloo: so, I think (although I can't swear I verified this) that the ManagementInterface is optional21:32
devanandaor just don't break it21:32
russell_hrloo: so if you're missing it, things will still work, although... if things get moved over into it... it might not :)21:33
rloorussell_h: I think it is optional, but I think some functions are going to move to that interface from whatever existing interface. Not sure though.21:33
russell_hright21:33
devanandaso the purpose of vendor-passthru, IMO, is to allow methods to "incubate" there21:34
JoshNanginstalled python-keystoneclient==0.8.0, still same sample config :/21:34
rloodevananda. can we add a note somewhere, that the driver API isn't yet stable or something like that?21:34
jrollJoshNang: o.o21:34
devanandaif we decide taht a given method is critical to the functionality of all drivers, it should mvoe to a core interface21:34
* jroll is *still* upgrading things21:34
russell_hat risk of sounding really aggressive, and as someone running a basically out-of-tree driver, kind of hypocritical, Ironic is very young to get bogged down worrying too much about backwards compatibility with itself21:34
devanandarloo: what part of the core driver API is not stable?21:34
devanandaJoshNang: is 14c61d69a20c9afaee8e534f0143874f751d1783 in your git history?21:35
*** rloo has quit IRC21:35
jrolldevananda: it could be argued that vendor_passthru is not stable (ever)21:35
devanandajroll: precisely21:35
jrollor more undefined then unstable21:35
*** rloo has joined #openstack-ironic21:35
rloodevananda. Nothing yet. russell_h and I were discussing backwards compatibility if we change the driver API.21:35
*** jrist has quit IRC21:36
JoshNangdevananda: "Update ironic.conf.sample" yup21:36
devanandavendor_passthru should never be considered stable -- it's whole purpose is to allow experimentation. perhaps we should rename it that way21:36
jrollrloo: are you saying we should document that it is not stable, and not worry about backwards compat?21:36
devananda(kidding on the rename)21:36
*** jrist has joined #openstack-ironic21:36
russell_hdevananda: the last service I built has a "/__experimental" namespace :)21:36
* NobodyCam was typing reply to rename comment21:36
linggaodevananda, jroll, installed 0.8.0, still has the same error. So I'l have to leave it as is since I am not changing the sample file anyway. :). It will work tomorrow just like last time. ;)21:36
rloojroll: anything so we can move forward w/o having to worry ;)21:36
linggaothanks for the help21:36
jrolllinggao: hm...21:37
russell_hdevananda: which hasn't stopped anyone from depending on it...21:37
openstackgerritlinggao proposed a change to openstack/ironic: Handle API error for XML client  https://review.openstack.org/8844821:37
russell_hespecially when they discovered the JSON Home API hidden in there21:37
jrollrloo: I'm ok with that, to be honest, but I like moving fast :)21:37
devanandarussell_h: the alternative here is not to allow *anything* in the API until it's deemed stable21:38
devanandanova addressed this with API extensions21:38
linggaohave to go. good night everyone.21:38
rloodevananda: this came up, cuz russell_h is adding a driver_vendor_passthru to VendorInterface, and I first thought why doesn't everything go through vendor_passthru(). But if we need to have two passthru's, then maybe vendor_passthru -> node_vendor_passthru. Dunno.21:38
NobodyCamnight linggao21:38
jrollnight ling21:38
devanandawhich ended up being a bit of a nightmare for them, since a) lots of extensions essentially becaome "core" functionalty, but stayed nominally extensions, and b) different deployers enabled different entensions21:38
devananda*extensions21:38
devanandaleading to different functionality across nova implementations of the same version21:38
*** linggao has quit IRC21:39
russell_hdevananda: I'm on your side here, you sold me on vendor_passthru a long time ago21:39
devanandarussell_h: :)21:39
russell_hthe particular implementation is a little ghetto, but basically fine21:39
devanandarloo, russell_h: by the way, did you see my reply on the BP?21:39
russell_hyep, replying to the reply now21:39
russell_hactually for like the last 2 hours21:40
russell_hkeep getting distracted21:40
* rloo looks now.21:40
devanandarloo: i get the reason for a driver-vendor-passthru. i think the name is too long, but the functionality seems correct - -for their use case. i'm concerned it breaks other uses though21:40
devananda(also, names are hard)21:40
*** rloo has quit IRC21:41
*** rloo has joined #openstack-ironic21:41
*** rloo has joined #openstack-ironic21:42
russell_hdevananda: replied to the BP21:43
russell_hnow, with my name on it21:43
devananda:)21:45
russell_hblueprints are horrible21:45
devanandayep21:45
*** vkozhukalov has quit IRC21:47
NobodyCamdevananda: lol... Discuss summit proposals Then do anything else, if we have time....21:49
jrolluh oh21:50
NobodyCamshould we add the bp repo / template idea to agenda?21:50
* jroll meant to put in more summit proposals today21:50
devanandaNobodyCam: yes21:52
devanandajroll: do it. sudnay's the deadline21:52
devanandaalso, we have 14 proposals right now and ONLY FOUR SLOTS21:52
devanandaso please, add more :)21:52
devananda(no, really)21:52
russell_hheh21:53
adam_gdevananda, what was the bug you mentioned re resource tracking?21:53
jrolldevananda: I know :|21:53
jrollI'll find time. somewhere. somehow.21:53
*** newell has quit IRC21:55
*** foexle has joined #openstack-ironic21:56
*** Haomeng has quit IRC21:57
devanandaadam_g: i'm not finding it right now...21:57
devanandaadam_g: IIRC, someone in -nova was trying to remove the host_manager classes since they aren't used any more, and then noticed that ironic seems to /think/ they're being used21:58
adam_ghmm21:58
comstudThe host manager class is used21:59
adam_git looks like exposing nodes /w no associated instance but powered on or in maitanence mode really screws with nova's resource tracking21:59
comstudsomeone would be on crack21:59
devanandacomstud: i may be completely misstating the problem22:00
comstudI saw something maybe related to no-db schedulding22:00
comstudor a discussion based off of it22:00
comstuddevananda: np, just confirming it's definitely used22:00
comstud:)22:00
devanandacomstud: i believe their conjecture was that http://git.openstack.org/cgit/openstack/ironic/tree/ironic/nova/scheduler/ironic_host_manager.py#n56 is not used22:01
comstudoh, ironic in specific, k22:02
devanandacomstud: ah. it's teh HostStates class22:03
devanandai dont see anywhere that it's used22:03
devanandain nova22:03
devanandagah22:04
devanandait helps if I don't have a typo in my grep line22:04
devanandaignore me -- it's totally used22:04
comstud:)22:04
NobodyCamdevananda: wanta take a quick look at : https://review.openstack.org/#/c/8837022:05
rloodevananda: I think it was this email about BaremetalHostManager http://lists.openstack.org/pipermail/openstack-dev/2014-April/031903.html22:06
russell_hrloo: I seem to type a lot of "i"s and "y"s out of order22:06
rloorussell_h: there weren't that many, were there? ;)22:06
devanandarloo: yes! that's it :)22:06
devanandaadam_g: ^22:06
adam_gah22:07
adam_gokay thats a separate issue than what im looking at22:08
rloobtw russell_h, because I added a comment to your review, you should probably reply so others know. And when I'm back on Tues, I hope that it will have been approved ;)22:08
adam_gfor ironic to expose nodes /w power_state ON or maintenance mode, nova will need a corresponding patch to its resource tracking22:09
devanandaNobodyCam: I dont think validate() is going to set the power state22:09
NobodyCamhumm22:10
devanandaadam_g: nova.virt.ironic, or nova.compute.resource_tracker?22:10
adam_gdevananda, nova.compute.resource_tracker22:10
devanandaadam_g: that's not pluggable, which means we cant do that out of tree22:11
devanandatime to get a patch merged to nova22:11
adam_gdevananda, the dict ironic return for resources contains keys that get reset by _update_usagte_from_instances()22:11
adam_ghttp://git.openstack.org/cgit/openstack/nova/tree/nova/compute/resource_tracker.py#n59622:12
adam_gattempting to calculate powered on or nodes in maitanence as part of the used resources from the driver does not good wrt actual resource tracking22:12
NobodyCamdevananda: https://github.com/openstack/ironic/blob/master/ironic/drivers/modules/ssh.py#L371-L37222:13
devanandaoh, hah22:13
devanandaadam_g: so even though our HostManager is consuming all the resources, nova is overriding that?22:13
NobodyCamwell22:13
NobodyCamthat check the ssh connection22:14
adam_gdevananda, looks that way, at least when there are no instances associated iwth the nodes22:14
devanandaNobodyCam: right. it doesn't set the node.power_state property22:14
devanandaNobodyCam: try just setting the power state there :)22:17
devanandaNobodyCam: though if you need error detection, you should set-and-poll22:18
NobodyCamlol that would be easy.. lol.. face in palms :-p22:18
devanandaok, i need to run for a while... didn't take a break earlier when i should have.22:19
devanandabbiah22:19
*** jgrimm has quit IRC22:20
*** dwalleck_ has quit IRC22:24
NobodyCamI to am going to run22:25
*** zdiN0bot has joined #openstack-ironic22:28
openstackgerritJosh Gachnang proposed a change to openstack/ironic: Adding a reference driver for the agent  https://review.openstack.org/8479522:34
jrollJoshNang: I wouldn't say that method "isn't implemented", I would say there's just no validation done there.22:39
jrollbut I guess it's fine22:39
openstackgerritA change was merged to openstack/ironic: Better handling of missing drivers  https://review.openstack.org/8357222:39
openstackgerritJosh Gachnang proposed a change to openstack/ironic: Adding a reference driver for the agent  https://review.openstack.org/8479522:40
JoshNang:)22:40
jrollgahhhhhhh22:40
* jroll +1's again22:40
jrollso many patchsets22:40
JoshNangonly a couple more and i'll break 100 emails in that thread!22:41
JoshNangi wonder if gmail has a limit before it starts a new thread.22:41
jrollheh22:41
*** derekh has joined #openstack-ironic22:41
jrollit does, but I forget what it is. 150 maybe22:41
*** rwsu has quit IRC22:44
*** jistr has quit IRC22:46
*** rwsu has joined #openstack-ironic22:48
*** zdiN0bot1 has joined #openstack-ironic22:49
*** zdiN0bot has quit IRC22:49
openstackgerritA change was merged to openstack/ironic: Refactor sync_power_states tests to not use DB  https://review.openstack.org/8707622:55
openstackgerritA change was merged to openstack/ironic: Make sync_power_states yield  https://review.openstack.org/8486222:55
*** zdiN0bot1 has quit IRC22:58
*** zdiN0bot has joined #openstack-ironic23:04
*** zdiN0bot has quit IRC23:04
*** zdiN0bot has joined #openstack-ironic23:06
*** zdiN0bot has quit IRC23:13
*** zdiN0bot has joined #openstack-ironic23:15
*** zdiN0bot has quit IRC23:17
*** zdiN0bot has joined #openstack-ironic23:21
*** eghobo has quit IRC23:25
*** zdiN0bot has quit IRC23:26
*** stevehuang has quit IRC23:38
devanandaback for a bit, then will be off to dinner23:56
* devananda tests https://review.openstack.org/#/c/83399/23:57

Generated by irclog2html.py 2.14.0 by Marius Gedminas - find it at mg.pov.lt!