Tuesday, 2018-03-20

*** suresh12 has joined #openstack-nova00:00
*** germs has joined #openstack-nova00:01
*** germs has quit IRC00:01
*** germs has joined #openstack-nova00:01
openstackgerritMerged openstack/nova master: Remove old flavor_extra_specs_delete db api method  https://review.openstack.org/53970200:04
openstackgerritMerged openstack/nova master: Report client: Remove version discovery comment  https://review.openstack.org/55425300:04
openstackgerritMerged openstack/nova master: Remove version/date from CLI documentation  https://review.openstack.org/55390300:05
*** germs has quit IRC00:05
*** pooja has joined #openstack-nova00:07
*** suresh12 has quit IRC00:11
*** ircuser-1 has joined #openstack-nova00:13
poojaHi.. I am seeing an issue with NumInstancesFilter in nova scheduler (Newton release) when provisioning multiple instances in parallel (not in one batch api call)00:13
poojaThe scheduler's view of host isn't updated and so multiple instances get placed on a host, which exceeds the max_instances value set for that host.00:14
poojaIs this a known issue and is there a solution for it in ocata/pike release?00:14
poojaAppreciate any pointers or change links. Thanks!00:15
melwittpooja: are you running a single scheduler? there was a change in pike to do resource claims in the scheduler via placement. I'm not yet familiar with the "max_instances" value you mentioned though00:16
melwittokay, so it's a config option00:16
poojamelwitt: Yes, I'm running a single instance of nova-scheduler00:21
poojaThis is the filter I'm referring to - https://github.com/openstack/nova/blob/master/nova/scheduler/filters/num_instances_filter.py#L2800:21
*** itlinux has joined #openstack-nova00:23
melwittI found it too, currently looking through the code. it looks like the problem you described should be fixed as of pike with the claims in the scheduler. let me see if I can find a patch related to that specific area00:23
poojaGreat! thanks for your help looking into it, melwitt!00:24
Spaz-HomeMorning00:25
*** yamamoto has quit IRC00:26
*** gjayavelu has quit IRC00:27
*** yamamoto has joined #openstack-nova00:28
*** yamamoto has quit IRC00:28
*** liverpooler has joined #openstack-nova00:30
melwittpooja: to be honest, I'm not sure if the issue is fixed as of the new code. to be sure, it would be better to ask someone like bauzas or edleafe. here's a link to where I started tracing, if that might help in the meantime https://github.com/openstack/nova/blob/master/nova/scheduler/host_manager.py#L283-L28400:30
melwittand the consume_from_request is used in filter_scheduler.py, looking at it more in filter_scheduler.py, it does seem like it would not be resilient to the issue of parallel requests00:33
melwittit looks like the report of "num_instances" comes from a compute node stat report, which may not be updating in real-time00:35
*** gjayavelu has joined #openstack-nova00:35
*** Dinesh_Bhor has joined #openstack-nova00:37
*** hongbin has joined #openstack-nova00:40
*** jichen has joined #openstack-nova00:41
*** amodi has quit IRC00:44
*** suresh12 has joined #openstack-nova00:44
*** elmaciej_ has quit IRC00:46
*** lifeless_ has joined #openstack-nova00:46
*** lifeless has quit IRC00:46
*** elmaciej has joined #openstack-nova00:46
*** gjayavelu has quit IRC00:48
poojamelwitt: Sure, let me check with bauzas or edleafe too.00:48
*** suresh12 has quit IRC00:49
poojaYes, the problem is that stats get updated asynchronously and num_instances value used by scheduler would be inaccurate based on that timing.00:49
melwittpooja: I see. I think I understand now, and based on that, it's probably still a problem in the current code now00:50
poojaOh okay.. should I file a bug for it?00:51
*** itlinux has quit IRC00:51
poojaDo these filters work the same way with the new Placement API?00:51
melwittsome do, some don't. the [Core|Ram|Disk]Filter became obsolete in the filter scheduler once we started calling placement. because we pre-filter based on answers from placement and the scheduler claims with placement along the way00:52
melwittbut the rest of the filters run after the placement call, as they did before00:54
*** tbachman has joined #openstack-nova00:54
*** wolverineav has quit IRC00:54
*** phuongnh has joined #openstack-nova00:54
*** hiro-kobayashi has joined #openstack-nova00:55
melwittagain, bauzas, edleafe, and co are the people to chat with about that00:56
melwittif you want to open a bug, I can point them to it and ask them to comment on it. it's up to you, however you want to do it00:56
*** wolverineav has joined #openstack-nova00:57
*** Dinesh_Bhor has quit IRC00:57
*** Dinesh_Bhor has joined #openstack-nova00:57
*** odyssey4me has quit IRC00:59
*** odyssey4me has joined #openstack-nova00:59
*** Dinesh_Bhor has quit IRC00:59
*** Dinesh_Bhor has joined #openstack-nova01:02
*** wolverineav has quit IRC01:03
*** wolverineav has joined #openstack-nova01:04
*** elmaciej has quit IRC01:07
*** yamamoto has joined #openstack-nova01:07
poojaSounds good! I will connect with them and see if I need to file a bug for this. Will let you know if I do that. Thanks again!01:09
melwittcool, thanks pooja01:09
*** wxy has joined #openstack-nova01:10
*** wolverineav has quit IRC01:11
*** Kevin_Zheng has joined #openstack-nova01:14
*** gjayavelu has joined #openstack-nova01:15
*** tiendc has joined #openstack-nova01:15
*** gjayavelu has quit IRC01:17
*** suresh12 has joined #openstack-nova01:19
*** suresh12 has quit IRC01:23
*** yamamoto has quit IRC01:27
*** Dinesh_Bhor has quit IRC01:49
*** Dinesh_Bhor has joined #openstack-nova01:52
*** Dinesh_Bhor has quit IRC01:52
*** suresh12 has joined #openstack-nova01:54
*** annp has joined #openstack-nova01:56
*** suresh12 has quit IRC01:58
*** suresh12 has joined #openstack-nova01:59
openstackgerritMerged openstack/nova master: VMware: fix TypeError while get console log  https://review.openstack.org/54918201:59
*** artom_ has joined #openstack-nova02:01
*** Dinesh_Bhor has joined #openstack-nova02:01
*** dikonoor has joined #openstack-nova02:02
*** Dinesh_Bhor has quit IRC02:02
*** germs has joined #openstack-nova02:02
*** artom has quit IRC02:03
*** germs has quit IRC02:06
*** Dinesh_Bhor has joined #openstack-nova02:08
*** wolverineav has joined #openstack-nova02:08
*** suresh12 has quit IRC02:10
*** dikonoor has quit IRC02:11
*** Nil_ has quit IRC02:12
*** fragatina has quit IRC02:15
*** chyka_ has joined #openstack-nova02:15
*** fragatina has joined #openstack-nova02:15
*** fragatin_ has joined #openstack-nova02:18
*** chyka has quit IRC02:18
*** fragatina has quit IRC02:19
*** yamahata has quit IRC02:21
*** fragatin_ has quit IRC02:22
*** chyka_ has quit IRC02:22
*** gongysh has joined #openstack-nova02:25
*** dave-mccowan has quit IRC02:31
*** hoangcx has joined #openstack-nova02:31
*** r-daneel has joined #openstack-nova02:32
*** r-daneel_ has joined #openstack-nova02:34
*** r-daneel has quit IRC02:36
*** r-daneel_ is now known as r-daneel02:36
*** suresh12 has joined #openstack-nova02:37
*** suresh12 has quit IRC02:39
*** suresh12 has joined #openstack-nova02:39
*** links has joined #openstack-nova02:40
*** itlinux has joined #openstack-nova02:42
*** suresh12 has quit IRC02:45
*** pooja has quit IRC02:46
*** salv-orl_ has joined #openstack-nova02:50
alex_xu_jaypipes: what do you think about the preferred_traits, that resovled the cyborg weigher problem02:52
*** salv-orlando has quit IRC02:52
*** hoangcx has quit IRC02:55
*** wolverineav has quit IRC02:58
*** hoangcx has joined #openstack-nova03:00
*** psachin has joined #openstack-nova03:02
*** suresh12 has joined #openstack-nova03:02
*** wolverineav has joined #openstack-nova03:04
*** yingjun has joined #openstack-nova03:07
*** suresh12 has quit IRC03:07
*** bkopilov has quit IRC03:08
*** sree has joined #openstack-nova03:13
*** sree has quit IRC03:14
*** sree has joined #openstack-nova03:14
*** wolverineav has quit IRC03:18
*** wolverineav has joined #openstack-nova03:22
*** chyka has joined #openstack-nova03:23
*** yingjun has quit IRC03:27
*** chyka has quit IRC03:28
*** gongysh has quit IRC03:42
*** suresh12 has joined #openstack-nova03:46
*** itlinux has quit IRC03:49
*** yamamoto has joined #openstack-nova03:52
*** suresh12 has quit IRC03:56
*** hongbin has quit IRC04:00
*** yamamoto has quit IRC04:01
*** germs has joined #openstack-nova04:03
*** germs has quit IRC04:03
*** germs has joined #openstack-nova04:03
*** hiro-kobayashi has quit IRC04:04
*** yamamoto has joined #openstack-nova04:04
*** germs has quit IRC04:07
*** yamamoto has quit IRC04:08
*** yamamoto has joined #openstack-nova04:09
*** yamamoto has quit IRC04:14
*** fragatina has joined #openstack-nova04:14
*** wolverineav has quit IRC04:15
*** yamamoto has joined #openstack-nova04:16
*** fragatina has quit IRC04:16
*** fragatina has joined #openstack-nova04:16
*** suresh12 has joined #openstack-nova04:17
*** suresh12 has quit IRC04:22
*** yamamoto has quit IRC04:27
openstackgerritMerged openstack/nova master: Add placeholder migrations for Queens backports  https://review.openstack.org/55383104:28
*** yamamoto has joined #openstack-nova04:28
*** vivsoni__ has joined #openstack-nova04:29
*** andreas_s has joined #openstack-nova04:29
*** andreas_s has quit IRC04:34
*** bkopilov has joined #openstack-nova04:41
*** wolverineav has joined #openstack-nova04:42
*** yamamoto has quit IRC04:43
*** suresh12 has joined #openstack-nova04:46
*** yamamoto has joined #openstack-nova04:46
*** vivsoni__ has quit IRC04:51
*** vivsoni__ has joined #openstack-nova04:53
*** Dinesh__Bhor has joined #openstack-nova04:58
*** Dinesh_Bhor has quit IRC04:58
*** wolverineav has quit IRC05:00
*** wolverineav has joined #openstack-nova05:01
*** lpetrut has joined #openstack-nova05:02
*** wolverineav has quit IRC05:05
*** ratailor has joined #openstack-nova05:06
*** imacdonn has quit IRC05:14
*** imacdonn has joined #openstack-nova05:14
openstackgerritNaichuan Sun proposed openstack/nova master: xenapi: Use XAPI pool instead of aggregate pool for shared SR migration  https://review.openstack.org/55415405:18
*** gjayavelu has joined #openstack-nova05:18
*** jmlowe has quit IRC05:18
*** hiro-kobayashi has joined #openstack-nova05:19
*** Aditya_ has joined #openstack-nova05:21
*** vivsoni__ has quit IRC05:25
*** hoangcx has quit IRC05:26
openstackgerritNaichuan Sun proposed openstack/nova master: xenapi: Use XAPI pool instead of aggregate pool for shared SR migration  https://review.openstack.org/55415405:27
*** suresh12 has quit IRC05:29
*** hoangcx has joined #openstack-nova05:29
*** Zames has joined #openstack-nova05:31
*** mdnadeem has joined #openstack-nova05:36
*** jmlowe has joined #openstack-nova05:37
*** sridharg has joined #openstack-nova05:38
*** Zames has quit IRC05:39
*** jmlowe has quit IRC05:42
*** yufei has joined #openstack-nova05:44
*** claudiub has joined #openstack-nova05:45
*** sidx64 has joined #openstack-nova05:51
*** sidx64 has quit IRC05:57
*** fragatina has quit IRC05:58
*** lpetrut has quit IRC06:00
openstackgerritPranab proposed openstack/os-vif master: Add abstract OVSDB API  https://review.openstack.org/47661206:03
*** sidx64 has joined #openstack-nova06:03
*** sidx64 has quit IRC06:04
*** germs has joined #openstack-nova06:04
*** dineshbhor__ has joined #openstack-nova06:07
*** germs has quit IRC06:08
*** OctopusZhang__ has joined #openstack-nova06:09
*** Dinesh__Bhor has quit IRC06:09
*** OctopusZhang__ has quit IRC06:09
openstackgerritOpenStack Proposal Bot proposed openstack/nova master: Imported Translations from Zanata  https://review.openstack.org/54877206:10
*** sidx64 has joined #openstack-nova06:11
*** lpetrut has joined #openstack-nova06:12
*** masber has quit IRC06:12
*** yufei has quit IRC06:12
*** sidx64 has quit IRC06:13
openstackgerritJianghua Wang proposed openstack/nova master: libvirt: Improve 'qemu-img convert' performance  https://review.openstack.org/52206706:15
*** OctopusZhang__ has joined #openstack-nova06:15
*** OctopusZhang__ is now known as yufei06:15
*** Aditya_ has quit IRC06:18
*** trinaths has joined #openstack-nova06:21
*** Jack_Iv has joined #openstack-nova06:22
*** Jack_Iv has quit IRC06:26
*** psachin has quit IRC06:27
*** ccamacho has quit IRC06:29
*** yufei has quit IRC06:37
*** sidx64 has joined #openstack-nova06:38
*** gjayavelu has quit IRC06:43
*** sidx64 has quit IRC06:44
*** masber has joined #openstack-nova06:44
*** gaoyan has joined #openstack-nova06:46
*** psachin has joined #openstack-nova06:47
*** masber has quit IRC06:47
*** masber has joined #openstack-nova06:47
*** Eran_Kuris has quit IRC06:48
*** Eran_Kuris has joined #openstack-nova06:50
*** avolkov has joined #openstack-nova06:54
*** yufei has joined #openstack-nova06:56
*** lpetrut has quit IRC06:57
*** avolkov has quit IRC06:57
*** avolkov has joined #openstack-nova06:57
*** yufei has quit IRC06:58
*** chyka has joined #openstack-nova07:00
*** sidx64 has joined #openstack-nova07:00
*** gjayavelu has joined #openstack-nova07:00
*** dineshbhor__ has quit IRC07:00
*** sidx64 has quit IRC07:01
*** suresh12 has joined #openstack-nova07:02
*** tetsuro has joined #openstack-nova07:03
*** chyka has quit IRC07:04
*** suresh12 has quit IRC07:06
*** sar has joined #openstack-nova07:08
*** abhishekk has joined #openstack-nova07:09
*** gjayavelu has quit IRC07:12
*** sahid has joined #openstack-nova07:13
*** kholkina has joined #openstack-nova07:15
*** rcernin has quit IRC07:23
*** rcernin has joined #openstack-nova07:24
*** rcernin has quit IRC07:24
*** ccamacho has joined #openstack-nova07:25
*** alexchadin has joined #openstack-nova07:30
*** trinaths has quit IRC07:31
*** tovin07 has joined #openstack-nova07:31
*** salv-orl_ has quit IRC07:32
*** salv-orlando has joined #openstack-nova07:33
*** salv-orlando has quit IRC07:33
*** salv-orlando has joined #openstack-nova07:33
*** pcaruana has joined #openstack-nova07:34
*** alexchadin has quit IRC07:36
*** alexchadin has joined #openstack-nova07:37
openstackgerritMichael Still proposed openstack/nova master: Move xenapi disk resizing to privsep.  https://review.openstack.org/55224207:45
openstackgerritMichael Still proposed openstack/nova master: Move xenapi partition copies to privsep.  https://review.openstack.org/55360507:45
openstackgerritMichael Still proposed openstack/nova master: Move image conversion to privsep.  https://review.openstack.org/55443707:45
openstackgerritMichael Still proposed openstack/nova master: We no longer need rootwrap.  https://review.openstack.org/55443807:45
openstackgerritMichael Still proposed openstack/nova master: We don't need utils.trycmd any more.  https://review.openstack.org/55443907:45
*** sidx64 has joined #openstack-nova07:52
*** lajoskatona has joined #openstack-nova07:52
*** sidx64 has quit IRC07:58
*** tesseract has joined #openstack-nova08:00
*** namnh has joined #openstack-nova08:01
*** andreas_s has joined #openstack-nova08:02
*** sidx64 has joined #openstack-nova08:03
*** AlexeyAbashkin has joined #openstack-nova08:04
*** germs has joined #openstack-nova08:05
*** germs has quit IRC08:05
*** germs has joined #openstack-nova08:05
*** sidx64 has quit IRC08:06
*** rmart04 has joined #openstack-nova08:07
*** sidx64 has joined #openstack-nova08:07
*** priteau has joined #openstack-nova08:08
openstackgerritZhenyu Zheng proposed openstack/nova-specs master: Allow abort live migrations in queued status  https://review.openstack.org/53672208:08
*** germs has quit IRC08:08
*** damien_r has joined #openstack-nova08:10
*** _pewp_ has quit IRC08:11
*** jpena|off is now known as jpena08:11
*** _pewp_ has joined #openstack-nova08:15
*** Zames has joined #openstack-nova08:19
*** ragiman has joined #openstack-nova08:22
*** sidx64 has quit IRC08:22
openstackgerritsahid proposed openstack/nova master: Revert "[libvirt] Add _get_vcpu_realtime_scheduler()"  https://review.openstack.org/55444808:22
openstackgerritsahid proposed openstack/nova master: Revert "[libvirt] Add _get_numa_memnode()"  https://review.openstack.org/55444908:22
*** trinaths has joined #openstack-nova08:24
*** gaoyan has quit IRC08:28
*** _ix has quit IRC08:29
*** _ix has joined #openstack-nova08:29
*** afaranha has joined #openstack-nova08:29
*** Kevin_Zheng has quit IRC08:30
*** redondo-mk has quit IRC08:30
sahidjaypipes: can you have a look at https://review.openstack.org/#/c/511188/08:30
openstackgerritlicanwei proposed openstack/nova master: Make nova-manage capable of syncing all cell databases  https://review.openstack.org/51927508:30
*** redondo-mk has joined #openstack-nova08:31
sahidi mean it would be nice if you cn cut on it we can make some progress, basically you brought that idea and we never see you again :)08:31
*** Kevin_Zheng has joined #openstack-nova08:31
*** moshele has joined #openstack-nova08:33
kaisers1efried: ping08:35
openstackgerritSilvan Kaiser proposed openstack/nova master: Exec systemd-run with privileges in Quobyte driver  https://review.openstack.org/55419508:37
*** tetsuro has left #openstack-nova08:40
*** _pewp_ has quit IRC08:41
*** Zames has quit IRC08:41
*** amoralej|off is now known as amoralej08:45
openstackgerritjichenjc proposed openstack/nova master: Move placement test cases from db to placement  https://review.openstack.org/55314908:46
*** _pewp_ has joined #openstack-nova08:47
*** sidx64 has joined #openstack-nova08:49
*** lpetrut has joined #openstack-nova08:49
*** cdent has joined #openstack-nova08:50
*** lpetrut_ has joined #openstack-nova08:50
*** lpetrut has quit IRC08:50
fanzhanghi, nova team. Can I add some specified parameter to control boot --min-count instances within one request on different hosts? Like forcing instances launched on different hosts?08:51
fanzhangI think there may be not functions like this, right?08:52
openstackgerritYikun Jiang (Kero) proposed openstack/nova-specs master: Complex (Anti)-Affinity Policies  https://review.openstack.org/54692508:54
*** Zames has joined #openstack-nova08:54
*** sidx64 has quit IRC08:58
*** alexchadin has quit IRC08:58
openstackgerritjichenjc proposed openstack/nova master: Avoid raise InstanceNotFound exception  https://review.openstack.org/54115208:58
*** Zames has quit IRC08:58
*** alexchadin has joined #openstack-nova08:59
*** lucas-pto is now known as lucasagomes09:00
*** ccamacho has quit IRC09:00
*** andreas_s has quit IRC09:01
*** andreas_s has joined #openstack-nova09:01
bauzasgood morning folks09:02
* bauzas is released from the magic kingdom09:02
*** tssurya has joined #openstack-nova09:02
Kevin_Zhengfanzhang, maybe you should try anti-affinity filter09:02
fanzhangKevin_Zheng: thanks so much. Reading docs about anti-affinity filter now :)09:03
openstackgerritJohn Garbutt proposed openstack/nova-specs master: Add PENDING vm state  https://review.openstack.org/55421209:05
*** alexchadin has quit IRC09:09
*** andreas_s has quit IRC09:10
*** alexchadin has joined #openstack-nova09:10
*** andreas_s has joined #openstack-nova09:13
*** baffle has quit IRC09:13
Kevin_Zhengfanzhang yw09:13
*** baffle has joined #openstack-nova09:14
*** ccamacho has joined #openstack-nova09:18
*** andreas_s has quit IRC09:22
*** andreas_s has joined #openstack-nova09:22
*** sidx64 has joined #openstack-nova09:24
*** _pewp_ has quit IRC09:24
*** _pewp_ has joined #openstack-nova09:26
*** mdbooth has joined #openstack-nova09:26
*** _ix has quit IRC09:33
*** _pewp_ has quit IRC09:33
jianghuaw_bauzas, good morning:-)09:33
bauzasjust catching up emails this morning09:34
jianghuaw_hope you enjoyed the time in the magic kingdom09:34
bauzaswell, my daughters did at least :)09:34
jianghuaw_good enough:-)09:34
*** ralonsoh has joined #openstack-nova09:35
*** abhishekk has quit IRC09:35
*** andreas_s has quit IRC09:36
*** andreas_s has joined #openstack-nova09:37
*** derekh has joined #openstack-nova09:40
*** _pewp_ has joined #openstack-nova09:40
*** _ix has joined #openstack-nova09:41
*** abhishekk has joined #openstack-nova09:48
*** josecastroleon has quit IRC09:49
*** josecastroleon has joined #openstack-nova09:50
*** hiro-kobayashi has quit IRC09:50
openstackgerritChris Dent proposed openstack/nova master: Provide framework for setting placement error codes  https://review.openstack.org/54617709:53
*** jichen has quit IRC09:53
*** liverpooler has quit IRC09:54
*** mgoddard has joined #openstack-nova09:54
*** andreas_s has quit IRC09:56
*** andreas_s has joined #openstack-nova09:57
gibimorning nova10:02
*** yamamoto has quit IRC10:04
*** alexchadin has quit IRC10:04
*** germs has joined #openstack-nova10:05
*** germs has quit IRC10:05
*** germs has joined #openstack-nova10:05
*** andreas_s has quit IRC10:07
*** andreas_s has joined #openstack-nova10:07
*** Zames has joined #openstack-nova10:08
*** yamamoto has joined #openstack-nova10:09
*** germs has quit IRC10:10
*** namnh has quit IRC10:10
*** chyka has joined #openstack-nova10:13
*** yamamoto has quit IRC10:14
*** Zames has quit IRC10:18
*** chyka has quit IRC10:19
*** Zames has joined #openstack-nova10:23
*** Zames has quit IRC10:25
*** sree has quit IRC10:28
openstackgerritSilvan Kaiser proposed openstack/nova master: Exec systemd-run with privileges in Quobyte driver  https://review.openstack.org/55419510:29
*** mvk has quit IRC10:31
*** ralonsoh_ has joined #openstack-nova10:33
*** sidx64 has quit IRC10:35
*** sambetts|afk is now known as sambetts10:36
*** ralonsoh has quit IRC10:37
*** jpena is now known as jpena|brb10:38
openstackgerritPranab proposed openstack/os-vif master: Add native implementation OVSDB API  https://review.openstack.org/48222610:41
gibimelwitt: I left a comment and a question in https://etherpad.openstack.org/p/nova-runways-rocky with '[gibi]' prefix. But overall I'm OK with the proposal.10:41
gibimelwitt: I think we have to start doing it to gather real experience and then we can improve the process iteratively10:42
*** masber has quit IRC10:43
*** elmaciej has joined #openstack-nova10:46
*** mvk has joined #openstack-nova10:46
*** tbachman has quit IRC10:49
*** sidx64 has joined #openstack-nova10:52
*** tiendc has quit IRC10:53
*** phuongnh has quit IRC10:56
*** vladikr has quit IRC11:00
*** sidx64 has quit IRC11:02
*** rcernin has joined #openstack-nova11:02
*** suresh12 has joined #openstack-nova11:02
openstackgerritSurya Seetharaman proposed openstack/nova master: Add disabled field to CellMapping object  https://review.openstack.org/55009011:03
*** ralonsoh__ has joined #openstack-nova11:03
Kevin_Zhenggibi, Hi, I might need some suggestion on https://review.openstack.org/#/c/553288/ about tests11:05
*** josecastroleon has quit IRC11:06
*** rcernin has quit IRC11:06
*** ralonsoh_ has quit IRC11:07
*** suresh12 has quit IRC11:07
*** masber has joined #openstack-nova11:08
*** yamamoto has joined #openstack-nova11:10
*** masuberu has joined #openstack-nova11:12
*** yamamoto has quit IRC11:12
*** yamamoto has joined #openstack-nova11:12
*** masber has quit IRC11:14
*** annp has quit IRC11:15
*** pcaruana has quit IRC11:23
*** psachin has quit IRC11:28
*** psachin has joined #openstack-nova11:34
*** chyka has joined #openstack-nova11:35
openstackgerritSurya Seetharaman proposed openstack/nova master: Add disabled field to CellMapping object  https://review.openstack.org/55009011:37
*** jpena|brb is now known as jpena11:37
*** sidx64 has joined #openstack-nova11:38
*** chyka has quit IRC11:39
*** yamamoto has quit IRC11:40
*** yamamoto has joined #openstack-nova11:41
*** abhishekk has quit IRC11:44
openstackgerritSurya Seetharaman proposed openstack/nova master: Add CellMappingList.get_all_enabled() query method  https://review.openstack.org/55018811:46
*** vladikr has joined #openstack-nova11:47
*** pchavva has joined #openstack-nova11:52
*** sidx64 has quit IRC11:53
*** sidx64 has joined #openstack-nova11:54
*** pcaruana has joined #openstack-nova11:55
*** sidx64 has quit IRC11:57
*** sidx64 has joined #openstack-nova11:58
*** sidx64 has quit IRC11:59
*** amoralej is now known as amoralej|lunch12:04
*** liuzz has quit IRC12:04
*** sree has joined #openstack-nova12:04
*** liuzz has joined #openstack-nova12:04
*** sree_ has joined #openstack-nova12:05
*** sree_ is now known as Guest6124512:05
*** lucasagomes is now known as lucas-hungry12:06
*** germs has joined #openstack-nova12:06
*** germs has quit IRC12:06
*** germs has joined #openstack-nova12:06
*** sidx64 has joined #openstack-nova12:07
*** sree has quit IRC12:08
*** jaosorior has quit IRC12:09
*** josecastroleon has joined #openstack-nova12:09
*** germs has quit IRC12:10
*** odyssey4me has quit IRC12:11
*** odyssey4me has joined #openstack-nova12:11
*** sidx64 has quit IRC12:11
gibiKevin_Zheng: I shortly checked your test, functionally it looks OK. Do you feel that adding the request_id everywhere in the test is too much?12:14
*** artom_ has quit IRC12:15
sean-k-mooney[m]bauzas:  o/ i believe you reviewed this last cycle would you mind taking a look at the re-proposal of the nic feature based scheduling spec when you have time. https://review.openstack.org/#/c/545951/12:16
Kevin_Zhenggibi, sort of, but I'm not be afraid to add them, just don't know whether we have a better way to do that, in the latest patchset, I did some nodification for test test_create_server_error, but somehow the req id is different, I will dig into it latter12:16
*** READ10 has joined #openstack-nova12:16
Kevin_Zhenggibi for those tests that tested multiple actions in one test, it seems a little bit complicated, so I wonder we have a better way to do it.12:17
sean-k-mooney[m]bauzas:  i have 1 or two nits to adress in the corresponding code but should have the series rebased and uploaded by the end of the week. im hoping we can get this all merged before milestone 112:17
*** edmondsw has joined #openstack-nova12:17
*** sidx64 has joined #openstack-nova12:18
openstackgerritsahid proposed openstack/nova master: only increment disk address unit for scsi devices  https://review.openstack.org/53831012:19
*** trinaths has quit IRC12:20
*** sidx64 has quit IRC12:21
sean-k-mooney[m]dansmith: just saw your comment on https://review.openstack.org/#/c/449257/59 you last comment on this topic was before i took this over from rodolfo. ill try and adress this when i do the rebase later this week.12:22
openstackgerritJan Zerebecki proposed openstack/os-vif stable/ocata: Check if interface belongs to a Linux Bridge before removing  https://review.openstack.org/55452312:22
*** tbachman has joined #openstack-nova12:22
sean-k-mooney[m]dansmith:  looking at your old comment you would like us to add a spec_object field in addtion to the spec field correct?12:23
*** jaosorior has joined #openstack-nova12:23
gibiKevin_Zheng: thanks for describing your concerns, I have to think about a bit. I will reply in the review12:23
Kevin_Zhenggibi thanks alot12:23
openstackgerritJan Zerebecki proposed openstack/os-vif stable/ocata: Check if interface belongs to a Linux Bridge before removing  https://review.openstack.org/55452312:24
*** sidx64 has joined #openstack-nova12:27
efriedkaisers1: Howdy12:28
efriedkaisers1: /me US Central time :)12:28
openstackgerritsahid proposed openstack/nova master: libvirt: handle DiskNotFound during update_available_resource  https://review.openstack.org/55306712:29
*** _ix has quit IRC12:31
*** bkopilov has quit IRC12:32
*** openstackgerrit has quit IRC12:33
*** sidx64 has quit IRC12:34
*** openstackgerrit has joined #openstack-nova12:37
openstackgerritMerged openstack/nova master: Fix message for unexpected external event  https://review.openstack.org/55438012:37
*** Zames has joined #openstack-nova12:38
*** _ix has joined #openstack-nova12:38
*** sidx64 has joined #openstack-nova12:40
*** Zames has quit IRC12:40
jaypipesalex_xu_: still around? not sure what you were asking about preferred_traits... I have no issue with decorating traits in flavors as being preferred. I wouldn't send them to placement, though... just allow the scheduler weighers to use them in their sorting. is that what you were thinking of?12:40
*** r-daneel has quit IRC12:41
*** jpena is now known as jpena|lunch12:44
*** AlexeyAbashkin has quit IRC12:48
openstackgerritsahid proposed openstack/nova-specs master: virt: allow instances to be booted with trusted VFs  https://review.openstack.org/48552212:50
*** sidx64 has quit IRC12:52
jaypipesefried, cdent: so based on you guys' and tetsuro's feedback on the "nested providers allocation candidates" series, I'm wondering if there's *any* reason to use non-granular request groups when nested providers are present. I spent all day reworking that series yesterday to get things working so that a non-granular request group could work against a tree of providers (but not by summing inventories across the entire tree). Instead, what I did12:54
jaypipeswas say that individual providers needed within the tree needed to satisfy each quantitative resource request and then *collectively* the tree needed to satisfy the traits request. do you think that's wrong as well?12:54
jaypipeslemme push what I have.. one se.c12:55
efriedjaypipes: That sounds correct to me.12:55
efriedWithin the request, one resource_class:amount needs to be satisfied by one provider in the tree, or one associated via aggregate with any provider in the tree.  And collectively, the RPs satisfying the resource request (which may in fact *exclude* some of the RPs in the tree) must satisfy the traits.12:57
efriedNot sure if that last thing is what you did.12:57
efriedjaypipes: But I think that's important.  I don't think we want to say traits are satisfied by a provider that's not providing any resource to the request.12:58
efried...and we do need to make sure agg-associated RPs (but only ones providing resource) are included in the traits calculation.12:58
jaypipesefried: what about when a trait is applied to the NUMA node that is the parent of a provider satisfying some part of the resource request?12:58
jaypipesefried: forget the shared stuff for right now...12:59
efriedjaypipes: Hmmm, the NUMA thing...  I said something yesterday, I think in a spec comment, related to this.  I'll dig it up, but I think taking that into account, the answer to the above will be "that will depend on different syntax".13:00
efriedjaypipes: The gist was that (what I recall from PTG discussions, early Wednesday) we want to support the NUMA subtree business (and cdent this may also play into the vmware cluster thing) via a syntax that expresses: "Get all the resources in this group from a subtree whose (sub)root is marked with trait X"13:02
efriedjaypipes: So if the NUMA node is marked with trait I_AM_A_NUMA_NODE, the syntax would be like GET /a_c?resources4=...&subtree_trait4=I_AM_A_NUMA_NODE13:03
openstackgerritJim Rollenhagen proposed openstack/nova master: ironic: stop lying to the RT when ironic is down  https://review.openstack.org/54547913:03
efriedThat would affect the calculation of which providers in the tree are eligible to provide resources - i.e. just the subtree rooted at a provider marked I_AM_A_NUMA_NODE - and that trait gets special treatment such that the NUMA root RP itself doesn't actually need to provide resources.13:04
openstackgerritEd Leafe proposed openstack/nova master: Address issues raised in adding member_of to GET /a-c  https://review.openstack.org/55435713:04
*** jianghuaw has joined #openstack-nova13:05
*** abhishekk has joined #openstack-nova13:05
openstackgerritJay Pipes proposed openstack/nova master: tests for alloc candidates with nested and traits  https://review.openstack.org/53189913:05
openstackgerritJay Pipes proposed openstack/nova master: placement: resource requests for nested providers  https://review.openstack.org/55452913:05
jaypipesefried, cdent: please see above (different series, forked from the base of the old one)13:06
*** sidx64 has joined #openstack-nova13:06
gibiefried, jaypipes: I agree that a single resource request need to be satisfied from a single RP but in case of traits I think the trait needs to be satisfied by the RPs on the path from the root (or subroot) to the RPs that are providing resources to that request13:08
sean-k-mooney[m]efried:  thanks for your review on https://review.openstack.org/#/c/449257/59 you spotted my **{} sed hack which is fair i was just hopping i would not need to manually convert them all. ill fix it in the next respin. i answered some of your other questions inline13:08
gibiefried, jaypipes: where the path can be defined also by taking the RPs that providing resources and collecting all the ancestors for those RPs and then checking that the traits are satisfied in the set of RPs or not13:10
*** AlexeyAbashkin has joined #openstack-nova13:10
*** lucas-hungry is now known as lucasagomes13:10
*** ratailor has quit IRC13:11
mdboothCould somebody take a look at this live migration bugfix for me: https://review.openstack.org/#/c/551302/ . It's got a bunch of +1s and I hacked a CI run to ensure coverage.13:11
jaypipesgibi, efried: ok, understood. which gets back to my original question... if nested providers are present, is there really any point in *not* requiring granular request groups?13:12
gibijaypipes, efried: do I remember correctly that a granular numbered request group means that both the resources and the traits needs to be fulfilled from a single RP?13:13
*** Guest61245 has quit IRC13:14
efriedgibi: I don't think I agree that we should collect traits from "tree paths".  That seems excessively complicated.  What use case does it satisfy?  (Hint: I don't think it satisfies the NUMA thing without further semantic work)13:14
efriedgibi: Correct.13:14
*** sree has joined #openstack-nova13:14
efriedjaypipes: Trying to think through whether it's always possible to express an un-numbered request group as one or more numbered groups.13:15
efriedI think there's some cases you can't express - but it's actually a good thing that you can't.13:15
gibijaypipes, efried: I've started thinking about the same13:15
efriedLike getting the a trait from a provider you weren't expecting.13:15
gibijaypipes, efried: i.e. trait is on the compute RP, resource inventory is on the PF13:16
gibithat would need two separate numbered group13:16
efriedgibi: Yeah.  It *should* be the case that traits on the one aren't applicable to the other.  But who knows?13:16
*** fragatina has joined #openstack-nova13:17
openstackgerritsahid proposed openstack/nova-specs master: libvirt: add support for virtio-net rx/tx queue sizes  https://review.openstack.org/53960513:17
*** mriedem has joined #openstack-nova13:18
jaypipesefried, gibi: yeah, that's been my dilemma :)13:18
*** jianghuaw has quit IRC13:18
gibijaypipes: you successfully shared your pain :)13:18
efriedjaypipes: I'm not opposed to this idea in principle - it makes a couple of things simpler, which is good.13:18
jaypipesgibi: you're welcome. ;)13:18
efriedjaypipes: My concern is how we express this to operators.13:19
jaypipesefried: agree with you. I'm hunting for ideas.13:19
*** sree has quit IRC13:19
efriedjaypipes: They need to have, what, separate flavors for nested-modeled hosts than for non?13:19
efriedI mean, even a non-nested host you can express requests with granular.13:19
efriedSo the line could just be: start using granular for everything, period.13:20
jaypipesefried: meh, I don't think that will be common. I'm more concerned about how to document the quirks of each "solving algorithm", depending on whether they use granular or not, nested or not, sharing providers or not, etc13:20
efriedBut that's kind of a dick punch to the traits-in-glance thing.13:20
efriedjaypipes: Don't think what will be common?  Environments where some are trees and some are not?13:20
*** psachin has quit IRC13:21
*** jianghuaw has joined #openstack-nova13:21
jaypipesefried: no, I mean a need for flavors that request the same resources/traits but "in different ways" (i.e. collectively met vs individually met)13:21
openstackgerritsahid proposed openstack/nova-specs master: libvirt: add support for virtio-net rx/tx queue sizes  https://review.openstack.org/53960513:22
kaisers1efried: Hey, ok :)13:22
*** sidx64 has quit IRC13:22
efriedkaisers1: What you have looks fine to me, but you'll want to take your lead from mikal since he's engaged at this point.13:23
efriedkaisers1: I'm not sure if he wants you to put the non-systemd exec into the privsep lib, or something.13:23
cdentjaypipes, efried: I think we should make the simple cases as simple to express as possible and for some deployments that ought to mean that some hardware doesn't "turn on" numa, so it just reports simple inventory.13:23
kaisers1ok, i just wanted to ask regarding the kwargs topic, did you read my reply on that?13:23
efriedkaisers1: Not yet, looking...13:23
cdentI'd like to think that it is possible to do some nested things without granular but I've not thought it all the way through13:24
dansmithsean-k-mooney[m]: yep, that's what needs to happen13:24
cdents/it is/ought to be/13:24
gibiefried, jaypipes: granularity in the flavor needs to express what resource needs to be collocated and what can be spread inside the selected host. This should not depend on the fact that the host provide nested RP tree that allows the spreading13:24
*** _ix has quit IRC13:24
kaisers1efried: ok. It's just that either i don'r fully grasp it or it doesn't make that much sense to me13:24
kaisers1*don't13:24
efriedkaisers1: Oh, I totally didn't see that response.13:25
efriedI'll answer in the patch.  Sorry about that.13:25
kaisers1efried: np, thanks for looking into the patch13:25
efriedgibi: I tend to agree.13:25
gibiefried, jaypipes: so If I don't care about to collocate cpu and ram to the same numa then I can create the cpu request in a different group than the memory request and that will work against nested and not nested hosts13:25
openstackgerritSurya Seetharaman proposed openstack/nova master: Add CellMappingList.get_all_enabled() query method  https://review.openstack.org/55018813:26
gibiefried, jaypipes: or does it?13:26
jaypipesgibi: in a non-nested representation, there's no reason to use granular request groups. because every request is a granular request group (it cannot be satisfied by >1 provider non-sharing provider)13:26
efriedgibi: Yes.  (Though that still doesn't help us collocate e.g. CPU and VF on the same NUMA, cause they'll be members of different providers.)13:26
openstackgerritSurya Seetharaman proposed openstack/nova master: Add disabled field to CellMapping object  https://review.openstack.org/55009013:26
efriedgibi: Same numbered request group == same provider.  Different numbered request groups == maybe same, maybe different providers13:26
gibiefried: your last point made my above statment false13:27
efriedgibi: We definitely have no way (other than unique traits) to express that two requests MUST be from separate providers.13:27
*** sree has joined #openstack-nova13:27
gibiefried: sorry, mixed up13:27
gibiefried: let me try again13:28
efriedgibi: Always same tree-or-associated-sharing, though, if it's in one request.13:28
openstackgerritSurya Seetharaman proposed openstack/nova master: Add CellMappingList.get_all_enabled() query method  https://review.openstack.org/55018813:28
gibiefried: so we can instruct the deployer to create flavors always with granular groups if he does not care about numa affinity as two separate group can be satisfied by the same RP13:29
*** sidx64 has joined #openstack-nova13:30
gibijaypipes: I guess we don't need granular in non-nested deployments, but in a mixed deployment, the granular groups works against both type of hosts, which is good13:30
efriedgibi: Even if he cares about NUMA affinity (assuming we're just talking about NUMA_CORE+MEMORY_MB - stuff in the same RP).  You would then specify both resources in the same numbered group, meaning they have to come from the same provider.13:30
gibiefried: true13:30
*** hongbin has joined #openstack-nova13:30
efried(still doesn't help the NUMA_CORE+PF case)13:30
efried(we still need new syntax for that)13:31
efried(and the new syntax has to be within granular)13:31
*** burt has joined #openstack-nova13:31
gibiefried: NUMA_CORE + PF case needs a group that is satisfied by a subtree specified by something that identifies the root of the subtree somehow13:32
*** r-daneel has joined #openstack-nova13:32
gibiefried: your I_M_A_NUMA_NODE trait based subtree specification makes sense to me13:32
openstackgerritSurya Seetharaman proposed openstack/nova master: [WIP] Allow scheduling only to enabled cells (Filter Scheduler)  https://review.openstack.org/55052713:32
cdentjaypipes: if you're able to go back and respond to the various comments on the reviews, in context, that would be awesome. It's sometimes hard to keep the many thread clear.13:33
gibiefried: I can even forsee that I_AM_A_NETWORKING_RP_WHICH_DEFINES_A_SUBTREE_BELONGIG_TO_THE_SAME_NEUTRON_AGENT trait :)13:34
jaypipescdent: sure, I will. was just looking to brainstorm.13:34
cdentYeah, sure, not complaining, definitely want the brainstorming too13:35
gibiefried: i mean I foresee a possible need of such trait based selection13:35
cdentgibi: I think you missed _PLEASE on the end of that trait13:35
*** r-daneel has quit IRC13:36
jaypipesin any case, unfortunately, I now have a dentist appointment I need to get to... and then drop the dogs off at the spa and sit while my car gets serviced (will have my lappie for the last part, though, so will be online later)13:36
efriedenjay :)13:37
sean-k-mooney[m]jaypipes: doggy spa days. you really do spoil them :)13:37
*** sree has quit IRC13:39
*** esberglu has quit IRC13:39
*** voelzmo has joined #openstack-nova13:40
gibicdent: I think there is a limit about the lenght of a trait :)13:40
cdent:)13:40
*** voelzmo has quit IRC13:41
gibijaypipes: good luck with your errands13:41
* gibi needs a spa day13:41
*** esberglu has joined #openstack-nova13:42
*** jpena|lunch is now known as jpena13:42
sean-k-mooney[m]gibi i belive its set in the db scema  probaly  64,128 or256 charaters13:42
*** voelzmo has joined #openstack-nova13:42
*** amoralej|lunch is now known as amoralej13:43
sean-k-mooney[m]gibi:  its 255 https://github.com/openstack/nova/blob/master/nova/db/sqlalchemy/api_migrations/migrate_repo/versions/041_resource_provider_traits.py#L4313:43
*** Spaz-Home has quit IRC13:44
sean-k-mooney[m]also aprently we set the charset to latin1 so no unicode traits. is that intentional ?13:44
gibisean-k-mooney[m]: OK, then I can fit the PLEASE at the end :)13:44
gibisean-k-mooney[m]: I think that is intentional https://github.com/openstack/nova/blob/f80b4e50093002f84b43ff245a605fbe44d34711/nova/api/openstack/placement/handlers/trait.py#L7413:45
alex_xu_jaypipes: For example, I want to a FPGA device with Funciton A, the request is 'resources=RC_FPGA_DEV:1&preferred=FPGA_FUNCTION_A'. Do you mean you only request one RC_FPGA_DEV, then weigh the RPs base on the summary of RPs in the response of allocation_candidates?13:47
efriedkaisers1: Responded.  Note that I'm not saying any of those things will necessarily ever happen in this case.  It's just one of those things that's good practice.13:47
sean-k-mooney[m]ah of url parsing so we dont get sql injections... we still might get requests form people with non latin based alphabets at some point but i guess we can cross that bridge when we come to it13:47
efriedsean-k-mooney[m]: The schema only allows A-Z and _13:48
efriedand 0-913:48
sean-k-mooney[m]efried:  the api schema yes i see that13:48
alex_xu_oh, jaypipes has a dentist appointment13:48
*** tbachman has quit IRC13:49
gibicdent: there is a debate about a need of microversion bump in https://review.openstack.org/#/c/502306/17/specs/rocky/approved/bandwidth-resource-provider.rst@142 this might be interesting to you as well13:50
* cdent wishes he never got associated with microversions :)13:51
sean-k-mooney[m]efried:  the api schema yes i see that13:51
*** bkopilov has joined #openstack-nova13:51
*** links has quit IRC13:51
sean-k-mooney[m]ed ::13:52
gibicdent: you can free to forward the ping to other API experts :)13:52
gibis/can/are/13:52
*** SamYaple has quit IRC13:52
cdentgibi: hmmm. that is an interesting one. my general rule is capture well here: http://eavesdrop.openstack.org/meetings/nova_scheduler/2018/nova_scheduler.2018-03-19-14.00.log.html#l-17813:53
alex_xu_jaypipes: one more question, what is the reason you didn't want to add prefered trait in placement13:53
*** tbachman has joined #openstack-nova13:53
* gibi clicks13:54
sean-k-mooney[m]efried:  damit lost the message. any  TL:DR CUSTOM_SEÁNS_TRAIT would be invalid which is not much of a deal for me but some operators may want to create custom traits in there own language in teh future that said we should try to converge in a set of standard traits to cover there needs and cross the non english traits bridge when we really need too13:54
efriedalex_xu_: Because placement is a placer, not a weigher13:56
efriedalex_xu_: That's my interpretation of jaypipes' opinion, not necessarily my own13:57
sean-k-mooney[m]alex_xu_:  placement is fitting hard constratins, a prefereed trait is a best effort request. filtering based on a best effort request would potentially result in no valid host when there were hosts that met the required traits but no the prefered hence why jaypipes wanted to handel prefred in a weigher not in placement13:57
*** jmlowe has joined #openstack-nova13:58
sean-k-mooney[m] what efried  said :)13:58
gibicdent: thanks, does this also means that in the old microversion we have to keep the old behavior which means not to fail the request just ignore the requested resources?13:58
alex_xu_efried: thanks for the high level explain, sean-k-mooney[m] thanks for the detail explain :)13:58
efriedheh13:59
sean-k-mooneyinteresting. i have been testing out riot.im as an irc bouncer/bridge. looks like there is a 5-10 secodn delay when i use riot vs direct irc at least on posting. i get messages almost instanly on both13:59
*** lajoskatona has quit IRC13:59
*** gouthamr has joined #openstack-nova14:00
sean-k-mooneysean-k-mooney[m] is my riot.im client14:00
cdentgibi: strictly speaking, yes, but it depends on the severity of the bad behavior14:00
cdentI'm not generally a fan of the strict interpretation14:00
*** voelzmo has quit IRC14:01
*** jianghuaw has quit IRC14:01
gibicdent: in this particular case the nova boot request was accepted and the requested resources was ignored. After the proposed change such a request would be rejected with an error message14:02
alex_xu_sean-k-mooney: Does that mean I require TRAIT_A, and prefer TRAIT_B, there is chance to return nothing when there is no TRAIT_B for any RP? not quite understand14:02
cdentgibi: Again, using the strict interpretation, the older microversions should continue to accept14:02
*** tbachman has quit IRC14:03
*** eharney has joined #openstack-nova14:03
gibicdent: assume for a second that we don't make it strict and allow failing in the old microversion. This would mean that there would be no difference between the old and the new microversion14:04
*** voelzmo has joined #openstack-nova14:04
gibicdent: the only difference would be in the documentation14:04
gibicdent: about the meaning of the accpet14:04
edleafealex_xu_: "preferred" doesn't mean anything in placement, since it doesn't order anything14:04
gibicdent: in this case do we need the microversion bump just for the doc?14:05
edleafe"required" implies a filtering, which is what placement does14:05
alex_xu_edleafe: 'preferred' order the number of traits which the RP has14:05
cdentgibi: assuming that, then I wouldn't think a microverison was required for a doc change, because the behavior is the same everywhered, right? But if there is in fact a behavior change (in the API itself), that seems like a microversion, right?14:07
*** germs has joined #openstack-nova14:07
*** germs has quit IRC14:07
*** germs has joined #openstack-nova14:07
*** finucannot is now known as stephenfin14:07
alex_xu_or you guys refer to the implement problem14:08
*** voelzmo has quit IRC14:08
edleafealex_xu_: if I prefer a trait, I will accept an RP that doesn't have that trait if necessary. So Placement will return all RPs that satisfy the requirements. It is then up to the scheduler to sort them based on various things, and preferred traits would be one of those things14:09
gibicdent: I'm hesitant. The external behaviour will change as nova will reject a previously accepted the request. But even if nova today accept such request nova actually lies and nova does not fulfill the request properly as nova does not consider the QoS bandwidth policy on the port that is included in the request during the scheduling14:10
*** SamYaple has joined #openstack-nova14:11
gibicdent: so by rejecting such request nova we clean up a lie14:11
alex_xu_edleafe: oh, the weigher will change the order again14:11
cdentgibi: and you break working code14:11
gibicdent: true14:11
*** germs has quit IRC14:11
*** sidx64 has quit IRC14:11
cdentgibi: Which is what microversions are supposed to prevent, even though it means broken (in other ways) client code gets to continue existing14:11
cdentI agree that it is a tricky problem and your hesitancy is warranted14:12
*** awaugama has joined #openstack-nova14:12
cdentDo you feel the "lie" is a security problem?14:12
*** yamamoto_ has joined #openstack-nova14:12
*** yamamoto has quit IRC14:12
gibicdent: no, the lie is not a security issue. It is just a resource allocation issue14:13
edleafegibi: would you call the "lie" a bug?14:13
edleafeIOW, Nova isn't doing what it's supposed to?14:13
gibiedleafe: Nova missing support for including the port QoS policy in the placement decision14:14
* alex_xu_ is suddenly enlightened14:14
edleafegibi: understood. What I'm asking is if nova not doing that is a bug14:14
edleafeor is it just a feature that has not yet been implemented14:14
gibiedleafe: it is a missing feature, not a bug.14:15
gibiedleafe: it never worked before14:15
edleafegibi: ok, then it definitely needs a microversion14:15
gibiedleafe, cdent: thank you for the discussion. I will link this discussion to the spec14:16
mriedemgibi: "The external behaviour will change as nova will reject a previously accepted the request. But even if nova today accept such request nova actually lies and nova does not fulfill the request properly as nova does not consider the QoS bandwidth policy on the port that is included in the request during the scheduling" - same story with rebuilding a volume-backed instance with a new image; we used to accept that and just n14:16
mriedemonor it, but changed it to a 400 in the API in queens14:16
alex_xu_each microversion discussion is a war14:16
edleafealex_xu_: https://twitter.com/EdLeafe/status/97610035465079193614:17
mriedemnot really14:17
alex_xu_edleafe: :)14:17
gibimriedem: so there is precedence not to bump microversion in this case even if it breaks client code14:18
*** chyka has joined #openstack-nova14:18
gibiedleafe, alex_xu_: :)14:19
mriedemgibi: https://review.openstack.org/#/c/520660/14:20
mriedemlike most things, it's a case by case basis, but in the case of ^ we said it wasn't a microversion because it was a silent failure on the compute side14:20
mriedemyou shouldn't have to opt into being not broken14:20
mriedembut i'm sure with enough time and examples we can spin that all ways14:21
*** pcaruana has quit IRC14:21
*** mlavalle has joined #openstack-nova14:21
openstackgerritChris Dent proposed openstack/nova master: DNM: Demo code for microversion parse extraction  https://review.openstack.org/55026514:22
*** eharney_ has joined #openstack-nova14:23
edleafemriedem: yeah, that was why I was asking if it was a bug or a new feature. You shouldn't have to opt into bug fixes.14:23
*** kholkina has quit IRC14:23
*** chyka has quit IRC14:23
*** fragatina has quit IRC14:23
alex_xu_mriedem: https://review.openstack.org/#/c/520660/ is a bug, I feel it is different14:23
gibimriedem, cdent, edleafe: in the bfv rebuild case that rebuild has never worked before14:24
openstackgerritSilvan Kaiser proposed openstack/nova master: Exec systemd-run with privileges in Quobyte driver  https://review.openstack.org/55419514:24
alex_xu_gibi: the qos policy works with neutron currently, right?14:24
*** eharney has quit IRC14:24
gibialex_xu_: the qos policy without placement support cannot really work properly even if neutron implemented some data plane enforcement for that policy14:25
mriedemgibi: depends on what you mean by 'worked',14:26
*** sree has joined #openstack-nova14:26
mriedemthe api used to accept that kind of request and return a 202, and nothing would fail, but it was wrong14:26
mriedemwrong in that it didn't actually replace the root disk14:26
gibialex_xu_: imagine that you promised 10G minimum bandwidth for two VMs on the same host and on the same 10G PF. That will cause inconsistency when both VM start using that minimum bandwidth heavily14:26
mriedemtoday you can create a port with a qos policy and attach it to an instance, right? nova allows that w/o changing any resource consumption for that qos policy.14:26
gibimriedem: right14:27
*** felipemonteiro has joined #openstack-nova14:27
mriedemif we followed the volume-backed rebuild + new image path here,14:27
mriedemwe'd merge a patch to check if the port (or network) has a qos policy and fail if so14:27
gibiyes14:27
mriedemand then land a microversion that allows supporting those types of ports14:27
gibimriedem: wait a bit14:28
mriedemwhich is what https://review.openstack.org/#/c/532407/ is for14:28
*** pcaruana has joined #openstack-nova14:28
mriedemthinking about volume multiattach, we kind of have the same situation - you can't attach a multiattach volume with microversion < 2.60, you have to opt into using those types of volumes14:29
gibimriedem: we talking about adding microversion bump to the first patch where we start rejecting such requests not just microversion bump when we later start supporting such request14:29
*** felipemonteiro_ has joined #openstack-nova14:29
bauzassean-k-mooney[m]: looking at https://review.openstack.org/#/c/545951/2/specs/rocky/approved/enable-sriov-nic-features.rst14:29
mriedemgibi: if you do that, anyone that doesn't have that code in their cloud can have users still attaching those kinds of ports14:29
mriedemif we consider it a bug that we don't support ports with qos policies, then we should make that a hard failure (bug fix) and then add a new microversion which adds the support for those types of ports14:30
*** sree has quit IRC14:30
gibimriedem: so making it a hard failure is a bugfix, I like that14:31
alex_xu_emm...wait, we don't check the network has qos policy, we check the port has required resource or not. That is something new added in neutron14:31
mriedemgibi: well that's what we did for the bfv rebuild case14:31
*** voelzmo has joined #openstack-nova14:32
gibialex_xu_: required resources is just another representation of the QoS policy to let neutron do the transformation between QoS policy entity on the API to the resource classes and traits in the Placement14:32
mriedemtbh, we're inconsistent on this type of thing. with the multiattach support in queens, we check if the volume is multiattach and if the microversion is high enough (2.60) and if not, we fail. but i'm pretty sure in pike you could probably attach a multiattach volume to at least one instance at a time without failures.14:32
*** felipemonteiro has quit IRC14:32
alex_xu_gibi: the neutron side add 'resources' field to the port API by extension? or just add directly?14:33
gibicdent, edleafe, mriedem: I start to get convinced that making the API hardfail instead of accept and lie is a bugfix14:33
*** eharney_ is now known as eharnye14:34
*** eharnye is now known as eharney14:34
mriedemso the bug is today we are potentially over-subscribing the qos bandwidth right?14:35
mriedemfor a given network14:35
*** gouthamr has quit IRC14:35
gibialex_xu_: that will come from the QoS neutron plugin based on my current understanding14:35
gibialex_xu_: the neutron spec will describe this in detail (writing is in progress)14:35
*** Nil_ has joined #openstack-nova14:35
mriedemthe neutron qos stuff has been around for a long time hasn't it? kind of surprised no one hasn't already reported this as a bug.14:35
*** hemna_ has joined #openstack-nova14:36
gibimriedem: yes, SRIOV ports with QoS minimum bandwidth policies are not properly enforced and the minimum bandwidth cannot be garanteed14:36
gibimriedem: only SRIOV ports supporting minimum bandwidth policies14:36
*** voelzmo has quit IRC14:37
gibimriedem: and there was a release notes that stated the problem14:37
gibimriedem: https://github.com/openstack/neutron/blob/49d614895f44c44f9e1735210498facf1886c404/releasenotes/notes/qos-min-egress-bw-rule-b1c80f5675a4c1c3.yaml14:37
gibimriedem: so maybe the deployers read the release notes and understood that the support for the minimum bandwidth rule is incomplete14:38
mriedemadded in newton14:38
gibimriedem: yes14:38
mriedemlooks like the policy defaults allow only admins to create these types of policies14:39
mriedemso there is at least that14:39
*** sar has quit IRC14:40
openstackgerritSurya Seetharaman proposed openstack/nova master: Add disabled field to CellMapping object  https://review.openstack.org/55009014:42
openstackgerritSurya Seetharaman proposed openstack/nova master: Add CellMappingList.get_all_enabled() query method  https://review.openstack.org/55018814:42
openstackgerritSurya Seetharaman proposed openstack/nova master: Allow scheduling only to enabled cells (Filter Scheduler)  https://review.openstack.org/55052714:42
bauzassean-k-mooney[m]: jaypipes: stephenfin: soft -1 on https://review.openstack.org/#/c/545951/2 but I need to understand why we can't just get the port info in the conductor, and pass the traits to the scheduler14:42
* stephenfin looks14:43
bauzasI could be wrong but AFAIK we haven't said why it wasn't possible to do the above ^14:43
*** liverpooler has joined #openstack-nova14:44
*** _ix has joined #openstack-nova14:44
*** elmaciej has quit IRC14:45
*** artom_ has joined #openstack-nova14:46
*** voelzmo has joined #openstack-nova14:47
*** sree has joined #openstack-nova14:47
*** itlinux has joined #openstack-nova14:49
*** eharney_ has joined #openstack-nova14:50
*** gouthamr has joined #openstack-nova14:50
*** sree has quit IRC14:51
*** eharney has quit IRC14:51
*** r-daneel has joined #openstack-nova14:55
*** artom_ is now known as artom14:55
dansmithjaypipes: sean-k-mooney[m]: do either of you know if/when we can remove the old vif plugging stuff from the libvirt driver? i.e. everything reachable past the "if not os-vif" line?14:57
openstackgerritEd Leafe proposed openstack/nova master: Address issues raised in adding member_of to GET /a-c  https://review.openstack.org/55435714:57
edleafeefried: ^^ yay pep8!14:57
*** amodi has joined #openstack-nova15:00
*** dtantsur|afk is now known as dtantsur15:00
*** jistr is now known as jistr|mtg15:01
*** voelzmo has quit IRC15:02
*** voelzmo has joined #openstack-nova15:03
*** sree has joined #openstack-nova15:10
*** Eran_Kuris has quit IRC15:10
*** moshele has quit IRC15:12
*** mdnadeem has quit IRC15:13
mriedemdansmith: presumably once all of the legacy methods in libvirt/vif.py are converted to using os-vif objects15:14
dansmithmriedem: I don't know how to tell that. are those just linux bridge and ovs right now?15:15
mriedemlook in nova.network.os_vif_util at the _nova_to_osvif_vif_* methods that raise NotImplementedError15:16
*** yassine has quit IRC15:16
mriedemthere are quite a few15:16
stephenfindansmith: Yeah, once all of those have been converted15:16
*** eharney_ is now known as eharney15:16
stephenfinjaypipes and I have discussed it before. The IVS driver has been converted to an os-vif plugin but I don't know about the rest of them. We might need to jettison them or bring them into os-vif core15:17
dansmithis that actually happening in the background?15:17
stephenfinJust this one so far https://bugs.launchpad.net/bugs/170412915:17
openstackLaunchpad bug 1704129 in networking-bigswitch "Add an IVS os-vif plugin" [Undecided,New] - Assigned to Aditya Vaja (wolverine-av)15:17
dansmithcouldn't we remove the ones from our tree that are converted already?15:17
stephenfinThey should already be removed15:17
stephenfine.g. OVS and linuxbridge15:18
dansmiththere are several still that have a method and just pass15:18
dansmithare those converted and "removed" or are those things we just don't do anything for?15:18
*** abhishekk has quit IRC15:18
mriedemcan you give an example?15:19
stephenfinFor example?15:19
dansmithdef plug_802qbg(self, instance, vif):15:19
dansmith        pass15:19
stephenfinI went through and removed some others ones a while back. IIRC, everything else was still needed. See commit 1b872996d08b01a1b8a1e82d13d6d7b06bc3aa0115:19
mriedemok that's in nova.virt.libvirt.vif15:19
dansmithI was probably reading ivs and ovs and just assumed we had orphaned all this code, but if we're removing them as we can, that's cool15:19
dansmithmriedem: right15:19
mriedemi guess there is no plug for 802qbg but there is a get_config method for it15:20
stephenfinHmm, I'm not sure about that one. We do getattr magic somewhere in there so it's tricky to figure out what's called and what's dead15:20
mriedem2.1 is changed to 2115:21
mriedemso it becomes 802qbg15:21
mriedemthe os-vif conversion code looks for _nova_to_osvif_vif_802_1qbg15:21
mriedembecause the vif type is "802.1qbg"15:21
stephenfinWhich raised NotImplemented?15:23
stephenfin*raises15:23
mriedemyes15:23
*** sree has quit IRC15:23
mriedemi don't know why we don't have to plug/unplug those, but we have to get the config apparently for the guest15:23
mriedemum https://github.com/openstack/os-vif/blob/master/os_vif/objects/vif.py#L26315:24
mriedemhttps://github.com/openstack/os-vif/blob/master/os_vif/objects/vif.py#L27715:24
mriedemseems we could convert two right there15:24
dansmiththe reason I'm asking is around removing our linux_net module15:25
openstackgerritJackie Truong proposed openstack/nova master: Add trusted_certs to instance_extra  https://review.openstack.org/53789715:25
openstackgerritJackie Truong proposed openstack/nova master: Add trusted_certs object  https://review.openstack.org/48940815:25
openstackgerritJackie Truong proposed openstack/nova master: Implement certificate_utils  https://review.openstack.org/47994915:25
openstackgerritJackie Truong proposed openstack/nova master: Add trusted_image_certificates to REST API  https://review.openstack.org/48620415:25
dansmithit's really hard to tell if/when it's used right now15:25
stephenfinthis is for nova-network deletion, I assume?15:26
*** yamamoto_ has quit IRC15:27
dansmithnot really, although that's part of it.. the linux_net module uses a lot of config in conf/network that we can't remove when we nuke nova-net15:27
*** sree has joined #openstack-nova15:28
*** dave-mccowan has joined #openstack-nova15:29
dansmithmriedem: are you cool with backporting this? https://review.openstack.org/#/c/552691/15:31
mriedemsure15:33
*** jistr|mtg is now known as jistr15:36
openstackgerritMatt Riedemann proposed openstack/nova master: Don't log a warning for InstanceNotFound with deleted VIFs  https://review.openstack.org/55459115:37
*** sree has quit IRC15:37
*** trinaths has joined #openstack-nova15:39
*** suresh12 has joined #openstack-nova15:41
mriedemwelp, volume multiattach does'nt work with libvirt 4.0.0 in the queens UCA15:43
*** trinaths has quit IRC15:43
*** trinaths has joined #openstack-nova15:44
mnasermriedem: honest question is there multiattach tempeset jobs15:44
mriedemhells yes15:45
mriedemhttps://review.openstack.org/#/c/554317/15:45
*** trinaths has quit IRC15:45
mriedemlooks like the shareable flag isn't getting set in the disk config xml for some reason, should be relatively easy to figure out why15:45
*** fragatina has joined #openstack-nova15:45
mnasermriedem: so does that mean that a certain company is releasing things without possibly running tempest jobs15:45
mriedemi've just been getting distracted with the amount of bullshit warnings in the n-cpu logs15:45
*** sree has joined #openstack-nova15:46
mriedemmnaser: libvirt is open source yeah? but even still, i wouldn't expect them to test openstack stuff against their code.15:46
mriedemor qemu for that matter15:46
mnaserwell not libvirt but if uca ships libvirt 4.0.015:46
mnaseri'd expect that they tested all of them together15:46
mriedemi don't expect ubuntu / canonical to test all of the permutations of openstack features15:47
mriedemsince multiattach is optional too15:47
*** Spazmotic has joined #openstack-nova15:47
mriedemmaybe if it were part of the interop guidelines, but it's not15:47
*** felipemonteiro_ has quit IRC15:48
*** chyka has joined #openstack-nova15:49
*** felipemonteiro_ has joined #openstack-nova15:49
openstackgerritDan Smith proposed openstack/nova stable/queens: Add --by-service to discover_hosts  https://review.openstack.org/55460015:50
*** trinaths has joined #openstack-nova15:51
openstackgerritStephen Finucane proposed openstack/nova master: Add CPUWeigher  https://review.openstack.org/37952515:52
openstackgerritDan Smith proposed openstack/nova stable/queens: Add --by-service to discover_hosts  https://review.openstack.org/55460015:53
*** tbachman has joined #openstack-nova15:54
openstackgerritDan Smith proposed openstack/nova stable/pike: Add --by-service to discover_hosts  https://review.openstack.org/55460315:54
*** tbachman_ has joined #openstack-nova15:56
openstackgerritTyler Blakeslee proposed openstack/nova stable/queens: Add method repr() to NovaException  https://review.openstack.org/55460415:56
*** salv-orlando has quit IRC15:56
*** salv-orlando has joined #openstack-nova15:57
*** salv-orlando has quit IRC15:57
*** salv-orlando has joined #openstack-nova15:57
*** tbachman has quit IRC15:59
*** tbachman_ is now known as tbachman15:59
*** sree has quit IRC16:01
openstackgerritTyler Blakeslee proposed openstack/nova master: Add method repr() to NovaException  https://review.openstack.org/55460716:03
*** zhaochao has quit IRC16:05
*** felipemonteiro_ has quit IRC16:06
*** voelzmo has quit IRC16:17
*** masuberu has quit IRC16:18
*** ragiman has quit IRC16:20
*** lyan has joined #openstack-nova16:21
*** lyan is now known as Guest6476816:22
*** andreas_s has quit IRC16:22
*** gyee has joined #openstack-nova16:26
*** tbachman has quit IRC16:26
*** andreas_s has joined #openstack-nova16:26
*** yamamoto has joined #openstack-nova16:28
*** salv-orlando has quit IRC16:28
*** salv-orlando has joined #openstack-nova16:28
openstackgerritTyler Blakeslee proposed openstack/nova master: Add method repr() to NovaException  https://review.openstack.org/55460716:30
*** salv-orlando has quit IRC16:33
*** wolverineav has joined #openstack-nova16:33
*** yamamoto has quit IRC16:34
*** ameeda has joined #openstack-nova16:34
*** andreas_s has quit IRC16:36
ameedaHi, I try to create stack using rest api. current api version is V1 . I POST request with auth key , I got this error message {     "explanation": "The server could not comply with the request since it is either malformed or otherwise incorrect.",     "code": 400,     "error": {         "message": "The server could not comply with the request since it is either malformed or otherwise incorrect.",         "traceback": null,16:36
*** sree has joined #openstack-nova16:37
*** wolverineav has quit IRC16:38
*** wolverineav has joined #openstack-nova16:39
*** tbachman has joined #openstack-nova16:40
kashyapHmm, `git fetch origin` again failing for me, with 'origin' == https://git.openstack.org/openstack/nova.git16:41
* kashyap wonders if anyone else sees that16:42
*** rmart04 has quit IRC16:42
openstackgerritMatt Riedemann proposed openstack/nova master: Clarify log in RT._update_usage_from_migration  https://review.openstack.org/55462316:42
*** sree has quit IRC16:42
mriedemwe should log this more often i think: "Defaulting the value of the field 'projects' to None in FlavorPayload due to 'Cannot call _load_projects on orphaned Flavor object'"16:44
lyarwoodkashyap: https://git.openstack.org/openstack/nova without the .git as a remote WORKSFORME16:44
mriedem1K+ times isn't enough16:44
kashyaplyarwood: Err, let me try that16:45
lyarwoodkashyap: as listed here - https://git.openstack.org/cgit/openstack/nova16:45
kashyaplyarwood: No luck for me; very bizarre; yesterday it worked briefly, but not anymore16:47
dansmithefried: around?16:47
efrieddansmith: yeaux16:47
efriedAm I reading backscroll or...?16:47
dansmithefried: trying to write a functional test that talks to placement and do some aggregate stuff16:47
kashyaplyarwood: Both (.git or otherwise) worked for me in the past.  Not anymore.  /me digs further16:47
dansmithefried: and the reportclient I have is complaining that it doesn't know about a provider, which I think might be a local cache thing and I need to refresh its view or something?16:47
efrieddansmith: Which module are you writing your test in?16:47
dansmithbut the provider definitely exists16:47
efrieddansmith: Wanna post what you've got?16:47
dansmithefried: see if this is enough: https://pastebin.com/ekYNZJWR16:48
*** mvk has quit IRC16:49
dansmithI did a raw get against placement and the provider uuid I'm using is definitely there16:49
efrieddansmith: Yeah, set_aggregates_for_provider does require that you've loaded up the provider previously.  I think it says so in the docstring.  Stand by...16:49
*** moshele has joined #openstack-nova16:50
efriedoh, it says the provider must exist.  Which isn't the same at all.16:50
dansmithI don't think it does16:50
dansmithyeah16:50
*** gjayavelu has joined #openstack-nova16:50
dansmithpresumably I need a .refresh() or some such?16:50
efrieddansmith: Well, the PUT ought to be working.16:50
dansmitht'aint16:50
efriedcause that's definitely not relying on the cache.16:50
efriedyeah, 'tis.  It's the cache update that ain't.16:51
efriedprovider_tree.update_aggregates is what's complaining about the provider not being found.16:51
dansmithyou mean the aggregate association completed and it's just trying to refresh its view and failing?16:51
efriedCorrect.16:52
dansmithokay, what I meant is "this call is failing"16:52
dansmith;)16:52
efriedIf you put an _ensure_resource_provider before your set_aggs, it'll fix it.16:52
dansmithokay, but that's a bug yes?16:52
efriedThinking...16:52
melwittbauzas: runways draft, please add feedback if you have any, want to kick off using it later this week https://etherpad.openstack.org/p/nova-runways-rocky16:53
*** AlexeyAbashkin has quit IRC16:53
dansmithefried: I think the deal here is that this makes report client unable to be generally useful to manage things, as it seems opinionated about only knowing about one provider16:53
efrieddansmith: It's not unreasonable for you to expect to be able to use that method as you are.  So yeah, I think it's a bug.  But tbh I don't know what the right fix would be.16:53
dansmithwhere it used to work well16:53
openstackgerritMathieu Gagné proposed openstack/nova master: Fix rebuild of baremetal instance when vm_state is ERROR  https://review.openstack.org/52355916:54
dansmithefried: I shall stick a #FIXME(efried) in my test16:54
efrieddansmith: wfm.  Or even open a bug.16:54
dansmithwell, just this for the moment16:55
efriedI'd like to fix this on top of the upt stack, cause I want to take advantage of the move of the cache refresh timer business.16:55
*** felipemonteiro has joined #openstack-nova16:55
dansmithI can muck with internals in my test, but for other uses it'd be more of a wart16:55
*** tbachman has quit IRC16:56
efrieddansmith: I think the answer is going to be to trap such errors and invalidate the cache for that RP.  Though after the aforementioned "move of the cache refresh timer business", the cache is already invalidated in this case.  So... we'll just wind up ignoring that exception, pretty much.  But there's similar code elsewhere.  Quite a bit of it, I fear.16:57
dansmithefried: why wouldn't we just _ensure_resource_provider(uuid) in the set_aggregates call?16:58
efrieddansmith: That would *usually* be an extraneous REST call.  Which I don't like (but cdent does).16:58
sean-k-mooneydansmith: currently we cant remove all the old vif plugging stuff yet. but i think we could get there in rocky16:58
dansmithefried: ah I figured it wouldn't do so if it's already in our known tree16:59
dansmithsean-k-mooney: ack, I got caught up by the others, thanks16:59
*** masuberu has joined #openstack-nova16:59
efrieddansmith: That may be true, looking...16:59
efrieddansmith: Well, you're correct that it wouldn't retrieve it if it's already in the cache.  But it does call _refresh_associations.17:00
dansmithwhich we're about to do after the put?17:01
efriedno, where?17:01
sean-k-mooneydansmith: ok cool, we have been trying to remove the old methods as they get converted to use os-vif. eventully they will all go.17:01
efrieddansmith: I mean, if you were gonna do that via some other call after set_aggregates_for_provider, that's different.17:01
*** sahid has quit IRC17:03
*** suresh12 has quit IRC17:03
*** lucasagomes is now known as lucas-afk17:03
dansmithefried: we're refreshing associations after the put, which is failing now, right? or is that something else?17:03
dansmithefried: anyway, it's your call on how to fix it, but I think this needs to be a thing you can do17:03
efriedNo, that's just updating the cache17:03
efriedwhich is local and fast.17:03
*** mvk has joined #openstack-nova17:03
dansmithokay, from the result?17:04
efrieddansmith: Anyway, the reason I wouldn't always want to do that _ensure within this method is because it will go do a potential buttload of placement calls to populate the report client's cache.  But if ALL you cared about was updating agg associations, and you were never going to use any of that other stuff, all of that would be wasted.17:04
efriedyes17:04
dansmithwouldn't you want the refresh of the associations first anyway so the cache ends up consistent with the server?17:04
*** sridharg has quit IRC17:04
*** felipemonteiro_ has joined #openstack-nova17:04
*** suresh12 has joined #openstack-nova17:04
efriedWell, yes, actually, once I get around to it, I'll need to make this guy take generation into account.17:04
efriedAt which point you'll HAVE to have previously retrieved the RP, cause that's where you'll get the generation from.17:05
*** damien_r has quit IRC17:06
efriedBut I agree that a public method such as this one ought to be able to be called without you having to know you needed to populate the cache beforehand.17:06
efriedSo yeah, I'll need to change it (and its brethren) to _ensure/_refresh.17:06
efriedBummer.17:07
*** felipemonteiro has quit IRC17:08
dansmithack17:08
*** felipemonteiro has joined #openstack-nova17:08
melwitthas anyone seen this error before "EndpointNotFound: Could not find requested endpoint for any of the following interfaces: ['internal', 'public']" when we try to send a notification?17:10
gibimelwitt: I did see this before17:11
*** felipemonteiro_ has quit IRC17:11
*** bkero- has quit IRC17:11
mriedemmnaser: figured it out; it's only a problem for resizing an instance that has multiattach volumes attached; i don't know why the newer libvirt package versions have anything to do with it, because it looks like a latent bug in how we actually handle resize + multiattach17:12
*** ameeda has quit IRC17:12
melwittgibi: do you know if it's a problem on our end or is it a problem with a deployment?17:12
gibimelwitt: is this EndpointNotFound a keystone exception?17:13
melwittgibi: here's the full trace https://bugs.launchpad.net/nova/+bug/1753550/comments/317:13
openstackLaunchpad bug 1753550 in OpenStack Compute (nova) "Status does not update to "Shutoff" when instance shuts down itself" [Undecided,Incomplete]17:13
mriedemmelwitt: for the given endpoint, which interfaces are available for it in the service catalog17:13
mriedemthe new ksa code will try to use internal and public by default if you don't specify a specific interface in config, like 'admin'17:13
mriedemit's trying to hit glance17:14
melwittokay. they said they have services listed under 'public' when they do a 'openstack endpoint list'17:14
mriedemso presumably they have glance configured in nova on an admin interface but don't have it configured that way17:14
*** mdbooth has quit IRC17:14
melwittoh, hm17:14
*** suresh12 has quit IRC17:15
mriedemhttps://docs.openstack.org/nova/latest/configuration/config.html#glance17:15
mriedemsee 'valid_interfaces'17:15
melwittwoo, thanks17:15
*** _ix has quit IRC17:16
gibimriedem, melwitt: hm, this means that sending a notification causes a REST call to glance. interesting...17:16
melwittthe trace says 'internal' and 'public' though so it seems like they haven't changed that17:17
mriedemFile "/opt/stack/nova/nova/notifications/base.py", line 398, in info_from_instance17:17
mriedem    context)17:17
mriedem  File "/opt/stack/nova/nova/image/api.py", line 65, in generate_image_url17:17
melwittgibi: yeah, that surprised me17:17
*** fragatina has quit IRC17:17
gibimelwitt: me too :)17:17
mriedemthere was also something related to this which we fixed at the end of queens17:18
gibimriedem: this one? https://review.openstack.org/#/c/511397/14/nova/notifications/base.py17:18
mriedemhttps://github.com/openstack/nova/commit/62ef6cfcf01d84813f71d1e8252b86c170ee39f017:18
gibimriedem: OK yours is different and seems more relevant17:19
openstackgerritKashyap Chamarthy proposed openstack/nova master: libvirt: Allow to specify granular CPU feature flags  https://review.openstack.org/53438417:20
mriedemthey likely need to update nova, or restack their devstack env, and try again17:20
melwittthe commit id they pulled is from march 4 though, so they should have that fix17:22
openstackgerritMerged openstack/nova master: Remove old flavor_extra_specs_get db api method  https://review.openstack.org/53970617:23
openstackgerritStephen Finucane proposed openstack/nova master: conf: Correct documentation for '[pci] passthrough_whitelist'  https://review.openstack.org/55287417:24
mriedemildikov: a real hum dinger https://bugs.launchpad.net/nova/+bug/175719017:26
openstackLaunchpad bug 1757190 in OpenStack Compute (nova) "resize fails with volume multiattach using with libvirt 4.0.0 (and qemu 2.11.1): Failed to get shared "write" lock" [Medium,Triaged] - Assigned to Matt Riedemann (mriedem)17:26
mriedemwill fix after lunch17:26
mriedemfrom what i can tell, it has nothing to do with the actual libvirt/qemu versions, except maybe older libvirt/qemu were masking a bug in nova17:26
*** vivsoni__ has joined #openstack-nova17:26
mriedemduring resize we blow away our special 'multiattach' flag in the connection_info here https://github.com/openstack/nova/blob/f80b4e50093002f84b43ff245a605fbe44d34711/nova/virt/block_device.py#L63917:27
*** liverpooler has quit IRC17:27
mriedemderp17:27
ildikovmriedem: nasty... :/17:28
ildikovmriedem: I guess it gives us another reason to find a better way to pass the multiattach info17:28
*** salv-orlando has joined #openstack-nova17:29
*** lpetrut_ has quit IRC17:30
gibimelwitt, mriedem: we call glance API from the notification sending due to the else branch here https://github.com/openstack/nova/blob/24379f1822e3ae1d4f7c8398e60af6e52b386c32/nova/image/glance.py#L12017:30
*** priteau has quit IRC17:30
*** yamamoto has joined #openstack-nova17:30
gibimelwitt,mriedem: s/glance/keystone/17:30
gibimelwitt, mriedem: and that ksa code was introduced here https://review.openstack.org/#/c/488137/23/nova/image/glance.py17:31
melwittokay, yeah, and we're failing on L126 on endpoint = utils.get_endpoint(ksa_adap)17:33
gibimelwitt, mriedem: before that ^^ the notification url generation only depened on the config params but after that it can fall back to keystone17:33
*** masuberu has quit IRC17:33
openstackgerritSurya Seetharaman proposed openstack/nova master: Add CellMappingList.get_enabled_or_disabled(disabled=False) query method  https://review.openstack.org/55018817:33
openstackgerritSurya Seetharaman proposed openstack/nova master: Allow scheduling only to enabled cells (Filter Scheduler)  https://review.openstack.org/55052717:33
gibimelwitt: I think so yes17:34
*** salv-orlando has quit IRC17:34
melwittI meant that's what's in the trace. so that fails before we ever get to the code that will try to strip the version from the url17:34
*** ralonsoh__ has quit IRC17:35
gibimelwitt: this also means that you can workaround the problem by setting CONF.glance.api_servers in the nova.conf17:35
openstackgerritMerged openstack/nova master: Remove old flavor_access_remove db api method  https://review.openstack.org/53970917:35
*** yamamoto has quit IRC17:36
melwittyeah, true. on the surface, it seems like there must be something wrong with their service catalog, because this is a straight call to keystone, or so it seems17:37
openstackgerritMerged openstack/nova master: Remove old flavor_access_add db api methods  https://review.openstack.org/53971417:38
gibimelwitt: yes, hence my classification of this change as only a workaround and not a real solution17:38
gibimriedem, melwitt: I have to leave now, but I will read back tomorrow to see if we have to do someting about not talking to keyston from the notification sending codepath17:39
melwittthanks gibi17:39
*** baoli has joined #openstack-nova17:39
*** baoli has quit IRC17:40
*** moshele has quit IRC17:48
edmondswmelwitt there are a ton of unapproved specs... is there any focus on reviewing those and knocking down that queue?17:48
*** suresh12 has joined #openstack-nova17:48
edmondswI know the powervm spec has been sitting without comment since Feb17:48
*** ccamacho has quit IRC17:48
*** suresh12_ has joined #openstack-nova17:50
*** suresh12 has quit IRC17:50
*** trinaths has quit IRC17:51
*** derekh has quit IRC17:52
*** jackie-truong has joined #openstack-nova17:56
melwittedmondsw: I think several people have been reviewing specs. I've been busy with PTG summary writeups and getting the runways proposal going. spec freeze is Apr 19 so I'm going to send email this week to get input on a spec review day date where everyone will focus on spec reviews17:57
*** _ix has joined #openstack-nova17:58
edmondswmelwitt I know a bunch of spec reviews are going on. Just wondering if there is a concerted focus on burning down the queue. Sounds like you're thinking about that17:58
edmondswkinda a prereq for the runways stuff17:59
edmondswgotta get the spec approved first :)17:59
dansmithedleafe: around?18:00
melwittyeah. as usual, the beginning of the cycle people are reviewing specs more because as you said, have to approve things before we can focus on reviewing the implementations. I think people are already doing that. last cycle we had a 79% approved spec/bp completion percentage so I think this cycle we're probably going to approve fewer things to increase that percentage significantly18:01
melwittlike last cycle, we'll have a dedicated spec review day to burn through a lot of them before spec freeze. that's what I'm going to send email about this week so we can pick a date that works for most18:02
edleafedansmith: yeah - somewhat distracted atm18:02
*** wolverineav has quit IRC18:03
*** wolverineav has joined #openstack-nova18:03
dansmithedleafe: okay, I'm having trouble getting member_of to work and there's kindof a missing case in your tests, which I thought maybe was covering up a bug, but I tweaked it and it still passes, so .. I'm still digging18:03
*** moshele has joined #openstack-nova18:03
dansmithedleafe: but, I've associated a provider with an aggregate, but when I member_of it, I get back no candidates18:04
edleafedansmith: that's... odd18:04
edmondswmelwitt just a little anxious not seeing reviews on the powervm spec. I'd have thought it was a fairly easy review. I believe the nova team is already committed to this effort and it's just a matter of how much or how little we bite off per release, not whether. I'd like to get the rocky content locked down.18:04
edleafedansmith: the gabbi tests show that exact case18:05
edmondswwe've been putting up commits for review, but until the spec is approved hard to ask anyone to look at functional commits18:05
melwittedmondsw: has it been previously approved? I don't know the history about it18:05
edmondswmelwitt yes, and the history is all in the spec at the end18:05
dansmithedleafe: yeah, I know, that's why I'm trying to diff what I'm doing with those18:05
edmondswmelwitt it's a multirelease effort18:06
edmondswmelwitt since it was considered too much to bite off in one release from a reviewer bandwidth perspective18:06
dansmithedleafe: you don't have a case where you only have one thing in member_of that actually returns results, but I shaved one down and it still seems to work (although it's hard to know if it's returning the thing we expect18:06
edmondswmelwitt https://review.openstack.org/#/c/545111/18:06
*** salv-orlando has joined #openstack-nova18:07
artomSo, do we handle *any* NUMA stuff during live migration? I know hugepages and CPU pinning aren't, but what about plain old NUMA topology?18:08
openstackgerritMerged openstack/nova master: Remove old flavor_access_get_by_flavor_id db api method  https://review.openstack.org/53972018:09
melwittedmondsw: okay, probably just needs some reminding then. looks like mriedem has reviewed it last month so maybe he can take another look at it soon18:09
edleafedansmith: you could inspect the returned a-c, but that's always too complicated to write in gabbi18:09
dansmithedleafe: yeah, I'm doing that in my real test and getting back [], but I dunno how to do much with the gabbit18:09
cdentedleafe: untrue, either make the test verbose: True or make it fail intentionally in the response_json_paths18:09
openstackgerritMerged openstack/nova master: Remove old flavor_destroy db api method  https://review.openstack.org/53972918:09
*** mgoddard has quit IRC18:09
dansmithbut it's checking the length of allocation_requests18:10
cdentand the error message will show the full response18:10
openstackgerritMerged openstack/nova master: Remove old flavor_get_by_flavor_id db api method  https://review.openstack.org/53973318:10
openstackgerritMerged openstack/nova master: Remove old flavor_get_by_name db api method  https://review.openstack.org/54437918:10
dansmithcdent: ah, duh, trying18:10
openstackgerritMerged openstack/nova master: Remove old flavor_get db api method  https://review.openstack.org/54462118:10
openstackgerritMerged openstack/nova master: Remove old flavor_get_all db api method  https://review.openstack.org/54468818:10
openstackgerritMerged openstack/nova master: Remove old flavor_create db api method  https://review.openstack.org/54470318:11
openstackgerritMerged openstack/python-novaclient master: Fix local test fails with pypy  https://review.openstack.org/55342618:11
dansmithcdent: I tried just setting the asserted length to something wrong, but it doesn't show me anything else other than 17 != 118:11
cdentdansmith: yeah, I meant something more breaking that that:18:12
cdentresponse_json_paths:\n  $: foo18:12
cdentwill try to compre the whole json object to foo18:12
* dansmith uses larger hammer18:12
*** sambetts is now known as sambetts|afk18:13
*** lpetrut has joined #openstack-nova18:13
*** AlexeyAbashkin has joined #openstack-nova18:14
vivsoni__In case of 'nova create' - NO cinder API is called18:14
vivsoni__In case of 'nova delete' - NO cinder API is called18:14
*** AlexeyAbashkin has quit IRC18:14
vivsoni__Hi Team, please correct if my understanding is wrong18:15
mriedemif you're booting from volume then of course cinder is called18:15
mriedemif you've attached volumes before you delete the instance, of course cinder is called18:15
*** tbachman has joined #openstack-nova18:15
vivsoni__mriedem: ok18:15
*** AlexeyAbashkin has joined #openstack-nova18:15
*** dtantsur is now known as dtantsur|afk18:16
edleafecdent: /me learns new gabbi trick18:16
dansmithcdent: edleafe: yeah, well, that clearly shows it's working as expected18:16
dansmithso I be stumped18:16
dansmithif I don't pass member_of, I get back the things I expect18:17
vivsoni__mriedem: so if my instance is attached to volume, then if i shutoff the nova instance and restart the instance... then cinder api of detach and attach is called is it ?18:17
dansmithinterestingly, the request logging from placement is urlencoded from inside my functional test, but not when the gabbit runs18:17
dansmithso I wonder if it's getting more than the uuid18:18
mriedemvivsoni__: no, if you're just stopping and starting the instance, nova doesn't detach the volume18:18
*** jpena is now known as jpena|away18:18
*** suresh12_ has quit IRC18:18
edleafedansmith: are you sending a single uuid string, or a 1-element list?18:19
*** suresh12 has joined #openstack-nova18:19
dansmithedleafe: I've tried both, initially just one, then tried in:$uuid18:19
vivsoni__mriedem: ok... i wanted to understand more on 'nova create/delete/live-migration' w.r.t cinder api call.. do you have some reference link, if yes, please share18:20
* edleafe is grasping at straws18:20
mriedemvivsoni__: not really18:21
mriedemvivsoni__: hemna has some diagrams of those flows though i think18:21
mriedemnot sure if they are published18:21
vivsoni__ok18:22
dansmithedleafe: yeah I know18:22
dansmithedleafe: I printed member_of from inside the normalize function and it is properly [$uuid]18:22
*** harlowja has joined #openstack-nova18:23
cdentdansmith, edleafe: any chance that this is the result of some base database condition that is different between the dan tests and the ed tests?18:23
dansmithcdent: such as what?18:23
dansmithcdent: I can get /resource_providers/$rp/aggregates and I get back $uuid18:24
dansmithso I feel like placement is working and storing the association I made18:24
dansmithohhhhhhh18:25
dansmithI may be completely stupid18:25
dansmithmaybe18:25
* edleafe bites tongue18:25
dansmithedleafe: I deserve whatever was coming before you bit your tongue for this one18:27
*** vivsoni__ has quit IRC18:28
dansmithgood news, member_of seems to work.. EOM18:29
efriedWell now you gotta tell us what you were doing wrong.18:29
efriedNot just so we can make fun of you - although that too.18:29
efriedBut mainly so we can avoid similar pratfalls ourselves.18:29
efriedAnd therefore avoid being made fun of in turn.18:29
edleafeuh, yeah - for *education*18:30
*** gyee has quit IRC18:30
dansmithI had a method that would create and add hosts to aggregates, both in nova and placement18:31
dansmithand the nova call I was making is additive, where the placement one is declarative18:31
dansmithso I was putting multiple hosts in the nova aggregates, but only in the latest placement one18:31
dansmithand the last placement one I created was "no-hosts"18:31
dansmithso, you know..18:31
dansmithand the order in which I was printing debug stuff was consistent, but.. not the full story18:32
*** dave-mccowan has quit IRC18:32
*** yamamoto has joined #openstack-nova18:32
edleafewish I could say I18:33
mriedemhttps://www.youtube.com/watch?v=AMQ8E3mTgY018:34
edleafeI've never done something similar18:34
dansmithmriedem: yeah I know, but thanks for putting into visual form18:34
mriedemha18:34
mriedemask smcginnis or jungleboyj, i was bashing my head for about 2 hours one morning over a unit test i was writing and couldn't figure out why it didn't work18:35
mriedemcan't remember what it was, but it was really dumb18:35
*** fragatina has joined #openstack-nova18:35
edleafecdent: ok, given the output using your trick: http://paste.openstack.org/show/706443/, how would I reference the resource provider uuid within the response_json_paths in order to compare it to the cn1uuid?18:36
*** yamamoto has quit IRC18:37
*** tssurya has quit IRC18:37
cdentedleafe: reading18:38
smcginnismriedem: What what?18:38
mriedemsmcginnis: last time i was at claddagh and we were upstairs, i was trying to get a gd unit test working most of the morning18:40
mriedemmaybe i was the only one that heard my swearing18:40
cdentedleafe: you can to check that the cn1uuid is in there, which in there?18:41
smcginnisOh yeah, I remember now. :)18:41
cdents/can/want/?18:41
edleafecdent: dunno, just thought that checking that it's in there wasn't definitive enough. Getting allocation_requests[0].keys()[0] or something like that18:42
melwittI'm able to repro the EndpointNotFound problem in an old devstack I have, and what I get from ksa_adapter.get_endpoint() in both the non-list and interface list cases is: "*** EmptyCatalog: The service catalog is empty."18:42
melwitteven though I have glance in 'openstack endpoint list'18:43
cdentedleafe: if you know the uuid in advance, it's generally easier to check for some key down its path. you can do environ and response expansions within the left hand side of a json path thing18:44
*** AlexeyAbashkin has quit IRC18:44
efriedmelwitt: Can you show me your openstack endpoint list for glance?18:45
efriedor 'show' would be better18:45
*** andreas_s has joined #openstack-nova18:45
*** dave-mccowan has joined #openstack-nova18:46
* edleafe tries to translate what cdent just typed18:46
efriedmelwitt: The ksa_adapter is going to be narrowed down based on (possibly defaulted) config options.  So as Matt was saying earlier, if for example your catalog has the admin endpoint, but your conf has (possibly by defaulting) internal and public, you'll get that EmptyCatalog/EndpointNotFound when you ask the adapter for an endpoint.18:46
cdentedleafe, sorry, I'm doing the usual too many things at once and then writing two thoughts down in one sentence18:47
melwittefried: okay, sec18:47
cdentedleafe: first example after the heading: https://gabbi.readthedocs.io/en/latest/jsonpath.html#substitution18:47
cdent(well, only example)18:47
edleafecdent: So given the example paste, what would I put in the left side? IOW, what is "nested.structure"?18:49
melwittefried: http://paste.openstack.org/show/706465/18:50
*** andreas_s has quit IRC18:50
openstackgerritEric Berglund proposed openstack/nova master: PowerVM Driver: vSCSI volume driver  https://review.openstack.org/52609418:50
efriedmelwitt: And the glance section of the conf?18:50
melwittno [glance] section, all default18:51
efriedmelwitt: ...and the version document from the endpoint?  (curl http://127.0.0.1/image)18:51
melwitthttp://paste.openstack.org/show/706468/18:51
cdentedleafe: $.allocation_requests[0].allocations.["$ENVIRON['CN1']"].resources.DISK_GB: 10018:52
cdentI think jay did this somewhere, will find example18:52
cdentedleafe:  $.allocation_requests..allocations["$ENVIRON['SS_UUID']"].resources[DISK_GB]: [100, 100] in the allocation-candidates.yaml18:53
efriedmelwitt: Yup, that ought to work.18:53
efriedmelwitt: You said you were using and old stack?  What version of ksa, and what commit of nova?18:54
melwittstepping through the code in pdb, when I get to keystoneauth1/access/service_catalog.py(362)endpoint_data_for() the self._catalog is an empty list [], even for the 'public' interface18:55
melwittthat's where it raises exceptions.EmptyCatalog('The service catalog is empty.')18:55
melwittsec18:55
edleafecdent: ok, thanks - that helps.18:55
melwittkeystoneauth1==3.4.018:56
melwittcommit a5a569d6670c29f995b1e8a2a2013471d57469d7 of nova18:56
melwittit raises exceptions.EmptyCatalog('The service catalog is empty.') for each of 'internal' and 'public'18:57
openstackgerritDan Smith proposed openstack/nova master: Make get_allocation_candidates() honor aggregate restrictions  https://review.openstack.org/54799018:57
openstackgerritDan Smith proposed openstack/nova master: Add require_tenant_aggregate request filter  https://review.openstack.org/54500218:57
openstackgerritDan Smith proposed openstack/nova master: WIP: Honor availability_zone hint via placement  https://review.openstack.org/54628218:57
dansmithedleafe: your stuff in action: https://review.openstack.org/#/c/545002/12/nova/tests/functional/test_aggregates.py18:57
dansmithL224 specifically18:58
efriedmelwitt: Is that ocata?18:58
melwittno?18:59
efriedsorry, trying the wrong way to find where that is chronologically :)19:00
melwittmaster from Feb 12, the commit I linked is dated Feb 1219:00
melwittit might have probably merged later19:00
efriedmelwitt: Is this code path using a RequestContext?19:00
melwittbut it's from around that time19:00
melwittyes, it's a RequestContext19:01
efriedWhat's in the RequestContext.service_catalog?19:01
melwittit's empty ... though I think that's my fault, I created an admin context to call glance.api_servers19:03
*** dave-mccowan has quit IRC19:03
*** AlexeyAbashkin has joined #openstack-nova19:03
efriedWell, unless I'm mistaken, the get_endpoint() stuff winds up in whatever context you're using.  So that would splain why it's empty in your pdb.19:05
melwittis the RequestContext.service_catalog supposed to get populated as a result of the get_endpoint() call? or is it supposed to be pre-populated before the get_endpoint call?19:06
*** amoralej is now known as amoralej|off19:06
efriedThe latter19:06
*** avolkov has quit IRC19:06
melwittokay, yeah, then I've messed up this attempt to repro the problem19:07
efriedget_endpoint gets its information *from* the context.19:07
*** maciejjozefczyk has quit IRC19:07
openstackgerritChris Dent proposed openstack/nova master: WIP: Parse placement forbidden traits query string  https://review.openstack.org/55466519:08
openstackgerritGiridhar Jayavelu proposed openstack/nova-specs master: VMware: place instances on resource pool  https://review.openstack.org/54906719:08
openstackgerritJim Rollenhagen proposed openstack/nova master: ironic: stop lying to the RT when ironic is down  https://review.openstack.org/54547919:13
jrollwhee, that should be good now19:13
openstackgerritMatt Riedemann proposed openstack/nova master: Use Queens UCA for nova-multiattach job  https://review.openstack.org/55431719:13
openstackgerritMatt Riedemann proposed openstack/nova master: Preserve multiattach flag when refreshing connection_info  https://review.openstack.org/55466719:13
*** rmart04 has joined #openstack-nova19:15
openstackgerritMatt Riedemann proposed openstack/nova master: Preserve multiattach flag when refreshing connection_info  https://review.openstack.org/55466719:16
openstackgerritMatt Riedemann proposed openstack/nova master: Use Queens UCA for nova-multiattach job  https://review.openstack.org/55431719:16
*** rmart04 has quit IRC19:20
*** suresh12 has quit IRC19:21
melwittokay, repro'd again by actually creating an instance, doing a virsh shutdown, then waiting for nova-compute to stop the instance, same error and again RequestContext.service_catalog = [] (I logged it from nova/image/glance.py)19:24
melwittso somewhere along the way, the context we pass isn't one with a properly populated service_catalog19:25
*** psachin has joined #openstack-nova19:28
*** _ix has quit IRC19:28
*** awaugama has quit IRC19:29
melwittuh oh, I think I know why. this path is being run through a periodic task which has been given an anonymous get_admin_context(), and when we eventually try to do something with a service_catalog in the RequestContext, there isn't going to be one19:31
efriedmelwitt: That'd do it.  Lemme dig up the patch that did the context-y stuff...19:32
efriedmelwitt: https://review.openstack.org/#/c/490057/19:32
melwittI'm not sure what the answer is here, other than gibi looking into whether we can avoid relying on service catalog stuff for sending notifications19:32
efriedmelwitt: That patch ought to help us isolate which context is giving us grief here.19:33
mriedemnova doesn't have admin creds to glance like we do for cinder and neutron, so you can't rely on that in a periodic either19:33
melwittbecause the context we use in periodic tasks is going to be a mostly empty admin one intended to read the database, etc. it's not going to have service catalog info in it19:33
*** suresh12 has joined #openstack-nova19:33
melwittoh, okay. so we are ok for cinder and neutron then19:33
*** _ix has joined #openstack-nova19:33
mriedemif properly configured19:33
mriedemyou have to configure nova to talk to neutron with an admin role token for port binding,19:34
efriedMm, yeah, it's coming back to me.  I don't remember where else we've seen this, but the answer was: if you want this to work, you have to supply creds in the conf so that we can build a proper admin context.19:34
mriedemthe cinder <> admin config was added in queens for a related bug with periodic tasks19:34
mriedemdoing things with volumes19:34
*** yamamoto has joined #openstack-nova19:34
efriedYeah, what mriedem said.19:34
melwittgotcha, okay19:34
efriedAnd... is that okay?19:34
efriedor do we need to "fix" it?19:34
mriedemfor notifications, we really shouldn't have to hit a REST API every time we send a notification, because that's kind of crazy19:35
openstackgerritsean mooney proposed openstack/nova master: add mtu to libvirt xml for ethernet and bridge types  https://review.openstack.org/55307219:35
efriedYeah, I'll agree with that.  Probably the first thing to look into, then - why we need to talk to glance to send a notification.19:35
mriedemi've mentioned this before,19:35
melwittyeah, gibi said he's going to investigate that tomorrrow19:35
mriedembut whenever we construct a glance "client" object in-tree, it goes thorugh the 'get endpoint url' stuff19:36
mriedemnotifications goes through info_from_instance to build a payload,19:36
mriedemwhich gets an image ref URL19:36
efriedmriedem: Yeah, I remember you mentioning it was doing it like 4000 times in a devstack run, or something.19:36
*** dave-mccowan has joined #openstack-nova19:36
mriedemefried: yup19:36
efrieds/devstack/tempest/19:36
mriedemwe f'ing love to hit glance19:36
*** d34dh0r53 has quit IRC19:37
mriedemif [glance]/api_servers is set, we just build a static string based on that19:37
sean-k-mooney[m]mriedem:  so the api call to glance when sending the notificaiton is from creating the glance client object?19:37
*** d34dh0r53 has joined #openstack-nova19:37
mriedembut if it's not set, we go through the ksa magik19:37
efriedWhich *should* be using cached values, I thought.19:37
efriedi.e. we're not actually hitting the API 4000 times.19:37
mriedemmaybe it is19:37
mriedemthat would be nice to know19:38
efriedmriedem: Yeah, we build the adapter every time, but the session & auth are cached.19:38
sean-k-mooney[m]efried:  well if the call to the api is coming from createing the client object perhaps we should be storing that between requests so that we dont have to keep creating it19:39
*** yamamoto has quit IRC19:39
openstackgerritJackie Truong proposed openstack/nova master: Add trusted_certs to instance_extra  https://review.openstack.org/53789719:39
openstackgerritJackie Truong proposed openstack/nova master: Add trusted_certs object  https://review.openstack.org/48940819:39
openstackgerritJackie Truong proposed openstack/nova master: Implement certificate_utils  https://review.openstack.org/47994919:39
openstackgerritJackie Truong proposed openstack/nova master: Add trusted_image_certificates to REST API  https://review.openstack.org/48620419:39
mriedemsean-k-mooney[m]: efried is saying we're not hitting the api every time19:40
mriedemjust the first time to get the service catalog entry19:40
mriedemand it's the identity api in this case19:40
sean-k-mooneymriedem: yes but we are still creating the client 4000 and hitting the cache 3999 times right?19:40
mriedemalternatively, don't put an image ref url in the notification payload, just the image uuid19:40
mriedemsean-k-mooney[m]: yeah19:40
mriedemthe notification payload attempts to mimic the GET /servers/detail API which returns the image id and bookmark link to the image19:41
sean-k-mooneyso if we create teh client once and resuit we definetlly dont hit the api and we get rid of 3999 calls to the client constructor?19:41
mriedemusing the same thing we're hitting here19:41
*** tssurya has joined #openstack-nova19:42
mriedemsure, but that's lower priority atm19:42
mriedemthe thing now is how to fix this periodic19:42
efriedIt'd be interesting to see if I can cache the adapter and stuff will still work.19:42
sean-k-mooneymriedem: image uuid might be better in general then the url19:43
mriedemsean-k-mooney: at least for the notification yeah i agree19:43
mriedemwe can change that with versioned notifications, but this also gets shoved into the legacy notifications and changing those is like breaking an api19:43
sean-k-mooneyglance can have multiple image urls correct? is there any guarntee the consumer of the url can reach that backend19:43
mriedemwell, it's over rpc so if you've configured nova to hit internal glance api endpoints, your notification consumer probably can too19:44
*** psachin has quit IRC19:44
mriedemif that consumer actually needs to get the image details, idk19:44
mriedemor why the consumer can't just take the image id and form it's own image api request, ...19:45
sean-k-mooneymriedem: sure but with the uuid they can query for which ever one they actully need so that is more generally useful i think19:45
mriedemit's not like we return links to volumes and ports in the notification either19:45
mriedemsean-k-mooney: i agree19:45
melwittyeah, the only other thing I can think of for fixing the periodic task is if we could somehow seed periodic tasks with the service catalog instead of the empty get_admin_context() one, but I don't know what would be involved there19:45
melwittI guess it anyway won't be able to get any info from glance even if it had the catalog, it would just fail later, right?19:46
mriedemthe notification path here isn't actually doing a GET to glance19:46
efriedYeah, I'm not worried about that bit at all.19:46
mriedemit's getting the service catalog from keystone via ksa19:46
melwittmeaning, it won't do that even if it has the catalog?19:46
mriedembut w/o a token19:46
melwittokay19:47
mriedemso for now, the easiest thing to do probably do is in this notification code, handle EndpointNotFound and just set the image_ref_url to the image id19:47
efriedare context hashable?19:47
mriedemthey are serializable19:47
dansmithedleafe: hmm, so member_of with multiple aggregates is doing an "or" of all the ones you pass, right? is that what we really want?19:47
openstackgerritMerged openstack/nova master: [libvirt] Add _get_XXXpin_cpuset()  https://review.openstack.org/52763119:47
efriedso probably19:47
mriedemefried: i'd rather not do crazy cache shit with contexts19:48
openstackgerritMerged openstack/nova master: api-ref: add a note in DELETE /os-services about deleting computes  https://review.openstack.org/55359819:48
*** psachin has joined #openstack-nova19:48
mriedemwe already have some craziness with contexts and periodics today pulling off the local thread storage19:48
mriedemwhich totally effs with request id log tracing19:48
efriedmriedem: I was just going to experiment, cool yer jets.19:48
mriedemi will not cool my jets or hold my horses19:48
melwittslow your roll19:48
mriedemthis is the part that blows up https://github.com/openstack/nova/blob/master/nova/notifications/base.py#L39719:48
* efried leafs back through the logs for assless chaps references19:48
openstackgerritChris Dent proposed openstack/nova master: Use nova.db.api directly  https://review.openstack.org/54326219:48
sean-k-mooneymriedem: is instance.image_ref the image id?19:49
melwittgdi I tried to set the importance at the same time as mriedem and set it differently. will set it back19:51
*** AlexeyAbashkin has quit IRC19:51
*** liverpooler has joined #openstack-nova19:51
*** salv-orlando has quit IRC19:51
*** salv-orlando has joined #openstack-nova19:52
openstackgerritEric Young proposed openstack/nova master: Support extending attached ScaleIO volumes  https://review.openstack.org/55467919:53
mriedemsean-k-mooney: yes19:54
melwittmriedem, efried: are one of you going to propose the patch or shall I?19:54
mriedemi can push a patch19:54
melwittk19:54
mriedemif LP would not timeout on me19:54
*** jackie-truong has quit IRC19:54
openstackgerritMichael Still proposed openstack/nova master: Move configurable mkfs to privsep.  https://review.openstack.org/55192119:55
openstackgerritMichael Still proposed openstack/nova master: Move xenapi xenstore_read's to privsep.  https://review.openstack.org/55224119:55
openstackgerritMichael Still proposed openstack/nova master: Move xenapi disk resizing to privsep.  https://review.openstack.org/55224219:55
openstackgerritMichael Still proposed openstack/nova master: Sync xenapi and libvirt on what flags to pass e2fsck.  https://review.openstack.org/55407819:55
openstackgerritMichael Still proposed openstack/nova master: Move xenapi partition copies to privsep.  https://review.openstack.org/55360519:55
openstackgerritMichael Still proposed openstack/nova master: Move image conversion to privsep.  https://review.openstack.org/55443719:55
openstackgerritMichael Still proposed openstack/nova master: We no longer need rootwrap.  https://review.openstack.org/55443819:55
openstackgerritMichael Still proposed openstack/nova master: We don't need utils.trycmd any more.  https://review.openstack.org/55443919:55
*** tesseract has quit IRC19:56
*** salv-orlando has quit IRC19:56
sean-k-mooneymriedem: one other tought. you said if you get an endpoint not found you would set image_ref_url to the image id. any reason to not always use the image id?19:56
mriedemsean-k-mooney: i said above, it changes the api19:57
mriedemin this case, meh19:57
mriedemwe can change the payload for the versioned notifications later,19:57
mriedembut this also goes in the legacy notifications19:57
jaypipesguh, this day turned into a giant disaster.19:57
sean-k-mooneymriedem: yes but we could have a microverion for that no? oh this is an unversioned notification19:57
mriedemjaypipes: i have a photo that might make your day better19:57
efriedjaypipes: Sokay, I still haven't gotten through those two patches yet.19:57
jaypipesefried: no worries, duder. today's pretty much a goner for me.19:57
sean-k-mooneyjaypipes: how did the dentist go.19:58
jaypipesefried: suffice to say it involves me cleaning up a giant pile of dog shit in the back of my car in the rain.19:58
jaypipesefried: ^ not related to the dentist19:58
mriedembut at least your teeth are clean19:58
efriedjaypipes: Geez, you try to do something nice for the dog...19:58
jaypipesindeed.19:58
mriedemunless there was...splatter19:58
sean-k-mooneymriedem: ew, let hop not for jaypipes sake19:59
efriedjaypipes: You want a quick feeling of satisfaction, https://review.openstack.org/#/c/545111/ ought to be eligible for quick-approve :)19:59
sean-k-mooney*hope19:59
mriedemefried: i'll take a look again at that after this19:59
jaypipesmy dogs have become a nested provider tree with inventory of DOG_SHIT_KG19:59
efriedrofl20:00
mriedemCUSTOM_DOG_SHIT_KG?20:00
tssuryalol20:00
jaypipesyes, sorry.20:00
jaypipeswell corrected, mriedem20:00
mriedem:)20:00
sean-k-mooneymriedem: i dont know me might want to standarise it20:00
jaypipessean-k-mooney: os-dog-poop?20:01
jaypipesthough that's a big specific. os-poop would be more generic and future proof.20:01
efriedWell, we already have os-brick20:01
jaypipeswell played efried20:01
sean-k-mooneyjaypipes: you know im surprised you have not come up with an os-pug yet20:01
*** sar has joined #openstack-nova20:01
*** masber has joined #openstack-nova20:02
jaypipessean-k-mooney: Provider Usage Group.20:03
* jaypipes gets to work...20:03
*** dave-mccowan has quit IRC20:03
*** artom has quit IRC20:03
openstackgerritMatt Riedemann proposed openstack/nova master: Don't log a warning for InstanceNotFound with deleted VIFs  https://review.openstack.org/55459120:03
*** dave-mccowan has joined #openstack-nova20:04
dansmithjaypipes: loaded question for you in here: https://review.openstack.org/#/c/544694/2/specs/rocky/approved/alloc-candidates-member-of.rst20:07
mriedemefried: edmondsw: question inline about evacuate https://review.openstack.org/#/c/545111/20:07
*** psachin has quit IRC20:08
*** psachin has joined #openstack-nova20:09
*** germs has joined #openstack-nova20:10
*** germs has quit IRC20:10
*** germs has joined #openstack-nova20:10
edmondswmriedem looking20:10
*** suresh12 has quit IRC20:10
*** suresh12 has joined #openstack-nova20:10
mriedemalso,20:10
mriedemat some point, we should not have powervm specs for 'implement random feature parity stuff in our driver',20:10
mriedemjust do what other virt drivers do and have specless feature parity blueprints per feature in questoin20:11
sean-k-mooneymriedem: for https://review.openstack.org/554591 are the network-vif-deleted event generated by neutron port being deleted as part of instance deleteion or are they neutron vif unplugged events form the ports being removed from ovs?20:11
*** suresh12 has quit IRC20:11
efriedmriedem: That would be lovely.  Are we at that point yet?20:11
*** suresh12 has joined #openstack-nova20:11
mriedemefried: i'm ok with that if melwitt is20:11
mriedemsean-k-mooney: the former20:12
melwittmriedem, efried: sounds fine to me20:12
efriedCool beans.  edmondsw esberglu ^.20:12
sean-k-mooneymriedem: ok then ya that makes sense to me. i just was not sure where the event was comming from20:12
efriedmriedem, melwitt: But for this release, since we're already here, use the bp/spec that's proposed, yah?20:13
mriedemshrug, i personally don't care for the wishlist20:13
mriedembecause it just seems really random20:13
melwittis the proposed spec all just feature parity stuff?20:13
*** germs has quit IRC20:13
mriedemand i see adding support for hot plugging vifs was added after PS320:14
efriedmelwitt: yes20:14
edmondswI'm happy to abandon this and start using specless blueprints this release20:14
efried++20:14
mriedemthere are no actual design details in the spec, so it's not really useful imo20:14
sean-k-mooneymriedem: so just to follow on from that, does that mean if i delete a neutron port that is bound to an instance that automatically results in a notifcation to nova to call detach interface on the virt driver?20:14
*** psachin has quit IRC20:14
mriedemso specless bp is fine20:14
edmondswyep20:14
melwittsounds reasonable to me20:14
mriedemsean-k-mooney: yup20:14
efriedSweet.  Swat that paperwork down!20:15
sean-k-mooneymriedem: hum ok is that documented anywhere. it kind of makes sense i just would not have expected neutron to allow you to delete teh port in that case20:15
*** awaugama has joined #openstack-nova20:15
edmondswmriedem to the evacuate question... I'm honestly not sure if there's anything more to it than a) supporting moving to another host in general and b) updating the support matrix20:16
jaypipesdansmith: will try to answer it first thing in the morning... been a day. having a beer.20:16
dansmithjaypipes: oh you said you were getting to work, so I figured you were around20:16
dansmithmah bad, tomorrow is fine20:16
melwittgetting to work cleaning dog poodoo20:17
mriedemedmondsw: someone on your dev or qa team could actually test it first20:18
mriedemto flush out any obvious problems20:18
edmondswmriedem oh definitely20:18
edmondswwas just talking from a code perspective20:18
mriedemsean-k-mooney: good question; i didn't trace the requests in the tempest job to see if the port was actually detached before it was deleted, or what the order was there,20:18
mriedemsean-k-mooney: should be pretty easy to find out - create a server from a pre-existing port and try to delete the port20:19
mriedemi don't have a devstack handy20:19
mriedemi know you can't delete an attached volume20:19
sean-k-mooneymriedem: ill give it a try i think i have an env running20:19
*** tidwellr has joined #openstack-nova20:21
*** tidwellr has left #openstack-nova20:22
openstackgerritEric Berglund proposed openstack/nova master: DNM: EXPERIMENTAL: Set proc_units_factor to 0.1 in VMBuilder init  https://review.openstack.org/55468820:23
sean-k-mooneymriedem: so ya i can create the port, boot with it then delete it and the interface gets detatched from the running vm20:23
sean-k-mooneymriedem: im just going to check the xml to confirm20:23
sean-k-mooneymriedem: yep the vm no has no nics20:24
sean-k-mooneyi wonder if the revers works. can i create a neutron port and then set its owner to a vm and trigger an attache20:25
cfriesenSo in the discussion around https://review.openstack.org/#/c/552924/ something interesting has come up...currently vcpus and 4KB pages can be consumed on a compute-node bases (where it floats across the whole compute node) or on a numa-node basis (where it's constrained to a single numa node) depending on whether the instance has a numa topology or not.   (https://github.com/openstack/nova/blob/master/nova/virt/libvirt/driver.20:27
cfriesenpy#L4435)   This makes it hard to reason about resource tracking.20:27
*** priteau has joined #openstack-nova20:29
mriedemsean-k-mooney: no the reverse doesn't work,20:30
mriedemnot might get a network-vif-plugged event but it won't know what to do with it and just log a warning20:31
mriedem*nova20:31
*** esberglu has quit IRC20:31
*** dave-mccowan has quit IRC20:31
sean-k-mooneymriedem: ya just tried it.20:31
mriedemthe fact you can delete a port while it's attached is pretty scary20:31
sean-k-mooneyi create another port and did openstack port set --device 55a08a3b-fd27-495f-85da-39eedd5fde9c --device-owner "compute:nova" --host "ubuntu" myport20:31
*** eharney has quit IRC20:31
sean-k-mooneyit ended up binding the port on the host correctly20:32
sean-k-mooneyim going to do a hard reboot but i think it will then update the xml correctly20:32
mriedemin the compute api we check instance states when detaching a port,20:32
edleafedansmith: sorry, was away. Yes, the idea for member_of was "member of at least one of these". The typical use case would be only 1 agg.20:32
mriedembut the networking api wouldn't have that20:32
mriedemso you could detach ports by deleting them while an instance is migrating20:32
*** masber has quit IRC20:32
mriedemwhich seems, not good20:33
*** yassine has joined #openstack-nova20:33
*** openstackgerrit has quit IRC20:33
cfriesendoes it mess up our accounting?20:34
*** esberglu has joined #openstack-nova20:34
cfriesenwhy is "while it's migrating" different than "while it's running"?20:34
*** yassine has quit IRC20:35
sean-k-mooneymriedem: strange the hard reboot did not cause the interface to show up either. i must be missing something that the attach/detach whould have done20:35
*** yamamoto has joined #openstack-nova20:35
dansmithedleafe: yeah, so that's kinda what I was thinking at first, but then I tried to apply it to two things at once, so:20:36
dansmithedleafe: imagine I have a tenant-restricted aggregate and the user boots with a requested AZ20:36
dansmithedleafe: if I just concatenate the tenant and az list, then I won't restrict them properly20:37
efriedcfriesen: Are you suggesting that somewhere in the stack there will be code responsible for translating the allocation_request for [PCPU:2 from NUMA_RP_1] to [PCPU:2 from NUMA_RP_1 + VCPU:2 from compute node RP] ??20:37
mriedemcfriesen: it was an example of a bad time to detach a port20:37
cfriesenefried: I'm suggesting that our logic right now is not rigorous. :)20:37
dansmithedleafe: I can restrict the tenant aggs to only the one that matches the az, but they might not be the same size (and thus are different)20:37
sean-k-mooneyright we could have already caluated the new xml and miss the port is being removed20:38
mriedemwe do refresh the nw info cache when we get network-vif-deleted events20:38
dansmithedleafe: I think placement _could_ know whether one is a subset of the other, but I can't really20:38
efrieddansmith, edleafe: Do we need member_of=all:... ?20:38
edleafedansmith: yeah, that really isn't how placement aggs should work20:38
dansmithedleafe: so, just trying to figure out what to do there.. maybe we can have a hangout with jaypipes (et al) tomorrow after I have stewed on it a bit and make sure we're happy with what we've got20:38
cfriesenefried: but other than that, I think we had been talking about tracking PCPU entirely separately from VCPU  (where PCPU is the number of "dedicated" pcpus we have, and VCPU is the number of VCPUs we can support)20:39
edleafedansmith: we could add an 'and' case by leaving off the 'in:' operator20:39
efriedI guess if pressed, I would have said we were implementing in: for parity with GET /rps, but wouldn't actually be using it.20:39
dansmithefried: well, I thought of that, but I'm not sure that's right either20:39
cfriesenefried: PCPUs are always consumed on a per-numa-node basis20:39
cfriesenefried: VCPUs can be consumed either per-numa-node or per-compute node, which is a problem for resource tracking.20:39
efriedcfriesen: Just so.  Which is why I was of the opinion that we should only be presenting that inventory from within the NUMA node RP.20:40
dansmithefried: if you have two groups of computes and both are legit for a tenant, then you don't want all20:40
efrieddansmith: Yeah, I wasn't really up on the use cases for GET /a_c?member_of20:40
dansmithefried: almost want member_of=one:$agg1,$agg2&member_of=$agg320:41
cfriesenefried: but that doesn't line up with the current logic, where you can consume a VCPU (and 4KB pages) from the entire compute node20:41
sean-k-mooneymriedem: fyi http://paste.openstack.org/show/706642/ any guess why the "attach" by recreating the port and hard rebooting did not work? the port look identical to me bar the fact the status is down on the new port20:41
efrieddansmith: But I knew we would want in:[any] for parity.20:41
*** yamamoto has quit IRC20:41
dansmithefried: edleafe: I'll try to come up with a few different hard examples to talk about and we can decide whether we care (or care right now) about them20:41
efriedYeah, sounds good.  I mean, the bp is approved and completed, but heck, let's throw another one after it.20:42
edleafedansmith: ok. If it's needed, it shouldn't be too big of a change20:42
cfriesenefried: and for various reasons it's really hard to track 4KB pages, since the host can consume them from either compute node at will.  (Unless you run host stuff in a separate cgroup or container to provide hard limits.)20:42
dansmithedleafe: okay20:42
efriedcfriesen: s/either compute node/either NUMA node/ ?20:42
mriedemsean-k-mooney: not really20:42
cfriesenefried: whoops,yes20:43
cfriesenefried: we try to account for memory fairly tightly on our compute nodes, and had to make 4KB pages use "preferred" rather than "strict" numa mempolicy because we were hitting the oom killer20:43
efriedcfriesen: You could have your memory expressed in terms of resource class MEMORY_4K_PAGE.20:44
efriedcfriesen: But I guess that's not the issue you're stuck on.20:44
efriedcfriesen: You're concerned about being allowed to get those pages from separate NUMA nodes in cases where a) NUMA affinity is not strictly required; and b) you can't get 'em all from one.20:44
efriedcfriesen: So what you can actually do here is express your resource requests as separate numbered groups.20:45
cfriesenefried: sort of.  currently if you have an instance numa_topology nova will strictly constrain your instance to a single host numa node.20:45
efriedcfriesen: It *sort of* requires knowing how many NUMA nodes are possible on a host.  What's the max for that, anyway?  2?  4?20:46
cfriesenefried: and if you don't have an instance numa_topology then nova will let you float over the whole compute node20:46
efriedcfriesen: I'm talking about in the idyllic future where we're doing all this with NRP and granular resource requests.20:46
cfriesenefried: common hardware is typically 2, some is 4.  exotic hardware has many more.20:46
efriedOkay, so if you care about strict affinity, you always ask for all your memory in one numbered request group.  If you don't - if you want to allow spread - you ask for blocks in separate numbered request groups.20:47
cfriesenefried: first we need to figure out how we want to handle host NUMA affinity for instances with no numa_topology.20:47
cfriesenefried: because if we continue to allow it to consume VCPUs and 4KB pages from the whole compute node, then we can't track those per numa node20:48
efriedBear with me, then20:48
efriedcfriesen: So let's say you want 4096 4K pages and 4 VCPUs.20:49
efriedcfriesen: If you want strict affinity, you say resources1=VCPU:4,MEMORY_4K_PAGE:409620:50
*** tssurya has quit IRC20:50
efriedcfriesen: But if you don't care about strict affinity...20:50
efriedlet's say you know your cloud doesn't have any hosts that support more than 4 NUMA nodes.20:50
*** edmondsw has quit IRC20:51
*** tssurya has joined #openstack-nova20:51
*** felipemonteiro_ has joined #openstack-nova20:51
efriedYour flavor that doesn't need strict affinity could say resources1=MEMORY_4K_PAGE:1024&resources2=MEMORY_4K_PAGE:1024&resources3=MEMORY_4K_PAGE:1024&resources4=MEMORY_4K_PAGE:1024&resources5=VCPU:1&resources6=VCPU:1&resources7=VCPU:1&resources8=VCPU:120:52
efriedNow, the results you get back can still include permutations where all of those resources come from the same RP.20:52
efriedAnd if you're using a weigher, you can choose those first.20:52
*** salv-orlando has joined #openstack-nova20:52
efriedBut it'll also allow for permutations where the resources come from different RPs.20:52
cfriesenefried: the problem arises when we actually start up the qemu process.  The memory is actually consumed on whatever numa node happens to request it...and now nova/placement no longer knows how much memory is available on each host numa node.20:52
sean-k-mooneyefried: there is nothing in the api preventing you form creating a vm with 4 virtual numa nodes on a host with on 2 phyical numanodes20:52
efriedcfriesen: That would be a libvirt fix.20:53
efriedsean-k-mooney: Cool, that's useful.20:53
sean-k-mooneyefried: the libvirt dirver does not allow that but the only guarentee that the api provides in this case is that if i request a vm with 1 numa node it will not span 2+ host numa nodes20:53
cfriesenefried: it's not a libvirt problem, it's how linux works.  If you haven't overridden the numa affinity, then by default memory allocations occur on the numa node that you're on.20:54
sean-k-mooneycfriesen: not quite20:54
efriedcfriesen: Then I don't understand how we ever wind up with memory coming from disparate NUMA nodes.20:54
sean-k-mooneythe linux kernel prefers to but numactl makes the desision and it may allocated form a remote numa node in some cases20:54
*** felipemonteiro has quit IRC20:54
*** tssurya has quit IRC20:55
sean-k-mooneyefried: a vm without hugepages can have its memory provided by any numa node20:55
*** tssurya has joined #openstack-nova20:55
cfriesenefried: if there is no instance numa topology, then nova doesn't specify any affinity.  this means the qemu threads are free to float across the whole compute node.  when they do a memory allocation they will by default be allocated memory from the numa node they're currently running on.20:55
efriedBut it's all provided by the same NUMA node?20:55
cfriesenefried: no20:55
sean-k-mooneyefried: infact today we do not have a facility to enforce that it comes form the same numa node in nova20:55
cfriesenwe explicitly let the host decide....we did for a while restrict it to a single numa node, then removed that for increased density20:56
cfriesen(we being nova)20:56
efriedAnd just so I'm clear, you want that kind of VM to be able to run on the same node as the strictly-affinitized one?20:56
efrieds/node/host/20:56
sean-k-mooneycfriesen: did that ever land upstream. i dont think we ever did a release with that behavior for 4k pages20:57
sean-k-mooneyefried: yes i dont see why not20:57
*** salv-orlando has quit IRC20:57
*** kaisers has joined #openstack-nova20:57
efriedWell, y'all seem to be splainin why not.20:57
sean-k-mooneyefried: the only way to request numa affined memory in openstack today is via hugepages20:57
cfriesenefried: same compute node, yes.   there's a spec under review right now to support shared and dedicated vcpus on the same compute node, and we already support 4KB and 2MB page backing on the same compute node20:57
cfriesenit's just that right now the resource tracking is kind of messed up for 4KB pages20:58
*** kaisers has quit IRC20:58
sean-k-mooneycfriesen: yes 4k pages are not tracked in the numa topology blob so we cant numa afine them20:59
sean-k-mooneycfriesen: you can request 4k pages specifically but the is special case code the skips the numa suff for them20:59
cfriesensean-k-mooney: aren't they part of mempages?20:59
*** openstackgerrit has joined #openstack-nova21:00
openstackgerritMerged openstack/nova master: Always pass 'NUMACell.siblings' to _pack_instance_onto_cores'  https://review.openstack.org/53736421:00
*** dave-mccowan has joined #openstack-nova21:00
sean-k-mooneymempages?21:00
cfriesenNUMACell.mempages21:00
sean-k-mooneyyou can set hw:mem_page_size=4k but i dont think they are stored in NUMACell.mempages21:01
sean-k-mooneyi guess i can check the db one seck21:01
*** pchavva has quit IRC21:02
sean-k-mooneythat is what the numatopology blob looks like http://paste.openstack.org/show/706682/21:04
sean-k-mooneycfriesen: so yes "nova_object.data": {"used": 0, "total": 2043576, "reserved": 0, "size_kb": 4} they are there21:05
sean-k-mooneycfriesen: they are actully tracked per numa node but we cant tie teh per numa node values back to the host memory_mb value simply today21:06
*** suresh12 has quit IRC21:06
*** suresh12 has joined #openstack-nova21:06
cfriesenI think we can, for ones that are strictly pinned. you just subtract the same value from both21:06
cfriesenbut for floating ones we don't know the actual per-host-numa-node consumption21:07
sean-k-mooneyi mean technically host memory_mb is the sum of all the cells 4k pages but im not sure if they will always agreee21:07
*** jmlowe has quit IRC21:08
sean-k-mooneycfriesen: the reall issue is the host reserved memory option. that is host wide and we jsut subtract it form the memory_mb value in the code but not sure how to translate that to per numa reserved values21:08
mriedemefried: melwitt: i reckon an upgrade release note will be in order for this endpoint not found thing with the legacy notification payload image_ref_url being an image id rather than a url21:09
sean-k-mooneycfriesen: that said we really need to deprecate it and replace it with a per numa version at some point21:09
cfriesensean-k-mooney: agreed21:09
efriedmriedem: Yeah, sounds like a plan.  Though I expect it to be a while before folks quit using api_servers IRL.21:09
sean-k-mooneyanyway its 9 so im going to go home and have dinner o/21:10
cfriesenlater21:10
mriedemefried: same21:10
efried...which (I think) makes the issue moot.21:10
efriedThis bug came out of a devstack?21:10
efriedCause that's the one place I know for sure we got rid of api_servers.21:10
mriedemyes it did21:10
mriedem"This on devstack with commit id: 5d2add74534719c5670b29152964a60e8f23b42b"21:10
melwittyeah, release note is always helpful. I do wonder if the field is nullable and if sending nothing would be better than breaking the contract, but meh, not sure21:11
efriedBut this is good - flushing out these bugs/corner cases in devstack before they hit the proverbial fan in production.21:11
*** jmlowe has joined #openstack-nova21:11
mriedemmelwitt: nullable is also breaking the contract a bit21:11
mriedemfwiw, this is already dumb if you're using bfv,21:12
mriedembecause instance.image_ref is '' for bfv21:12
mriedemso we're sending an image ref url with no image id in it21:12
melwittheh, okay21:12
cfriesensean-k-mooney: prior to 1231c469d (circa 2014) we did pin instances without a numa_topology to a single host numa node21:12
melwittefried: yeah, we need to check when this broke so we know how far back to backport. and also if it's in ocata then people are probably hitting this in production21:13
efriedmelwitt: https://review.openstack.org/#/c/490057/21:14
melwittit's maybe not "as noticeable" since it's likely only this sync power states periodic that results in notifications maybe21:14
melwittefried: okay, cool.21:14
melwittso only need to backport to queens21:15
efriedmelwitt: And here's the change that removed api_servers from devstack: https://review.openstack.org/#/c/490031/21:15
*** suresh12 has quit IRC21:15
efriedmelwitt: Well, I'll be a little surprised if that's the only place we're using a tokenless auth context to look up a glance endpoint.21:16
efriedgiven the scope of 49005721:16
*** AlexeyAbashkin has joined #openstack-nova21:19
melwittefried: yeah ... probably. but since it's so recent, I think you're right this is being caught a lot sooner than most people upgrade to queens. I was worried it was going to go back to ocata or something crazy. I know you linked the patch earlier but I didn't notice it was pretty recent21:19
*** suresh12 has joined #openstack-nova21:21
melwittmriedem: what I was thinking with the nullable thing, would be in the worst case if someone had automation parsing image_ref_url and we make it not a url, but empty string isn't one either. not sure if they're equally bad. just thinking out loud21:22
mriedemif someone blindly takes the image_ref_url and makes a GET curl request with it, it's going to blow up for all volume-backed instances21:22
openstackgerritMatt Riedemann proposed openstack/nova master: Handle EndpointNotFound when building image_ref_url in notifications  https://review.openstack.org/55470321:23
mriedemi think in the long ago, you could also create servers by passing an image URL to the imageRef parameter21:23
melwittyeah. I was thinking before that as far as the format of it. probably overthinking it21:23
mriedemand nova would parse it21:23
*** AlexeyAbashkin has quit IRC21:23
mriedemmultiattach is fixed with the queens UCA21:29
mriedemhttps://review.openstack.org/#/c/554667/21:29
mriedemthat's the only thing blocking us from using the queens UCA in devstack now21:29
imacdonnooh21:30
imacdonnis there any actually need to use the UCA? ZFSSA CI seems to be doing OK with out .. but that's cinder21:31
mriedemwell, a few things,21:31
*** itlinux has quit IRC21:31
mriedemwe get newer libvirt and qemu,21:31
mriedemtesting rocky against pike UCA seems weird, we should at least use queens if it's there21:31
mriedemand it allows us to remove these weird workarounds in devstack for the multiattach job to *not* use the UCA b/c the pike UCA didn't have the right package versions, but queens UCA does https://review.openstack.org/#/c/554317/21:32
mriedemhttps://review.openstack.org/#/c/554314/21:32
imacdonnyeah, I have one of those workarounds (per your recommendation)21:32
*** moshele has quit IRC21:32
imacdonnI suppose it does make sense to test nova stuff with the latest virt stuff21:32
mriedems/nova/openstack/21:33
imacdonnseems it'd matter more for nova... but yeah21:33
mriedemsure, for libvirt and qemu yes, but you also get newer things like rados and tgtd21:33
mriedemovs21:33
mriedemetc21:33
imacdonnOK. Has devstack master already been updated to use the Queens UCA?21:35
mriedemhttps://review.openstack.org/#/c/554314/21:35
mriedemdepends on fixing this bug in nova first21:35
*** esberglu has quit IRC21:35
imacdonnk... I'll subscribe to the bug21:36
mriedemget stvnoyes on the oracle phone21:36
imacdonnnot sure if he's still in EU, but even if not, he's probably gone for the day (east coast)21:37
melwittmriedem: +221:37
*** yamamoto has joined #openstack-nova21:37
mriedemthanks21:37
openstackgerritMerged openstack/nova stable/queens: Revert "Refine waiting for vif plug events during _hard_reboot"  https://review.openstack.org/55381721:39
imacdonnSomething completely unrelated to bounce off you, mriedem (or anyone else)21:39
imacdonnper Queens release notes, I tried to remove neutron.url from my nova.conf, but it seems to be unable to get the neutron endpoint from the service catalog21:40
mriedemefried: replied to all comments in https://review.openstack.org/#/c/554703/ and i don't think any of them are worth changing21:40
imacdonnI had someone in #openstack buddy-check my config, and it seems sane21:40
mriedemimacdonn: there is a bug fix that you need,21:40
mriedemsec21:40
*** esberglu_ has joined #openstack-nova21:40
imacdonnk ;)21:40
mriedemimacdonn: https://github.com/openstack/nova/commit/3a3b0f09db318faf1a1ea711a73bb365cab8b23321:40
imacdonnmriedem: Interesting, Looks pertinent. Will try it. Thanks!21:41
*** yamamoto has quit IRC21:43
efriedmriedem: Brain fart, sorry 'bout that.  +1.21:44
*** lpetrut has quit IRC21:44
mriedemefried: np, thanks for the quick review21:45
*** esberglu_ has quit IRC21:45
*** josecastroleon has quit IRC21:47
mriedemhongbin: yikun: Kevin_Zheng: easy bug https://bugs.launchpad.net/nova/+bug/175727321:47
openstackLaunchpad bug 1757273 in OpenStack Compute (nova) "nova-compute fails to start even if [placement]/region_name is set" [Medium,Triaged]21:47
*** gouthamr has quit IRC21:49
*** salv-orlando has joined #openstack-nova21:53
openstackgerritMatt Riedemann proposed openstack/nova master: Remove RequestContext.instance_lock_checked  https://review.openstack.org/55437821:53
*** READ10 has quit IRC21:56
*** tssurya has quit IRC22:00
*** gjayavelu has quit IRC22:02
*** sar has quit IRC22:09
*** germs has joined #openstack-nova22:10
*** germs has quit IRC22:10
*** germs has joined #openstack-nova22:10
*** rcernin has joined #openstack-nova22:13
*** Guest64768 has quit IRC22:13
*** germs has quit IRC22:14
*** priteau has quit IRC22:15
*** priteau has joined #openstack-nova22:15
*** slagle has quit IRC22:16
*** dave-mccowan has quit IRC22:17
*** slagle has joined #openstack-nova22:17
*** felipemonteiro_ has quit IRC22:18
*** oomichi has joined #openstack-nova22:18
*** AlexeyAbashkin has joined #openstack-nova22:19
*** priteau has quit IRC22:20
*** awaugama has quit IRC22:20
*** liverpooler has quit IRC22:23
*** AlexeyAbashkin has quit IRC22:23
*** dave-mccowan has joined #openstack-nova22:28
mriedemi wonder if anyone that uses DVR also uses shared ephemeral storage (rbd imagebackend) and live migration (HP cloud anyone?)22:29
mriedembecause i'm pretty sure we don't cleanup on failed live migration properly in that case22:29
mriedemhttps://github.com/openstack/nova/blob/3fd863d8bf2fa1fc09acd08d976689462cffd2e3/nova/compute/manager.py#L6506 will also cleanup some stuff we put in the port's binding profile for DVR during live migration,22:30
mriedembut if do_cleanup is False, which it is if you're using shared ephemeral storage, then we don't clean that up22:30
*** yamamoto has joined #openstack-nova22:39
efriedmriedem: Are you talking specifically libvirt?22:40
efriedCause that sounds like something we either support already or will support shortly in PowerVM (out of tree).22:41
mriedemfor shared local disk?22:41
mriedemhttps://github.com/openstack/nova/blob/3fd863d8bf2fa1fc09acd08d976689462cffd2e3/nova/compute/manager.py#L618522:42
mriedemlooks like that is also checked for xen and hyperv22:42
mriedemanyway, https://github.com/openstack/nova/blob/3fd863d8bf2fa1fc09acd08d976689462cffd2e3/nova/compute/manager.py#L6524 is totally doing more than just cleaning up local disk created on the dest host for non-shared storage22:43
mriedemit's also cleaning up network stuff on the dest host22:43
*** hongbin has quit IRC22:44
mriedemadded back in mitaka https://review.openstack.org/#/c/227897/22:45
mriedemoh nvm, that was the error handling22:45
*** yamamoto has quit IRC22:45
*** andreas_s has joined #openstack-nova22:47
mriedemwow added long ago https://review.openstack.org/#/c/4646/22:47
mriedemessex22:47
efriedGotta run22:48
*** andreas_s has quit IRC22:51
*** gjayavelu has joined #openstack-nova22:56
*** harlowja has quit IRC22:57
*** _ix has quit IRC22:57
*** chyka has quit IRC23:00
*** chyka has joined #openstack-nova23:01
*** vladikr has quit IRC23:01
*** vladikr has joined #openstack-nova23:01
*** masber has joined #openstack-nova23:03
*** masuberu has joined #openstack-nova23:04
*** masber has quit IRC23:08
*** AlexeyAbashkin has joined #openstack-nova23:19
*** jpena|away is now known as jpena|off23:21
*** AlexeyAbashkin has quit IRC23:23
*** Zames has joined #openstack-nova23:25
openstackgerritMichael Still proposed openstack/nova master: Move configurable mkfs to privsep.  https://review.openstack.org/55192123:27
openstackgerritMichael Still proposed openstack/nova master: Move xenapi xenstore_read's to privsep.  https://review.openstack.org/55224123:27
openstackgerritMichael Still proposed openstack/nova master: Move xenapi disk resizing to privsep.  https://review.openstack.org/55224223:27
openstackgerritMichael Still proposed openstack/nova master: Sync xenapi and libvirt on what flags to pass e2fsck.  https://review.openstack.org/55407823:27
openstackgerritMichael Still proposed openstack/nova master: Move xenapi partition copies to privsep.  https://review.openstack.org/55360523:27
openstackgerritMichael Still proposed openstack/nova master: Move image conversion to privsep.  https://review.openstack.org/55443723:27
openstackgerritMichael Still proposed openstack/nova master: We no longer need rootwrap.  https://review.openstack.org/55443823:27
openstackgerritMichael Still proposed openstack/nova master: We don't need utils.trycmd any more.  https://review.openstack.org/55443923:27
*** Anticime1 is now known as Anticimex23:29
*** Zames has quit IRC23:30
*** mriedem has quit IRC23:34
*** mlavalle has quit IRC23:36
*** yamamoto has joined #openstack-nova23:41
*** chyka has quit IRC23:46
*** yamamoto has quit IRC23:46
*** amodi has quit IRC23:46
*** harlowja has joined #openstack-nova23:49
*** cdent has quit IRC23:50
*** mriedem has joined #openstack-nova23:52
openstackgerritmelanie witt proposed openstack/nova master: Remove useless run_periodic_tasks call in ClientRouter  https://review.openstack.org/55438123:59

Generated by irclog2html.py 2.15.3 by Marius Gedminas - find it at mg.pov.lt!