Wednesday, 2019-02-20

*** akhil_jain has joined #openstack-monasca05:49
*** openstackgerrit has joined #openstack-monasca07:44
openstackgerritDobroslaw Zybort proposed openstack/monasca-api master: Fix docker hub login variable  https://review.openstack.org/63810007:44
*** pcaruana has joined #openstack-monasca07:59
*** dougsz has joined #openstack-monasca09:09
*** witek has joined #openstack-monasca09:17
openstackgerritDobroslaw Zybort proposed openstack/python-monascaclient master: Update hacking version  https://review.openstack.org/62772510:11
openstackgerritMerged openstack/monasca-persister master: Update hacking version to latest  https://review.openstack.org/62876910:43
openstackgerritMerged openstack/monasca-api master: Add keystonemiddleware to oslo-config-generator conf  https://review.openstack.org/63352211:06
*** akhil_jain has quit IRC11:33
openstackgerritMerged openstack/monasca-api master: Fix docker hub login variable  https://review.openstack.org/63810011:41
openstackgerritDobroslaw Zybort proposed openstack/monasca-api master: Update default docker monasca-api conf  https://review.openstack.org/63813912:42
*** spsurya has joined #openstack-monasca13:44
openstackgerritWitold Bedyk proposed openstack/monasca-api master: Set up DB schema in DevStack plugin using Alembic  https://review.openstack.org/62236113:53
openstackgerritDobroslaw Zybort proposed openstack/monasca-api master: Update default docker monasca-api conf  https://review.openstack.org/63813913:56
openstackgerritDobroslaw Zybort proposed openstack/monasca-api master: Update default docker monasca-api conf  https://review.openstack.org/63813914:05
*** mohankumar has joined #openstack-monasca14:54
*** koji_n has joined #openstack-monasca15:00
witek#startmeeting monasca15:01
openstackMeeting started Wed Feb 20 15:01:33 2019 UTC and is due to finish in 60 minutes.  The chair is witek. Information about MeetBot at http://wiki.debian.org/MeetBot.15:01
openstackUseful Commands: #action #agreed #help #info #idea #link #topic #startvote.15:01
*** openstack changes topic to " (Meeting topic: monasca)"15:01
openstackThe meeting name has been set to 'monasca'15:01
witekhello everyone15:01
koji_nhello15:01
witekhi koji_n15:01
DobroslawHi15:02
mohankumarhi eveyone15:02
witekhi Dobroslaw and mohankumar15:02
witekCourtesy Monasca meeting reminder in #openstack-monasca: witek, jayahn,iurygregory,ezpz,igorn,haad,sc,joadavis, akiraY,tobiajo,dougsz_,fouadben, amofakhar, aagate, haruki,kaiokmo,pandiyan,charana,guilhermesp,chaconpiza,toabctl15:03
dougszhi all15:03
schere I'm15:03
witekagenda:15:03
witekhttps://etherpad.openstack.org/p/monasca-team-meeting-agenda15:03
witek#topic reviews15:03
*** openstack changes topic to "reviews (Meeting topic: monasca)"15:03
witekI have just one new one:15:03
witekhttps://review.openstack.org/63633715:04
joadavismore tests are appreciated. will have to read that one15:04
witekI have also updated the alembic DevStack change to install missing jira dependency for monasca-notification15:05
witekhttps://review.openstack.org/62236115:05
witeksome other reviews to pull attention to?15:06
dougszThere was this one: https://review.openstack.org/#/c/637190/15:07
witekoh yes, thanks dougsz15:08
Dobroslawdougsz: thx for comments15:08
Dobroslawmichal is still thinking about it15:08
Dobroslawhe said that api is gladly taking metrics with old timestamp but persister is not so happy about them15:09
Dobroslawhe will look into it more15:09
dougszSounds good, great to have another contributor.15:09
dougszthanks Dobroslaw15:09
Dobroslawyep, hi is learning quite fast15:10
Dobroslaw*he*15:10
witekpersister will drop messages based on retention period15:10
*** mayankkapoor has joined #openstack-monasca15:11
Dobroslawshouldn't they be dropped already in the API?15:12
mohankumarwitek : qq , persister have rentention policy ? or you are referring InfluxDB retention DB retention policies ?15:13
witekwe don't control retention policy in the API15:13
witekmohankumar: InfluxDB retention period15:13
mohankumarwitek : okay15:14
Dobroslawok15:14
witekDobroslaw: we could do that, but that's additional logic which have to be done on every message15:14
Dobroslawhmmm, yea15:15
witeklet's discuss it further in review15:16
Dobroslawok15:16
witekcan we move on to the next topic?15:17
witek#topic Reliance Jio deployment15:17
*** openstack changes topic to "Reliance Jio deployment (Meeting topic: monasca)"15:17
mayankkapoorHi all15:17
witekhi mayankkapoor15:17
mayankkapoorSorry it's been a while15:17
mayankkapoorWanted to give a quick update15:17
mayankkapoorI've mentioned status of deployment in meeting agenda15:18
witekthanks, great to hear that15:18
mayankkapoorDeployed across 352 bare-metals at the moment (single openstack cluster), working fine, few issues we're working through when they come15:18
mayankkapoorAny specific items I should talk about?15:19
joadavisAre there any of the monasca metrics you are finding particularly useful?15:19
witekyou've deployed with Docker Swarm, would it be possible to share configuration?15:20
mayankkapoor@joadavis: We've started with CPU and RAM mainly. We have built a custom console for our users, and we're showing these in that UI.15:20
openstackgerritweizj proposed openstack/python-monascaclient master: Update hacking version  https://review.openstack.org/62772515:20
mayankkapoor@witek: Sure, yes sure no problem. How do you recommend we share the config? github.com?15:21
openstackgerritweizj proposed openstack/python-monascaclient master: Update hacking version  https://review.openstack.org/62772515:22
witekyes, github would be great15:22
mayankkapoorOk let me upload to github.com and send the link here15:22
witekgreat, thanks15:22
Dobroslawmayankkapoor: any monasca components working slower/unexpectly than others?15:22
witekhave you managed to improve persister performance15:22
witek?15:22
mayankkapoor@Dobroslaw: We had some issues with Timeouts on InfluxDB, but that was mainly due to bombarding InfluxDB with large batch sizes (50k) and lots of persisters (5).15:23
mayankkapoorNow we're using 5k batch size and 10 persisters. mohankumar works with me and can confirm latest config15:24
openstackgerritweizj proposed openstack/python-monascaclient master: Update hacking version  https://review.openstack.org/62772515:24
mayankkapoor@witek: Things are working fine with persisters and InfluxDB now with 5k batch and 10 persisters15:25
mayankkapoorStill early days through15:25
dougszmayankkapoor: How many metrics a second do you ingest?15:25
mayankkapoorWe've also built our own monitoring scripts for errors/warnings we got, which we'll share in github.com repo15:25
mohankumarwitek: hi , I m concerned about DB writing speed . We use API post to write into DB , Cluster DB would help to scale DB and perforamnce15:26
mayankkapoor@dougsz: Think I will need some help how to calculate the exact number. So 352 bare-metals with monasca-agent sending libvirt plugin data every 30 seconds. Roughly 5 VMs on each bare-metal.15:27
joadavisIf I had more expertise, I'd love to share the Cassandra setup with the docker based installs. :/15:27
joadaviswe use Cassandra in a nicely clustered setup, but we still install using our own methods15:28
witekmohankumar: InfluxDB can ingest >1.000.000 measurements/s, the bottleneck is the persister15:29
witekin particular the old Kafka consumer15:29
witekI hope I can provide the new implementation soon15:29
mohankumarwitek : im getting InfluxDB timeout error if i add more batch size15:30
mayankkapoor@dougsz: We can share some data from our Kafka-admin page to give you some idea about TPS15:30
mayankkapoor^^^ increase batch size15:30
dougszthanks mayankkapoor, i'm always interested to hear about performance at scale15:31
mayankkapoorMain thing we did for this setup was use GlusterFS for HA across three docker worker VMs15:31
mayankkapoorThis was a huge risk15:31
mohankumarwitek : just to add on Persister giving  InfluxDB timeout error if i increasing batch size15:32
mayankkapoorHowever, we reasoned that we're not running active-active containers, so might be ok15:32
mayankkapoorSo when a stateful container dies, it respawns on another node and it has access to the same data it had previously15:33
mayankkapoorWe tested each component, MySQL, InfluxDB and monasca containers individually for HA with GlusterFS. Then proceeding for prod deployment.15:34
witekso you add HA to InfluxDB that way15:34
mayankkapoorYup15:34
witekdo you know what is performance impact?15:34
joadaviscoool15:35
mayankkapoorHmm no we don't yet, haven't gotten around to testing a setup without GlusterFS and comparing15:35
dougsz mayankkapoor: Are you using rdma transport for Gluster share?15:36
mayankkapoor@dougsz: Hmm need to check. We're using a GlusterFS replicated volume with 3 replicas, and we haven't changed any of the defaults.15:36
dougszCool, there is also nufa which is quite neat - if your gluster storage is local to the docker workers it can right directly to a local drive15:37
dougszThis might be a useful reference, we use it for HPC activities: https://github.com/stackhpc/ansible-role-gluster-cluster15:38
mayankkapoorBased on reading the GlusterFS docs, RDMA transport needs to be enabled. So no, we haven't enabled RDMA yet.15:38
mayankkapoorYeah our gluster storage is local to the worker VMs15:39
mayankkapoorHmm wait, we're using Ceph15:39
mayankkapoorSo not local15:39
mayankkapoorUsed Ceph RBD15:39
dougszAh ok15:39
mayankkapoorSo we'll share our deployment doc and config on github.com for review15:40
witekyou mentioned problems with network metrics, do you mean standard system plugin metrics?15:41
mayankkapoor@witek: We've disabled following in libvirt plugin: ping_check=false vm_ping_check_enable=false vm_network_check_enable=false15:43
witeki see, thanks15:44
mayankkapoorIf we enable vm_network_check_enable, it loads our OpenContrail controllers too much. We tested load balancing on our OpenContrail controllers, and that worked fine15:44
mayankkapoorCurrent hypothesis that we need to test is that monasca-agent is getting some unexpected response from the SDN controllers, and keeps querying the controllers rapidly15:45
mayankkapoorHaven't gotten around to checking this at the moment.15:45
mayankkapoorRather than every 30 sec, it querying faster15:45
witekplease report a bug if you can confirm that15:46
mayankkapoorsure15:46
witekwhat OpenStack distribution do you use?15:47
mayankkapoorMonasca deployment is for Ocata cloud. However, We have 7 production clouds with between 100-500 bare-metals each, and staging environments for each. Various versions of openstack. Oldest are on Liberty, latest on Pike.15:49
mohankumarwitek : Ocata15:49
witekwhich OS?15:49
mayankkapoorUbuntu160415:50
witekmeaning, Ubuntu agent packages would be handy15:50
mayankkapoorYes. However, we use the Mirantis distribution of openstack (they provide L4 support for us). So getting these bundled in the OS is a bit challenging for us.15:51
witekthanks for the update15:53
mayankkapoorYou're welcome, and thanks for the great work on Monasca15:53
witekthat great to hear your feedback15:53
mayankkapoorWe'll try to figure out how we can contribute further.15:53
mohankumarquestion : does the current monasca help me to get vm disk usage , if I'm using ceph storage. I can see baremetal (cmp nodes) disk usage , not from VM . If I enable monasca ceph plugin as per document   https://github.com/openstack/monasca-agent/blob/master/docs/Plugins.md#ceph . I'll get only ceph cluster metrics not from Individual VM . Is there any way in current monasca i can get from each vm with  ceph storage?15:53
mohankumarI hope this question based on the agenda line what we ve :)15:55
witekI'm afraid I cannot answer that right now15:55
witekI'll check15:56
mohankumarwitek : sure thanks .15:56
witekgood support for Ceph is important, so if there's anything missing, we should think about closing the gaps15:57
dougszWe've got update for Luminous support in the pipeline15:57
dougszThe existing plugin can't handle the newer releases15:58
witekthanks again for the update15:58
witekI have to finish the meeting soon15:58
witek#topic backlog15:58
*** openstack changes topic to "backlog (Meeting topic: monasca)"15:58
witekshort update15:58
dougszyep, thanks mohankumar15:58
witekI have added two stories to the backlog15:59
witekespecially running Python3 unit tests for monasca-agent15:59
witekwe still don't run them16:00
witekall from me16:00
witekthanks for joining16:00
witeksee you next time16:00
Dobroslawone info16:00
DobroslawAPI docker image is now pushed from zuul on master, need to wait for tagging to see if it will be pushed with proper tag and then I will replace first image on github.com/monasca/monasca-docker with this one16:00
witek#endmeeting16:00
*** openstack changes topic to "OpenStack Monitoring as a Service | https://wiki.openstack.org/wiki/Monasca"16:00
koji_nthx16:00
openstackMeeting ended Wed Feb 20 16:00:26 2019 UTC.  Information about MeetBot at http://wiki.debian.org/MeetBot . (v 0.1.4)16:00
openstackMinutes:        http://eavesdrop.openstack.org/meetings/monasca/2019/monasca.2019-02-20-15.01.html16:00
openstackMinutes (text): http://eavesdrop.openstack.org/meetings/monasca/2019/monasca.2019-02-20-15.01.txt16:00
Dobroslawthx16:00
openstackLog:            http://eavesdrop.openstack.org/meetings/monasca/2019/monasca.2019-02-20-15.01.log.html16:00
*** koji_n has quit IRC16:00
joadavisThanks mayankkapoor and mohankumar. Really find feedback valuable16:01
mayankkapoorYou're welcome and looking forward to chatting again.16:01
*** mayankkapoor has quit IRC16:01
joadavisWoohoo!  The session "Heat and Monasca" was accepted for the Denver summit.  The Monasca Events submission was not, but that's ok.16:34
joadavisI see that Pandiyan's presentation has an overlapping time slot on Wednesday with the Monitoring and Analyzing workstop.  Too bad, I want to see both16:43
*** mohankumar has quit IRC16:52
*** pcaruana has quit IRC17:13
witekperhaps it would be possible to reschedule one of these17:15
*** dougsz has quit IRC18:01
*** dougsz has joined #openstack-monasca18:01
*** dougsz has quit IRC18:01
*** joadavis has quit IRC18:29
*** joadavis has joined #openstack-monasca18:32
*** witek has quit IRC20:37
*** spsurya has quit IRC20:42

Generated by irclog2html.py 2.15.3 by Marius Gedminas - find it at mg.pov.lt!