Tuesday, 2022-03-15

*** dviroel|ruck is now known as dviroel|out00:04
opendevreviewMerged openstack/openstack-ansible-galera_server master: Change location of ipaddr filter  https://review.opendev.org/c/openstack/openstack-ansible-galera_server/+/83152600:04
*** aussielunix is now known as aussielunix_01:20
*** aussielunix_ is now known as aussielunix01:23
*** anbanerj is now known as frenzyfriday07:38
*** arxcruz|off is now known as arxcruz07:47
BraceI've lost VM networking on my openstack cluster, any idea what this error might mean - https://pastebin.com/EfW5WRsk09:08
BraceSo far I've tried to restart the various neutron components and also rebooted (separately) all of my controller nodes09:08
noonedeadpunkBrace: have an idea09:12
noonedeadpunkBrace: try this out https://paste.openstack.org/show/bY235whPe5LKkFFzo6pn/09:13
noonedeadpunkYou might want to apply smth simmilar to cinder_volume_init_overrides and nova_compute_init_overrides as well09:14
Bracenoonedeadpunk: I'll try that out, thank you so much!09:19
*** arxcruz is now known as arxcruz|brb09:20
opendevreviewAndrew Bonney proposed openstack/openstack-ansible-os_nova master: Add configuration option for heartbeat_in_pthread  https://review.opendev.org/c/openstack/openstack-ansible-os_nova/+/83323610:41
opendevreviewAndrew Bonney proposed openstack/openstack-ansible-os_neutron master: Add configuration option for heartbeat_in_pthread  https://review.opendev.org/c/openstack/openstack-ansible-os_neutron/+/83323710:42
opendevreviewAndrew Bonney proposed openstack/openstack-ansible-os_cinder master: Add configuration option for heartbeat_in_pthread  https://review.opendev.org/c/openstack/openstack-ansible-os_cinder/+/83323810:42
opendevreviewAndrew Bonney proposed openstack/openstack-ansible-os_cinder master: Add configuration option for heartbeat_in_pthread  https://review.opendev.org/c/openstack/openstack-ansible-os_cinder/+/83323811:00
opendevreviewAndrew Bonney proposed openstack/openstack-ansible-os_nova master: Add configuration option for heartbeat_in_pthread  https://review.opendev.org/c/openstack/openstack-ansible-os_nova/+/83323611:00
*** dviroel|out is now known as dviroel|ruck11:09
opendevreviewMerged openstack/openstack-ansible-tests master: Add ansible.utils collection requirement  https://review.opendev.org/c/openstack/openstack-ansible-tests/+/83359611:24
opendevreviewDmitriy Rabotyagov proposed openstack/openstack-ansible-galera_server master: Use separate tmp directory  https://review.opendev.org/c/openstack/openstack-ansible-galera_server/+/83155011:26
opendevreviewDmitriy Rabotyagov proposed openstack/openstack-ansible-galera_server master: Add galera_data_dir variable  https://review.opendev.org/c/openstack/openstack-ansible-galera_server/+/83155211:28
gokhaniHi folks, How can we use Mellanox infiniband ports with OSA? 11:33
opendevreviewAndrew Bonney proposed openstack/openstack-ansible-os_neutron master: Add parameters to limit the number of DHCP or L3 agents  https://review.opendev.org/c/openstack/openstack-ansible-os_neutron/+/83376911:34
gokhaniHi folks, how can we use Mellanox İnfiniband ports with OSA?11:35
jrossergokhani: that can mean several things but you could look at this https://satishdotpatel.github.io/HPC-on-openstack/11:36
opendevreviewDmitriy Rabotyagov proposed openstack/openstack-ansible-galera_server master: Update MariDB version to 10.6.7  https://review.opendev.org/c/openstack/openstack-ansible-galera_server/+/83325911:37
noonedeadpunkgokhani: yep, I did that and it was nasty :)11:39
noonedeadpunkgokhani: the main thing is that you will need to use sr-iov to pass ib devices inside containers11:39
noonedeadpunksince they can't be bridged11:39
noonedeadpunkif you're talking about IB for storage network as an example11:39
gokhanijrosser: thanks I will try11:41
*** arxcruz|brb is now known as arxcruz11:42
gokhaninoonedeadpunk: yes firstly I tried to use IB for storage network11:42
jrossergokhani: as noonedeadpunk says it depends what you want11:42
jrosserbecasue that link i give you is nothing to do with storage, just VM<>VM MPI workloads11:42
noonedeadpunkhm, I think I need to document that :)11:43
gokhaninoonedeadpunk: I need your documentation :) preliminary comments also will also help 11:46
opendevreviewDmitriy Rabotyagov proposed openstack/openstack-ansible master: [doc] Add doc how to pass SR-IOV inside containers  https://review.opendev.org/c/openstack/openstack-ansible/+/83377512:11
opendevreviewMerged openstack/openstack-ansible-lxc_container_create master: Change location of ipaddr filter  https://review.opendev.org/c/openstack/openstack-ansible-lxc_container_create/+/83311812:17
opendevreviewDmitriy Rabotyagov proposed openstack/openstack-ansible-galera_server master: Add galera_data_dir variable  https://review.opendev.org/c/openstack/openstack-ansible-galera_server/+/83155212:27
opendevreviewMerged openstack/openstack-ansible-plugins master: Change location of ipaddr filter  https://review.opendev.org/c/openstack/openstack-ansible-plugins/+/83153012:42
opendevreviewMerged openstack/openstack-ansible-os_keystone stable/xena: add oauth support  https://review.opendev.org/c/openstack/openstack-ansible-os_keystone/+/83355212:53
opendevreviewDmitriy Rabotyagov proposed openstack/openstack-ansible-plugins master: Add reference_group support to provider_networks module  https://review.opendev.org/c/openstack/openstack-ansible-plugins/+/82974113:19
opendevreviewJames Denton proposed openstack/openstack-ansible-os_neutron master: Update Mellanox ASAP^2 Documentation  https://review.opendev.org/c/openstack/openstack-ansible-os_neutron/+/83379514:05
opendevreviewMerged openstack/openstack-ansible-plugins master: Add reference_group support to provider_networks module  https://review.opendev.org/c/openstack/openstack-ansible-plugins/+/82974114:15
*** dviroel|ruck is now known as dviroel|ruck|mtg14:16
spatelnoonedeadpunk around? i have dumb question related kubernetes14:18
spatelI have deploy k8s cluster on my lab openstack using magnum. I don't have octavia lb. How do i expose my hello-world ngnix to outside world? 14:20
spatelare there anyway or octavia is hard stop here? 14:20
gokhaninoonedeadpunk: one of my customer requested using zfs storage for their HPC clusters. Can I use zfs pool which shared with nfs for vm disks or glance images? 14:22
noonedeadpunkum, yes, nfs can be used as backend for cinder and glance and nova14:24
gokhanispatel: https://docs.openstack.org/magnum/latest/user/#ingress_controller you can use also trafeik or nginx14:25
noonedeadpunkspatel: I never ran magnum without octavia :) I guess you can jsut not in HA manner14:25
spatelnoonedeadpunk i am learning and trying to understand how kubernetes work if octavia not exist 14:27
gokhaninoonedeadpunk: thanks for https://review.opendev.org/c/openstack/openstack-ansible/+/833775 and ı think I need to also enable sriov 14:27
spatelI can understand HA and redundancy won't be there if no octavia. This is just for learning 14:27
spatelgokhani thanks 14:29
spatelgokhani Are you doing HPC on openstack? because i am also doing it and would like to understand your setup also :)14:30
opendevreviewDmitriy Rabotyagov proposed openstack/openstack-ansible-galera_server master: Update MariDB version to 10.6.7  https://review.opendev.org/c/openstack/openstack-ansible-galera_server/+/83325914:33
gokhanispatel: yes but I am newbie to HPC. I struggled to how can we achive run hpc cluster on openstack14:34
gokhaniI am reading your blogpost :)14:34
spatelwe can help each other :)14:34
spatelI am trying to use glusterfs with manila to provide shared space for MPI job application 14:35
gokhaniyes definitely :) My customer requested using zfs storage and they say it is must :)14:37
noonedeadpunkjust in case - nfs is pretty bad idea to use as shared storage 14:38
noonedeadpunkany network issue and you stuck with mount that can't be unmounted so you need to reboot all computes14:38
gokhaninoonedeadpunk: you are right, I had a lof issues about nfs in some of my deployments, but I don't know how can I use zfs storage instead of nfs 14:42
spatelnoonedeadpunk nfs can be force umount with -l (lazy option) 14:43
noonedeadpunkspatel: when it's passed to libvirt?14:45
noonedeadpunkbecause it's not local mount14:45
noonedeadpunkand connection is not re-initiated14:45
spatelyou can use fuse command to kill attached pid 14:45
spatelbut that won't be clean 14:45
noonedeadpunkand then you have VM without disk?:)14:46
spatelyou are correct it won't be clean and neat14:46
noonedeadpunkso it's really all mess, corrupted databases and etc14:46
noonedeadpunkgokhani: and how you plan to distinguish userspace with zfs with openstack space? 14:47
noonedeadpunkIt sounds like you would need to have several storages anyway?14:47
gokhanispatel: my customer is running their applications on 4 gpu servers with 8 gpu cards (NVIDIA RTX A6000). I am planning tu use gpu passthrough. for gpu virtualization we need to nvidia vgpu license14:48
spatelMy customer doesn't have license so we decided to use passthrough 14:51
spatelvgpu is good if you have multiple folks trying to use infra. in my case customer itself using and he own everything in that case doesn't need vgpu14:52
spateldo you have infiniband network? 14:52
gokhanispatel: yes I have infiniband network and I am planing use as storage network 14:57
jrosseri think i also vote for mandating zfs being kind of orthoganal to using filesystems in openstack14:57
spatelIn my case IB is just for mpi job and storage will be on 10G dedicated nic14:57
spateljrosser ZFS is great but you need storage protocol like NFS or iSCSI 15:00
jrosseryes thats what i mean15:00
noonedeadpunk#startmeeting openstack_ansible_meeting15:01
opendevmeetMeeting started Tue Mar 15 15:01:21 2022 UTC and is due to finish in 60 minutes.  The chair is noonedeadpunk. Information about MeetBot at http://wiki.debian.org/MeetBot.15:01
opendevmeetUseful Commands: #action #agreed #help #info #idea #link #topic #startvote.15:01
opendevmeetThe meeting name has been set to 'openstack_ansible_meeting'15:01
*** dviroel|ruck|mtg is now known as dviroel|ruck15:01
noonedeadpunk#topic rollcall15:01
spatel\o/15:01
NeilHanlono/ heya folks, hope everyone is doing well15:02
mgariepy\o15:04
noonedeadpunk#topic office hours15:10
noonedeadpunksorry I bit distracted internally15:10
noonedeadpunkhaving some nesty issue for weeks now with mariadb that drops connection but thinks it alive so leaves lock on table until timeout for connection ends15:11
damiandabrowski[m]hey! (sorry for being late)15:13
noonedeadpunkWe've got approval that https://bugs.launchpad.net/openstack-ansible/+bug/1955676 works nicely15:15
noonedeadpunk*solution for 15:15
noonedeadpunkWhich is https://review.opendev.org/c/openstack/openstack-ansible-os_gnocchi/+/82286015:15
noonedeadpunkSo probably would be great to have it reviewed15:15
NeilHanlonbeen looking at the graylog role in the ops repo, seems fairly straightforward but was thinking of swapping out the current journal-to-graylog converter that's deprecated with https://github.com/parse-nl/SystemdJournal2Gelf, maybe?15:16
noonedeadpunkAnother nasty thing we also hitted after X upgrade was https://review.opendev.org/q/topic:bug%252F1961603 so worth checking it as well15:16
noonedeadpunkNeilHanlon: well... the only tricky thing is that it requires go everywhere?15:17
noonedeadpunkwhich is.... meh...15:17
NeilHanlonprobably would create deb/rpm packages for it15:18
NeilHanloni'm not against updating the current python one either15:18
noonedeadpunkwe just don't have anything that requires go atm, and we're trying to leave things as minimalistic as possible. 15:19
noonedeadpunkbut yeah. I saw that journal2gelf is not supported for quite a while which is quite sad15:20
NeilHanlongotcha, makes sense. I'll look at updating the journal2gelf module. it didn't look too bad, just some py2 -> py3 stuff that needs fixing15:20
NeilHanlonmostly around bytes and strings and all that fun :) 15:20
spatelNeilHanlon why don't we use journalbeat to push journal to graylog?15:20
noonedeadpunkwell, https://github.com/nailgun/journal2gelf has 2.1.0 jsut in case. So it lloks like it jsut wasn't pushed to pypi?15:21
NeilHanlonspatel: i'm not familiar w/ journalbeat, but sounds like an interesting option, maybe?15:22
spatelNeilHanlon that is what i am using and works great 15:22
NeilHanlonyeah I installed right from the master branch on a test box noonedeadpunk and it still seemed to have some issues15:22
noonedeadpunkbut um, what point of having journalbeat with graylog?:)15:22
noonedeadpunkisn't this part of elk?:)15:22
NeilHanlonyeah that was sorta my next thing heh15:23
noonedeadpunkah I see15:23
spatelI don't have ELK infra (I am using graylog outside OSA and beat to send logs)15:23
noonedeadpunkNext to that I tried to use system-scope stuff and keystone changes, but that has barely merged to openstack collection as well as openstacksdk, and ofc is broken in several places. Was trying to investigate but have quite a lot of internal issues I have to address first...15:24
NeilHanloni think journalbeat is also golang, fwiw15:25
noonedeadpunkThen I'd prefer SystemdJournal2Gelf, since at least it's BSD licensed...15:26
spatelYes but single RPM and i have not installed journalbeat on all containers but just on physical node like infra/compute. 15:27
spateli point Journalbeat to point /openstack/logs/ to push all container logs to graylog15:28
noonedeadpunkBut journal from all containers is in /var/log/journal/ ?15:29
spateldamn it correct sorry my bad - https://paste.opendev.org/show/bYg0RxxDdBZjW2gRuIIg/15:32
*** dviroel|ruck is now known as dviroel|ruck|lunch15:35
jrosserhello15:37
spatelhello15:38
NeilHanlonlooking again with fresh eyes, it appears someone's already done some of the work for journal2gelf working https://github.com/nailgun/journal2gelf/pull/6/files15:38
jrosseri was just taking another look at the molecule stuff if we wanted to discuss role tests15:39
opendevreviewJonathan Rosser proposed openstack/ansible-role-pki master: Refactor conditional generation of CA and certificates  https://review.opendev.org/c/openstack/ansible-role-pki/+/83079415:48
opendevreviewAndrew Bonney proposed openstack/openstack-ansible-os_neutron master: Add parameters to limit the number of DHCP or L3 agents  https://review.opendev.org/c/openstack/openstack-ansible-os_neutron/+/83376915:56
opendevreviewMerged openstack/openstack-ansible-haproxy_server master: Change location of ipaddr filter  https://review.opendev.org/c/openstack/openstack-ansible-haproxy_server/+/83152815:58
noonedeadpunk#endmeeting16:03
opendevmeetMeeting ended Tue Mar 15 16:03:18 2022 UTC.  Information about MeetBot at http://wiki.debian.org/MeetBot . (v 0.1.4)16:03
opendevmeetMinutes:        https://meetings.opendev.org/meetings/openstack_ansible_meeting/2022/openstack_ansible_meeting.2022-03-15-15.01.html16:03
opendevmeetMinutes (text): https://meetings.opendev.org/meetings/openstack_ansible_meeting/2022/openstack_ansible_meeting.2022-03-15-15.01.txt16:03
opendevmeetLog:            https://meetings.opendev.org/meetings/openstack_ansible_meeting/2022/openstack_ansible_meeting.2022-03-15-15.01.log.html16:03
*** dviroel|ruck|lunch is now known as dviroel|ruck16:03
noonedeadpunkjrosser: I had no time for functional tests despite I started them :(16:04
jrosserheh no problem - i have been away for ~2 weeks also16:04
jrosserthough i don't want to waste my time on it either16:04
noonedeadpunkso the whole idea was to use integrated tests and jsut run tests/test.yml with tests/inventory.ini if present. and based on zuul var we know what project is that16:05
noonedeadpunkas alternative to molecule that should be easier to manage16:05
noonedeadpunkand then define such jobs and define vars for jobs if needed, that would end up in user_variables16:05
noonedeadpunkjobs can be defined in any project, jsut have same parent 16:06
jrosseri would like something thats really useful for role development16:11
jrosserand quick turnaround teardown/re-run is really helpful for that16:11
jrosserand for a role unit test there might be more than one situation to test too16:12
jrosser"does it work with just role defaults" / "test things A" / "test things B"16:13
opendevreviewMerged openstack/openstack-ansible-galera_server master: Improve incremental backups rotation in mariabackup script  https://review.opendev.org/c/openstack/openstack-ansible-galera_server/+/82817016:15
noonedeadpunkyes, but we need to have depends-on working as tons of stuff rely on each other. And with that I was thinking to run jsut bootstrap-ansible without bootstrap-aio16:17
noonedeadpunkso it should be really fast16:17
jrosserdo you think we concentrate too much on integration testing?16:17
jrosserthere is almost no unit testing16:17
noonedeadpunkbut why we can't launch unit testing same way?16:18
jrosseri was adding more stuff to the PKI role for the unfinished keystone patch and just don't know if i break something else16:18
noonedeadpunkI mean - it's same scenarious that will be launched, jsut not with molecule, but with zuul job?16:18
jrosseroh well right, so its a question of the test environment then i guess16:19
noonedeadpunkyup16:19
jrossereither VM + openstack-ansible + lxc + tests/test.yml16:19
jrosseror VM + moecule/role + docker/podman + many scenarios16:20
noonedeadpunk(we kind of don't need even LXC I guess)16:20
jrosseroh really we do16:20
jrosserbecause if you run the tests once then you can't guarantee to clean everything up properly if you fix things then want to try again16:20
jrosserreally i am most concerned about local development being nice rather than zuul stuff16:21
noonedeadpunkI just see nightmare of managing molecule in terms of supported os, when we'd need to manually patch all roles, not working dependancies, ansible versions....16:21
noonedeadpunkah, local development is question indeed.16:21
noonedeadpunkI was thinking purely CI16:21
jrosserit takes really very long to boostrap some LXC with openstack-ansible16:22
jrosserand same again to tear down / re-create16:22
noonedeadpunkwhere we don't need to destroy lxc and re-run16:22
jrossercurrently i work on the PKI role16:23
noonedeadpunkbut I don't see how we can manage molecule either. It's even more messier then we had functional testing with tests repo16:23
jrosserand it's complex enough that i'm very much concerned that i break it16:23
jrosserare there specific things that need addressing?16:23
opendevreviewNeil Hanlon proposed openstack/openstack-ansible master: Use the short form nodeset definition once it's merged upstream  https://review.opendev.org/c/openstack/openstack-ansible/+/83387716:24
noonedeadpunkI'd say whole https://review.opendev.org/c/openstack/ansible-role-pki/+/831236/3/molecule/molecule.yml ?16:25
noonedeadpunklike platforms, ansible-lint version and test-galaxy-requirements.yml as well16:26
noonedeadpunknot saying about ansible version itself...16:26
jrosserwell, this is the very first patch i make16:27
jrosserwith no attempt to make any of that proper16:27
noonedeadpunkI mean when we will have that everywhere, we would need to patch all roles each release kind of...16:27
noonedeadpunkand with stuff like jsut raised with netcommon....16:27
noonedeadpunkwe need to have env defined in one place16:28
jrosseri have removed the lint as it's duplicate16:28
jrosseri was going to copy the requirements file from openstack-ansible repo16:28
jrosseras in, copy at run-time16:29
noonedeadpunkand we can set image as variable as well?16:32
noonedeadpunkok, then we will jsut don't have depends-on working which is likely fine if we don't ever depend on some other role16:33
noonedeadpunkwhich we shouldn't have anyway for unit testing16:33
noonedeadpunkand we need to somehow control molecule version16:34
jrosserwe have a place for that sort of thing already https://github.com/openstack/openstack-ansible/blob/master/test-requirements.txt16:35
jrosserwell anyway - sounds like i should leave this alone for a while16:36
noonedeadpunkI;'d say you have point here for sure....16:37
jrossersome things we test quite thoroughly with infra scenario already, like DB cluster16:38
jrosserso i don't think there is any merit in refactoring things like that16:39
jrosserbut kind of 'tool' roles like pki, keypairs, config_template and so on could do with a refresh particularly if we want to deprecate openstack-ansible-tests16:40
jrosserthey are all pretty complex and need a test suite rather than just saying that AIO deploys OK16:40
noonedeadpunkyes, totally16:40
noonedeadpunkbut indeed I haven't thought about local development when was thinking through my idea with jsut adding functional scenario into integrational testing16:41
noonedeadpunkI just thought that we have everything prepared and it's super easy instead of aio just run tests.yml in exact same env we run everywhere16:42
noonedeadpunkthat would be super tricky indeed for local testing16:43
jrossersounds like we had started thinking about this from opposite ends :)16:43
noonedeadpunkyeah, indeed)16:43
gokhani  vb  vx xv  vgh<qvsxnnxvvsvsbsvzzzzzzzzzzbbbbbb                                                                x f jfze4munwök Plı*85  17:09
noonedeadpunkthat looks like yubikey hash...17:18
opendevreviewMerged openstack/openstack-ansible-os_cinder master: Add configuration option for heartbeat_in_pthread  https://review.opendev.org/c/openstack/openstack-ansible-os_cinder/+/83323817:28
opendevreviewMerged openstack/openstack-ansible-os_nova master: Add configuration option for heartbeat_in_pthread  https://review.opendev.org/c/openstack/openstack-ansible-os_nova/+/83323617:30
*** arxcruz is now known as arxcruz|off17:55
tbarronspatel: noonedeadpunk was reading backlog, don't see gokhani here now but manila has a zfs-on-linux driver.  dunno if that'w what gokhani has or if it's oracle/solaris zfs.19:05
noonedeadpunkI guess it was linux yeah...19:05
spateltbarron he left for the day. 19:06
noonedeadpunkbut dunno if he was looking for shared filesystem, felt like more for glance and cinder which is quite different layer anyway19:06
noonedeadpunkbut actually I didn't know manila does support zfs!19:06
tbarrongot it, ty.  There zfs via nfs seems dicey as you say.19:07
spatelif we want to share ZFS then we need ganesha-nfs or iscsi protocol 19:07
tbarronhere's the manila zfs on linux driver doc: https://opendev.org/openstack/manila/src/branch/master/doc/source/admin/zfs_on_linux_driver.rst19:09
spatelso its NFS last mile correct?19:12
tbarronspatel: but not *ganesha* gatewayed.  Others in #openstack-manila have done more with it than I have.  But I think to share even native oracle zfs you use NFS or SMB, right?19:19
spatelI don't have much experience with ZFS but all i know we need server protocol to export filesystem (CIFS/SMB or NFS) 19:20
spateldoes ZFS has own client server protocol ?19:21
tbarronit leverages nfs or smb for network file sharing; manila driver uses nfs for zfs-backed shares (smb wasn't implemented for zfs for manila)19:26
tbarronso thereis no separate zfs-on-the wire client server protocol, but in theory19:27
tbarronyou get cool back end features like replication, zfs snapshots, etc.19:28
spatelnice!! 19:49
spateldoes ZFS has clustering feature? not DAS based which is hardware based cluster. 19:50
jrossernot really19:52
jrosserits a server plus as many block devices as you need, attached over nvme/sata/sas/FC/whatever19:54
spatel+120:01
jrosseri have a small zfs setup for cinder backup, so that the backups do not have a shared fate with the ceph cluster20:02
jrossernow if any users actually bother to use it is another thing......20:02
tbarronjrosser: since you mentioned ceph, I just did an OSA AIO for the first time. Used 20:10
tbarron"export SCENARIO='aio_manila'" before running bootstrap-aio.sh and20:11
tbarronit all seemed to "just work.".  Nice, not a very big footprint and I don't have to20:11
tbarronrebuild kolla containers to test a code change.20:12
tbarronjrosser: But I'm led to ask: the upstream ceph community is deprecating20:12
tbarronceph-ansible in favor of cephadm.20:13
tbarronDoes OSA have plans to adapt to cephadm for future Ceph deployments?20:13
jrosserthat scenario is exactly what we run when merging patches to our os_manila ansible role20:13
jrosserlike here https://review.opendev.org/c/openstack/openstack-ansible-os_manila/+/82760420:14
tbarroncool20:14
tbarronjrosser: manila (and devstack-plugin ceph, and tripleo) will be moving to the ceph20:15
jrossertbarron: do you mean ceph-ansible itself is deprecated20:15
tbarronorchestrator backed by cephadm20:15
jrosseror just its use by redhat in their stuff is switching to cephadm20:15
tbarronsinde yes, ceph-ansible itself is being deprecated, and since20:15
tbarronthe *upstream ceph community* will be only supporting new needed features like20:16
tbarronactive-active ganesha (ceph-nfs daemon) with the orchestrator20:16
tbarron(which is either backed by cephadm or by rook (for k8s).20:17
jrosserimho ceph orchestrator intersects quite heavily with what tools such as ansible are trying to acheive20:17
jrosserand so if you want your deployment defined by your ansible code that this is going to be tricky20:17
tbarronjrosser: note that I am not here as a Red Hat guy (am actually independent now) and am20:17
tbarronjust reporting.20:18
tbarronI liked my experience with OSA and manila and ceph and it would be cool if20:19
*** dviroel|ruck is now known as dviroel|ruck|brb20:19
tbarronOSA with manila and Ceph will be able to use the new stuff where Ganesha runs20:19
tbarronactive-active inside the Ceph cluster with its own "ingress" (HA proxy) and20:20
tbarronlooks to consumers like any other ceph daemon.20:20
tbarronjrosser: we can talk more about it later.  devstack-plugin-ceph changes to do this are in progress.20:21
tbarronThat plugin was pretty primitive, pre-ceph-ansible.20:22
tbarronBut it did via shell commands what ceph-ansible does.20:22
jrosseryeah, though i think we maybe come from a deployer-first presepective20:22
tbarronjrosser: Yeah, that's the right perspective for OSA!20:22
jrosserwe need battle hard, and safe not to burn your cluster when doing upgrades or scaling out20:23
jrosserceph-ansible has been excellent in that regard20:23
tbarronnote that I don't "have a dog in the fight" w.r.t. ceph-ansible vs cephadm, but20:23
tbarronI think my claim that upstream ceph community is, rightly or wrongly, dropping20:24
tbarronceph-ansible in favor of cephadm is correct.  Take it as a data claim to be checked.20:24
jrosserwe are also not wedded to ceph-ansible for production20:25
jrosserin fact most havey users of OSA do not use the inbuilt ceph stuff at all and prefer to separate the concerns20:25
jrosserwe provide hooks to integrate an external cluster20:25
tbarronThat makes perfect sense.  20:26
jrosserceph-ansible gives us a quick solution for CI, and also for people who want a one-stop solution20:26
tbarronWhen I worked with TripleO it was pretty much the same.  But when tripelO set20:26
jrosserbut that integration causes major stress at upgrade time as the requirements of ceph(-ansible) and a particular openstack release are not necessarily co-incident20:27
tbarronup openstack to deploy the ceph cluster instead of just referenciing an externally deployed cluster, we tried to modify the tripleo triggered deployment to do it by the current best practice and look20:27
tbarronthe same.20:27
jrosserbtw i don't have manila in my deployments, it's always seemed too hard to get real multitenancy and performance at the same time20:28
jrosserbut i would really like to20:28
tbarronjrosser: so pls. just take my remarks now as a "heads up".  If I'm correct then20:28
tbarronexternal deployments of Ceph will move in time from ceph-ansible to cephadm.20:29
tbarronAnd at some point it make sense for OSA CI/dev/test deployments of Ceph to do the same.20:29
tbarronjrosser: your point about *real* multitenancy and performance tradeoff is fair.20:30
jrosserunless there is a way to not be docker/podman based i think it is highly unlikely you would see it used in OSA20:31
tbarronThere was some anti-container sentiment within the Ceph community w.r.t. cephadm but20:32
tbarronso far I don't think it prevailed.20:32
opendevreviewMerged openstack/openstack-ansible-lxc_hosts master: Change location of ipaddr filter  https://review.opendev.org/c/openstack/openstack-ansible-lxc_hosts/+/83311920:33
tbarronMy *rough* take on cephadm is that it aims to give one k8s-like declarative orchestration of a ceph cluster w/o actually using k8s.20:34
jrosseri was just reading the docs yes and it looks very much like that20:36
tbarronw.r.t. the multi-tenancy/performance tradeoff, cephadm *promises* to support 20:40
tbarrondynamically spawning new nfs clusters (each a set of active-active nfs ganesha daemons)20:41
tbarronwhere each cluster can have its own ingress (HA proxy implemented, at least for now).20:42
tbarronSo in theory manila could trigger these per-tenant, and set up per-tenant networking.20:42
tbarronnfs would still be in the data path from native CephFS20:43
jrosserhow would you see that networking work?20:43
tbarronbut there wouldn't be a single bottleneck.20:43
tbarronjrosser: I don't know if we'd need to take advantage of the BGP stuff now getting20:45
tbarronOpenStack support or not.  Likely manila would need to inject return routes to the20:45
tbarronclient network.  So even if the ceph-nfs ingress is pingable from everywhere20:46
tbarron(and firewall/security rules could be set up by manila to restrict) 20:46
tbarronpackets could only return route in tenant-appropriate manner.20:47
tbarronAdmittedly  I am waving my hands.  Originally the vision with Sage et. al.20:47
tbarronwas to have k8s-managed ceph clusters and to use kuryr to tie the NFS ingresses20:48
jrosserso far the only thing that doesnt make me go ewwwwwww is future support for virtiofs20:48
tbarronback to tenant private neutron networks.  So the goal would be something like that.20:48
jrosseras an example my storage network doesnt route anywhere20:48
jrosserabsolutely nowhere near the tenants20:48
tbarronyeah that's the way I did it for the ceph daemons proper.  But not for nfs :) and yeah,20:49
tbarronI get it, and that's why I pitched virtiofs to nova.20:49
jrosseri have another use case for that to pass intel RAPL power usage data into virtual machines20:50
tbarronThey think they'll have something for Zed.20:50
tbarronBut for bare metal compute instances we'd still need something like CephFS w/o20:50
tbarronvirtiofs.20:50
jrosserwe wrote an article for superuser here https://superuser.openstack.org/articles/environmental-reporting-dashboards-for-openstack-from-bbc-rd/20:51
jrosserthat describes a use case where having a static virtiofs mount from the hypervisor into the VM would allow per process energy accounting20:51
tbarronI hadn't seen that artticle, thank you!20:52
jrosserhttps://github.com/hubblo-org/scaphandre/issues/6020:54
tbarronawesome20:55
*** dviroel|ruck|brb is now known as dviroel|ruck20:57
spateljrosser is virtiofs ready for prod, i think not 21:00
jrosseri do not think so21:00
spatelit would be awesome when ever its ready 21:02
spateltbarron blog out my manila+glusterfs setup here https://satishdotpatel.github.io/openstack-manila-integration-with-glusterfs/21:02
spatelnow i need to test how i can provide manila SRIOV interface for dedicated NFS mount point 21:03
spateltbarron by the way thank for the manila patch 21:05
*** dviroel|ruck is now known as dviroel|ruck|afk21:51
*** dviroel|ruck|afk is now known as dviroel|ruck23:55

Generated by irclog2html.py 2.17.3 by Marius Gedminas - find it at https://mg.pov.lt/irclog2html/!