Monday, 2023-02-27

noonedeadpunkI will be semi-around today just in case09:54
jrossergood morning10:12
noonedeadpunko/10:30
noonedeadpunklet's try to land backports as would be great to make another releases for stable branches https://review.opendev.org/q/parentproject:openstack/openstack-ansible+branch:%255Estable/.*+status:open+10:34
jrosserthats the ones for Z on the way10:42
jrosserthough it looks like one of the infra mirrors might be broken so we could be in for some rechecks10:43
opendevreviewDmitriy Rabotyagov proposed openstack/openstack-ansible-galera_server master: Upgrade MariaDB to 10.11  https://review.opendev.org/c/openstack/openstack-ansible-galera_server/+/87538110:46
noonedeadpunkI'm gonna spawn some sandbox to check out haproxy changes....10:47
noonedeadpunkAs I don't quite understand some things still10:47
jrosserthere is something about handlers that i want to understand10:47
noonedeadpunkyeah10:48
jrosserand i was wondering if actually the problem there was we need to make the hander names more specific in the PKI role10:48
jrosseri think you can `notify: "{{ handler_name }}"`10:48
jrosserso we could have `glance cert installed` rather than just `cert installed` but tbh i was not even understading the underlying issue at all right now10:49
noonedeadpunkyeah, I wanted to try out couple of things as it's tough to understand just looking at code10:51
jrosserright - i am really struggling to find time to make AIO/understand which is why my review of this has been so tough to do10:52
jrosserthere was also a bug for LE which needed a throttle: 1 that should be called out as it's own patch10:52
noonedeadpunkyeah, kind of same here... but will invest time today at least to spawn env and can give you access it to play around if you want to10:54
noonedeadpunkBut yes, I was going to check these things as well10:55
jrosseri was sure that we had some other role that allowed a list of handler names to be passed in10:59
jrosserbut not found that obviously10:59
Elnaz^ Hi11:00
jrossernoonedeadpunk: i was also thinking that `add_host` could be used to make temporary group for running a play against all the haproxy hosts whilst having them access the group vars from some other service group11:04
jrosserthat would be really neat and take away some of the un-ansible-ness of whats there now11:05
noonedeadpunkOh, yes, that's really good idea. We indeed have temp groups in couple of places as of today11:06
jrosseryes we do, like lxc stuff11:06
jrosserthen it could be just an extra play in the service playbooks and super obvious how it works11:07
noonedeadpunk++11:08
opendevreviewDmitriy Rabotyagov proposed openstack/openstack-ansible-plugins master: Allow to manage more the one vhost with mq_setup  https://review.opendev.org/c/openstack/openstack-ansible-plugins/+/87539911:13
opendevreviewDmitriy Rabotyagov proposed openstack/openstack-ansible master: Replace HA policies for RabbitMQ with quorum  https://review.opendev.org/c/openstack/openstack-ansible/+/87361811:25
opendevreviewDmitriy Rabotyagov proposed openstack/openstack-ansible-os_cinder master: Add quorum support for service  https://review.opendev.org/c/openstack/openstack-ansible-os_cinder/+/87540811:26
*** dviroel_ is now known as dviroel11:28
opendevreviewDmitriy Rabotyagov proposed openstack/openstack-ansible-os_glance master: Add quorum support for glance  https://review.opendev.org/c/openstack/openstack-ansible-os_glance/+/87363211:30
noonedeadpunkjrosser: btw, have you checked for fixed haproxy packages? As  I still don't see them, but maybe I'm blind?12:26
jrosseri think they were available for focal pretty quick when i looked12:26
jrosseri did need to check the ubuntu changelog for the package though12:27
noonedeadpunkhm....12:27
noonedeadpunkIt was just mentioning 2.0.31  but might be indeed that just backported patch to 2.0.29 for some reason....12:29
noonedeadpunkThough they used to bump minor version quite freely...12:29
noonedeadpunkAh, yes... It's indeed 2.0.29-0ubuntu1.312:31
opendevreviewMerged openstack/openstack-ansible-os_nova stable/xena: Add authentication for [cinder] section of nova.conf  https://review.opendev.org/c/openstack/openstack-ansible-os_nova/+/87357112:41
opendevreviewMerged openstack/openstack-ansible-os_nova stable/zed: Add authentication for [cinder] section of nova.conf  https://review.opendev.org/c/openstack/openstack-ansible-os_nova/+/87346912:41
opendevreviewMerged openstack/openstack-ansible-os_nova stable/zed: Use SSL database connections with nova-manage  https://review.opendev.org/c/openstack/openstack-ansible-os_nova/+/87346612:41
*** lowercase_ is now known as lowercase15:29
lowercaseHey guys, im having an issue where the playbook keeps returning this error: "sqlalchemy.exc.OperationalError: (pymysql.err.OperationalError) (1091, \"Can't DROP COLUMN `vm_state`; check that it exists\")", "[SQL: ALTER TABLE build_requests DROP COLUMN vm_state]". The table, build_requests does not exist. and continues to cause api sync to fail. I confirmed this by logging into the db and confirming it indeed doesn't exist.15:31
jrosserlowercase: which playbook is this?15:32
lowercasenova15:32
lowercase"/openstack/venvs/nova-25.2.1.dev3/bin/nova-manage", "api_db", "sync"] is the ocmmand being run.15:32
lowercaseim on 25.3.015:33
jrosserlowercase: this is the patch in qestion - part of nova https://review.opendev.org/c/openstack/nova/+/81214515:42
lowercasethat's it.15:46
jrosserlowercase: this really is a nova issue tbh - i wonder if there is some error with the version of alembic/sqlalchemy causing issues here15:47
lowercaseim looking at seeing if there is a better way to write this so it tests if the table and column exists, before attempting to remove it15:47
lowercasealright, ill ping them15:47
jrossernote that this is the first migration that is not a "legacy" migration too15:48
jrosserso there could easily be a gotcha there15:48
lowercasewhat?15:48
lowercasewhat do you mean by legacy, sorry?15:48
jrosserhere https://github.com/openstack/nova/tree/master/nova/db/api/migrations/versions15:49
jrosservs here https://github.com/openstack/nova/tree/master/nova/db/api/legacy_migrations/versions15:49
jrosserbut tbh this is the extent of my understanding of the db migrations, knowing that they exist :)15:49
lowercasesame.. same..15:50
jrosseri think nova-manage puts stuff in syslog15:50
jrosserpossibly there is more there, and i believe that there are debug options too15:51
lowercase Running upgrade d67eeaabee36 -> b30f573d3377, Remove unused build_requests columns15:51
lowercasethat's all it says sadly15:51
lowercaseand then it spits out the python error i linked earlier15:52
jrosserwell at least that matches up with the version numbers in the migrations/ folder15:52
jrosseris this a previous install, then upgrade to 25.x, now minor upgrade to 25.y ?15:52
jrosserit would be interesting to know if in a fresh AIO running `nova-manage api_db sync` more than once runs into this issue15:54
lowercaseit does15:58
lowercasei dropped the whole nova database, and reran the api_db sync15:59
jrosserso delete db -> nova-manage api_db sync (works) -> nova-manage api_db sync (fails)16:04
jrosseri wonder what `nova-manage api_db version` says during those steps16:06
lowercaseapi_db sync fails with the same error16:20
lowercaseRunning upgrade d67eeaabee36 -> b30f573d3377, Remove unused build_requests columns16:20
jrosserlowercase: did you do W->Y upgrade, missing out X? i expect nova to say thats not supported16:36
lowercaseno, that was me misremebering version names16:37
lowercasewe are going x -> y16:37
lowercasehttps://bugs.launchpad.net/nova/+bug/200871616:38
lowercaseI just created that to see where it goes, in the meantime im going to start writing a patch16:38
jrosseri would edit that to avoid a distraction on skipping versions16:39
lowercasebut first gym and lunch16:39
lowercasei just edited it16:39
jrossercool16:39
lowercaserefresh just to confirm i did it right?16:39
jrosseryes i see it correct now16:40
spatelis this SSD good for ceph ? - https://www.dell.com/en-us/shop/dell-384tb-ssd-sas-mixed-use-12gbps-fips-140-512e-25in-pm5-v-3-dwpd-21024-tbw/apd/400-beoy/storage-drives-media17:04
jrosserspatel: seems pretty expensive :) and thats an older toshiba from before they rebrand to kioxia?17:09
spatelDoes it good for ceph.. ?17:10
jrosserit's impossible to say17:10
spatelhmm17:10
jrosserfirst it doesnt say if it has PLP17:10
spatelwhat is PLP?17:10
jrosserand it looks like "enterprise" type drive rather than scale out type17:10
jrosserPower Loss Protection17:11
spatelIf its enterprise then PLP must be there.. 17:11
spatelbased on cost17:11
jrosserthe only SSD drives that are good for ceph are those with sufficient power storage on the SSD to flush the RAM cache to the flash when the power fails17:12
jrosserand then depending on use case you might not need anywhere near 3 DWPD17:12
opendevreviewMerged openstack/openstack-ansible-os_nova stable/yoga: Use SSL database connections with nova-manage  https://review.opendev.org/c/openstack/openstack-ansible-os_nova/+/87346717:13
jrosserlike a 12Tb kioxia CD6 is ~2000eur17:15
opendevreviewMerged openstack/openstack-ansible-os_neutron stable/zed: Fix typo in ansible_facts['pkg_mgr']  https://review.opendev.org/c/openstack/openstack-ansible-os_neutron/+/87474217:58
opendevreviewMerged openstack/ansible-role-python_venv_build stable/zed: Drop empty elements from constraint/requirement files  https://review.opendev.org/c/openstack/ansible-role-python_venv_build/+/87473417:58
spateljrosser is Kioxia is good brand 17:58
spatelnever heard of it 17:58
lowercaseour considerations for drives are: does dell certify the drive? And price. 11k is way to high. We use microns mostly. we are targeting nvme's for our next cluster.19:02
lowercasei don't have any experience with fips drives tho so that's a completly different animal19:03
spatelWe have bunch of drives in stock spare.. thinking to use for ceph but not sure if they are good or not. I have to run some stuff to find good result 19:07
opendevreviewMerged openstack/openstack-ansible-os_nova stable/xena: Use SSL database connections with nova-manage  https://review.opendev.org/c/openstack/openstack-ansible-os_nova/+/87346819:35
ElnazHi19:36
ElnazDo you log this channel somewhere?19:37
ElnazBy 'Access` channel on the provider network, it didn't work in VLAN mode. Why it needs to have trunk channel between the nodes and TOR switch?19:37
-opendevstatus- NOTICE: The Gerrit service on review.opendev.org experienced severe performance degradation between 17:50 and 19:45 due to excessive API query activity; the addresses involved are now blocked but any changes missing job results from that timeframe should be rechecked19:54
supamattspatel: koxia is a good brand, cm6's are beasts20:06
supamattthe new cm8's are just gonzo fast20:06
spatelsupamatt thanks for your input20:07
supamattthe cm8 is hands down the fastest NVMe on the market today20:07
supamattit's u.3, gets 2Miops+ on reads, and writes are around 1.2Miops... Ceph though will never see that performnace. 20:08
supamatttoo much overhead20:08
supamattany enterprise NVMe that can do 200Kiops on reads, or 80Kiops on writes is going to be fast enough for Ceph today. until crismon goes ga20:09
BobZannapolisAnyone know how to make Rally's  "NovaServers.boot_and_delete_server" execute on EVERY hypervisor we have so we can verify that all compute nodes are being accessed ? tia 20:39

Generated by irclog2html.py 2.17.3 by Marius Gedminas - find it at https://mg.pov.lt/irclog2html/!