)]}'
{"id":"openstack%2Foslo.cache~742193","triplet_id":"openstack%2Foslo.cache~master~I8e6826bfb2c85e7ceed03e1667bd6a06b3dff469","project":"openstack/oslo.cache","branch":"master","topic":"bug/1888394","hashtags":[],"change_id":"I8e6826bfb2c85e7ceed03e1667bd6a06b3dff469","subject":"Do not hardcode flush_on_reconnect, move to oslo.cache config","status":"MERGED","created":"2020-07-21 13:42:33.000000000","updated":"2021-02-03 12:29:10.000000000","submitted":"2021-02-03 12:27:25.000000000","submitter":{"_account_id":22348,"name":"Zuul","username":"zuul","tags":["SERVICE_USER"]},"total_comment_count":38,"unresolved_comment_count":4,"has_review_started":true,"submission_id":"742193-bug/1888394","meta_rev_id":"337dc0cb7ec762edf417160f1a992c7900b37470","_number":742193,"virtual_id_number":742193,"owner":{"_account_id":27339,"name":"Michal Arbet","email":"michal.arbet@ultimum.io","username":"michalarbet"},"actions":{},"labels":{"Verified":{"approved":{"_account_id":22348,"name":"Zuul","username":"zuul","tags":["SERVICE_USER"]},"all":[{"tag":"autogenerated:zuul:gate","value":2,"date":"2021-02-03 12:27:24.000000000","permitted_voting_range":{"min":2,"max":2},"_account_id":22348,"name":"Zuul","username":"zuul","tags":["SERVICE_USER"]},{"value":0,"_account_id":32428,"name":"Tony Liu","email":"tonyliu0592@hotmail.com"},{"value":0,"_account_id":22072,"name":"Viktor Křivák","email":"viktor.krivak@ultimum.io","username":"viktor-krivak"},{"value":0,"_account_id":28522,"name":"Hervé Beraud","email":"herveberaud.pro@gmail.com","username":"hberaud"},{"value":0,"_account_id":27339,"name":"Michal Arbet","email":"michal.arbet@ultimum.io","username":"michalarbet"},{"value":0,"_account_id":5046,"name":"Lance Bragstad","email":"lbragstad@redhat.com","username":"ldbragst"},{"value":0,"_account_id":15334,"name":"Stephen Finucane","display_name":"stephenfin","email":"stephenfin@redhat.com","username":"sfinucan"},{"value":0,"_account_id":2903,"name":"Morgan Fainberg","email":"morgan.fainberg@gmail.com","username":"mdrnstm"},{"value":0,"_account_id":6928,"name":"Ben Nemec","email":"openstack@nemebean.com","username":"bnemec"},{"value":0,"_account_id":11904,"name":"Sean McGinnis","email":"sean.mcginnis@gmail.com","username":"SeanM"},{"value":0,"_account_id":27954,"name":"Moisés Guimarães de Medeiros","email":"guimaraes@pm.me","username":"moguimar"},{"value":0,"_account_id":8482,"name":"Colleen Murphy","email":"colleen@gazlene.net","username":"krinkle"}],"values":{"-2":"Fails","-1":"Doesn\u0027t seem to work"," 0":"No score","+1":"Works for me","+2":"Verified"},"description":"","default_value":0,"optional":true},"Code-Review":{"approved":{"_account_id":15334,"name":"Stephen Finucane","display_name":"stephenfin","email":"stephenfin@redhat.com","username":"sfinucan"},"all":[{"value":0,"_account_id":22348,"name":"Zuul","username":"zuul","tags":["SERVICE_USER"]},{"value":0,"_account_id":32428,"name":"Tony Liu","email":"tonyliu0592@hotmail.com"},{"value":0,"_account_id":22072,"name":"Viktor Křivák","email":"viktor.krivak@ultimum.io","username":"viktor-krivak"},{"value":0,"_account_id":28522,"name":"Hervé Beraud","email":"herveberaud.pro@gmail.com","username":"hberaud"},{"value":0,"_account_id":27339,"name":"Michal Arbet","email":"michal.arbet@ultimum.io","username":"michalarbet"},{"value":0,"_account_id":5046,"name":"Lance Bragstad","email":"lbragstad@redhat.com","username":"ldbragst"},{"value":2,"date":"2021-02-03 11:21:37.000000000","permitted_voting_range":{"min":2,"max":2},"_account_id":15334,"name":"Stephen Finucane","display_name":"stephenfin","email":"stephenfin@redhat.com","username":"sfinucan"},{"value":0,"_account_id":2903,"name":"Morgan Fainberg","email":"morgan.fainberg@gmail.com","username":"mdrnstm"},{"value":2,"date":"2021-01-25 16:26:57.000000000","permitted_voting_range":{"min":2,"max":2},"_account_id":6928,"name":"Ben Nemec","email":"openstack@nemebean.com","username":"bnemec"},{"value":0,"_account_id":11904,"name":"Sean McGinnis","email":"sean.mcginnis@gmail.com","username":"SeanM"},{"value":2,"date":"2021-01-25 14:44:46.000000000","_account_id":27954,"name":"Moisés Guimarães de Medeiros","email":"guimaraes@pm.me","username":"moguimar"},{"value":0,"_account_id":8482,"name":"Colleen Murphy","email":"colleen@gazlene.net","username":"krinkle"}],"values":{"-2":"Do not merge","-1":"This patch needs further work before it can be merged"," 0":"No score","+1":"Looks good to me, but someone else must approve","+2":"Looks good to me (core reviewer)"},"description":"","default_value":0,"optional":true},"Workflow":{"approved":{"_account_id":15334,"name":"Stephen Finucane","display_name":"stephenfin","email":"stephenfin@redhat.com","username":"sfinucan"},"all":[{"value":0,"_account_id":22348,"name":"Zuul","username":"zuul","tags":["SERVICE_USER"]},{"value":0,"_account_id":32428,"name":"Tony Liu","email":"tonyliu0592@hotmail.com"},{"value":0,"_account_id":22072,"name":"Viktor Křivák","email":"viktor.krivak@ultimum.io","username":"viktor-krivak"},{"value":0,"_account_id":28522,"name":"Hervé Beraud","email":"herveberaud.pro@gmail.com","username":"hberaud"},{"value":0,"_account_id":27339,"name":"Michal Arbet","email":"michal.arbet@ultimum.io","username":"michalarbet"},{"value":0,"_account_id":5046,"name":"Lance Bragstad","email":"lbragstad@redhat.com","username":"ldbragst"},{"value":1,"date":"2021-02-03 11:21:37.000000000","permitted_voting_range":{"min":1,"max":1},"_account_id":15334,"name":"Stephen Finucane","display_name":"stephenfin","email":"stephenfin@redhat.com","username":"sfinucan"},{"value":0,"_account_id":2903,"name":"Morgan Fainberg","email":"morgan.fainberg@gmail.com","username":"mdrnstm"},{"value":0,"_account_id":6928,"name":"Ben Nemec","email":"openstack@nemebean.com","username":"bnemec"},{"value":0,"_account_id":11904,"name":"Sean McGinnis","email":"sean.mcginnis@gmail.com","username":"SeanM"},{"value":0,"_account_id":27954,"name":"Moisés Guimarães de Medeiros","email":"guimaraes@pm.me","username":"moguimar"},{"value":0,"_account_id":8482,"name":"Colleen Murphy","email":"colleen@gazlene.net","username":"krinkle"}],"values":{"-1":"Work in progress"," 0":"Ready for reviews","+1":"Approved"},"description":"","default_value":0,"optional":true},"Backport-Candidate":{"all":[{"value":0,"_account_id":22348,"name":"Zuul","username":"zuul","tags":["SERVICE_USER"]},{"value":0,"_account_id":32428,"name":"Tony Liu","email":"tonyliu0592@hotmail.com"},{"value":0,"_account_id":22072,"name":"Viktor Křivák","email":"viktor.krivak@ultimum.io","username":"viktor-krivak"},{"value":0,"_account_id":28522,"name":"Hervé Beraud","email":"herveberaud.pro@gmail.com","username":"hberaud"},{"value":0,"_account_id":27339,"name":"Michal Arbet","email":"michal.arbet@ultimum.io","username":"michalarbet"},{"value":0,"_account_id":5046,"name":"Lance Bragstad","email":"lbragstad@redhat.com","username":"ldbragst"},{"value":0,"_account_id":15334,"name":"Stephen Finucane","display_name":"stephenfin","email":"stephenfin@redhat.com","username":"sfinucan"},{"value":0,"_account_id":2903,"name":"Morgan Fainberg","email":"morgan.fainberg@gmail.com","username":"mdrnstm"},{"value":0,"_account_id":6928,"name":"Ben Nemec","email":"openstack@nemebean.com","username":"bnemec"},{"value":0,"_account_id":11904,"name":"Sean McGinnis","email":"sean.mcginnis@gmail.com","username":"SeanM"},{"value":0,"_account_id":27954,"name":"Moisés Guimarães de Medeiros","email":"guimaraes@pm.me","username":"moguimar"},{"value":0,"_account_id":8482,"name":"Colleen Murphy","email":"colleen@gazlene.net","username":"krinkle"}],"values":{"-2":"Do Not Backport","-1":"Not A Backport Candidate"," 0":"Backport Review Needed","+1":"Proposed Backport","+2":"Should Backport"},"description":"","default_value":0,"optional":true}},"removable_reviewers":[],"reviewers":{"CC":[{"_account_id":11816,"name":"mike_mp@zzzcomputing.com","display_name":"Mike Bayer","email":"mike_mp@zzzcomputing.com","username":"zzzeek","status":"Red Hat"}],"REVIEWER":[{"_account_id":2903,"name":"Morgan Fainberg","email":"morgan.fainberg@gmail.com","username":"mdrnstm"},{"_account_id":5046,"name":"Lance Bragstad","email":"lbragstad@redhat.com","username":"ldbragst"},{"_account_id":6928,"name":"Ben Nemec","email":"openstack@nemebean.com","username":"bnemec"},{"_account_id":8482,"name":"Colleen Murphy","email":"colleen@gazlene.net","username":"krinkle"},{"_account_id":11904,"name":"Sean McGinnis","email":"sean.mcginnis@gmail.com","username":"SeanM"},{"_account_id":15334,"name":"Stephen Finucane","display_name":"stephenfin","email":"stephenfin@redhat.com","username":"sfinucan"},{"_account_id":22072,"name":"Viktor Křivák","email":"viktor.krivak@ultimum.io","username":"viktor-krivak"},{"_account_id":22348,"name":"Zuul","username":"zuul","tags":["SERVICE_USER"]},{"_account_id":27339,"name":"Michal Arbet","email":"michal.arbet@ultimum.io","username":"michalarbet"},{"_account_id":27954,"name":"Moisés Guimarães de Medeiros","email":"guimaraes@pm.me","username":"moguimar"},{"_account_id":28522,"name":"Hervé Beraud","email":"herveberaud.pro@gmail.com","username":"hberaud"},{"_account_id":32428,"name":"Tony Liu","email":"tonyliu0592@hotmail.com"}]},"pending_reviewers":{},"reviewer_updates":[{"updated":"2020-07-23 12:39:38.000000000","updated_by":{"_account_id":2903,"name":"Morgan Fainberg","email":"morgan.fainberg@gmail.com","username":"mdrnstm"},"reviewer":{"_account_id":2903,"name":"Morgan Fainberg","email":"morgan.fainberg@gmail.com","username":"mdrnstm"},"state":"REVIEWER"},{"updated":"2020-07-23 12:39:50.000000000","updated_by":{"_account_id":8482,"name":"Colleen Murphy","email":"colleen@gazlene.net","username":"krinkle"},"reviewer":{"_account_id":8482,"name":"Colleen Murphy","email":"colleen@gazlene.net","username":"krinkle"},"state":"REVIEWER"},{"updated":"2020-07-27 13:25:27.000000000","updated_by":{"_account_id":28522,"name":"Hervé Beraud","email":"herveberaud.pro@gmail.com","username":"hberaud"},"reviewer":{"_account_id":28522,"name":"Hervé Beraud","email":"herveberaud.pro@gmail.com","username":"hberaud"},"state":"REVIEWER"},{"updated":"2020-07-27 15:20:13.000000000","updated_by":{"_account_id":27954,"name":"Moisés Guimarães de Medeiros","email":"guimaraes@pm.me","username":"moguimar"},"reviewer":{"_account_id":27954,"name":"Moisés Guimarães de Medeiros","email":"guimaraes@pm.me","username":"moguimar"},"state":"REVIEWER"},{"updated":"2020-07-27 15:27:26.000000000","updated_by":{"_account_id":11904,"name":"Sean McGinnis","email":"sean.mcginnis@gmail.com","username":"SeanM"},"reviewer":{"_account_id":11904,"name":"Sean McGinnis","email":"sean.mcginnis@gmail.com","username":"SeanM"},"state":"REVIEWER"},{"updated":"2020-07-27 18:21:07.000000000","updated_by":{"_account_id":22348,"name":"Zuul","username":"zuul","tags":["SERVICE_USER"]},"reviewer":{"_account_id":22348,"name":"Zuul","username":"zuul","tags":["SERVICE_USER"]},"state":"REVIEWER"},{"updated":"2020-07-27 20:29:52.000000000","updated_by":{"_account_id":5046,"name":"Lance Bragstad","email":"lbragstad@redhat.com","username":"ldbragst"},"reviewer":{"_account_id":5046,"name":"Lance Bragstad","email":"lbragstad@redhat.com","username":"ldbragst"},"state":"REVIEWER"},{"updated":"2020-07-28 14:31:02.000000000","updated_by":{"_account_id":22072,"name":"Viktor Křivák","email":"viktor.krivak@ultimum.io","username":"viktor-krivak"},"reviewer":{"_account_id":22072,"name":"Viktor Křivák","email":"viktor.krivak@ultimum.io","username":"viktor-krivak"},"state":"REVIEWER"},{"updated":"2020-08-10 15:41:48.000000000","updated_by":{"_account_id":6928,"name":"Ben Nemec","email":"openstack@nemebean.com","username":"bnemec"},"reviewer":{"_account_id":6928,"name":"Ben Nemec","email":"openstack@nemebean.com","username":"bnemec"},"state":"REVIEWER"},{"updated":"2020-09-16 15:49:25.000000000","updated_by":{"_account_id":32428,"name":"Tony Liu","email":"tonyliu0592@hotmail.com"},"reviewer":{"_account_id":32428,"name":"Tony Liu","email":"tonyliu0592@hotmail.com"},"state":"REVIEWER"},{"updated":"2021-01-28 14:08:06.000000000","updated_by":{"_account_id":11816,"name":"mike_mp@zzzcomputing.com","display_name":"Mike Bayer","email":"mike_mp@zzzcomputing.com","username":"zzzeek","status":"Red Hat"},"reviewer":{"_account_id":11816,"name":"mike_mp@zzzcomputing.com","display_name":"Mike Bayer","email":"mike_mp@zzzcomputing.com","username":"zzzeek","status":"Red Hat"},"state":"CC"},{"updated":"2021-02-03 11:21:37.000000000","updated_by":{"_account_id":15334,"name":"Stephen Finucane","display_name":"stephenfin","email":"stephenfin@redhat.com","username":"sfinucan"},"reviewer":{"_account_id":15334,"name":"Stephen Finucane","display_name":"stephenfin","email":"stephenfin@redhat.com","username":"sfinucan"},"state":"REVIEWER"}],"messages":[{"id":"678a3e6404b6f5a55fa79b2ec41fda677b19841c","author":{"_account_id":27339,"name":"Michal Arbet","email":"michal.arbet@ultimum.io","username":"michalarbet"},"date":"2020-07-21 13:42:33.000000000","message":"Uploaded patch set 1.","accounts_in_message":[],"_revision_number":1},{"id":"2c0e5a9e75ef99041a856d4288f75dfde81a5908","author":{"_account_id":28522,"name":"Hervé Beraud","email":"herveberaud.pro@gmail.com","username":"hberaud"},"date":"2020-07-21 13:43:35.000000000","message":"Patch Set 1:\n\nHello,\n\nPlease add a release note to highlight the config stuff and the behavior changes.","accounts_in_message":[],"_revision_number":1},{"id":"5e29ec5de3512dd26882a5bc554696196f246273","author":{"_account_id":28522,"name":"Hervé Beraud","email":"herveberaud.pro@gmail.com","username":"hberaud"},"date":"2020-07-21 13:47:34.000000000","message":"Patch Set 1: Code-Review-1\n\nhum I think some changes are needed, we need to declare this new config in our options (1], else I don\u0027t know how it will be retrieved.\n\n[1] https://github.com/openstack/oslo.cache/blob/master/oslo_cache/_opts.py","accounts_in_message":[],"_revision_number":1},{"id":"327fbe5e1478f4a3b4c31365fa3bacaa7cc6b569","author":{"_account_id":27339,"name":"Michal Arbet","email":"michal.arbet@ultimum.io","username":"michalarbet"},"date":"2020-07-21 13:51:40.000000000","message":"Patch Set 1:\n\n\u003e hum I think some changes are needed, we need to declare this new\n \u003e config in our options (1], else I don\u0027t know how it will be\n \u003e retrieved.\n \u003e \n \u003e [1] https://github.com/openstack/oslo.cache/blob/master/oslo_cache/_opts.py\n\nWell, firstly i wanted to add configuration option. But it is not straightforward as it might seem at first glance. Oslo.cache could have several backends , for example redis. Oslo.cache\u0027s option backend_argument is especially for this -\u003e to pass some option to backend which I currently using.","accounts_in_message":[],"_revision_number":1},{"id":"2640995c51eae7d11b2aecf2b7c15c956cf9ddc3","author":{"_account_id":28522,"name":"Hervé Beraud","email":"herveberaud.pro@gmail.com","username":"hberaud"},"date":"2020-07-21 13:52:00.000000000","message":"Patch Set 1:\n\n\u003e hum I think some changes are needed, we need to declare this new\n \u003e config in our options (1], else I don\u0027t know how it will be\n \u003e retrieved.\n \u003e \n \u003e [1] https://github.com/openstack/oslo.cache/blob/master/oslo_cache/_opts.py\n\nI mean it could be more useful to allow user to handle this params by using config like the other options, isn\u0027t?","accounts_in_message":[],"_revision_number":1},{"id":"c366e1323417dfcea99f6494ea3a62628aff8064","author":{"_account_id":28522,"name":"Hervé Beraud","email":"herveberaud.pro@gmail.com","username":"hberaud"},"date":"2020-07-21 13:57:03.000000000","message":"Patch Set 1:\n\n\u003e \u003e hum I think some changes are needed, we need to declare this new\n \u003e \u003e config in our options (1], else I don\u0027t know how it will be\n \u003e \u003e retrieved.\n \u003e \u003e\n \u003e \u003e [1] https://github.com/openstack/oslo.cache/blob/master/oslo_cache/_opts.py\n \u003e \n \u003e Well, firstly i wanted to add configuration option. But it is not\n \u003e straightforward as it might seem at first glance. Oslo.cache could\n \u003e have several backends , for example redis. Oslo.cache\u0027s option\n \u003e backend_argument is especially for this -\u003e to pass some option to\n \u003e backend which I currently using.\n\nYep I know, but you can take example on this code [1] to target a specific backend, if you follow it then it\u0027s not too hard either.\n\n[1] https://github.com/openstack/oslo.cache/blob/master/oslo_cache/core.py#L144,L147","accounts_in_message":[],"_revision_number":1},{"id":"79f41d3d2b9b3b2619658b2b7bec34686d8fd924","tag":"autogenerated:zuul:check","author":{"_account_id":22348,"name":"Zuul","username":"zuul","tags":["SERVICE_USER"]},"date":"2020-07-21 15:26:54.000000000","message":"Patch Set 1: Verified+1\n\nBuild succeeded (check pipeline).\n\n- tempest-full-py3 https://zuul.opendev.org/t/openstack/build/ad0608f74d404f2bbdaba1b2239efbfb : SUCCESS in 1h 35m 59s\n- openstack-tox-lower-constraints https://zuul.opendev.org/t/openstack/build/24d838c8290b474f938968e360e400d0 : SUCCESS in 5m 29s\n- openstack-tox-pep8 https://zuul.opendev.org/t/openstack/build/321c69d8177c4e508cb0bb0e5afb8425 : SUCCESS in 4m 29s\n- openstack-tox-py36 https://zuul.opendev.org/t/openstack/build/de619fc701864bfb96245a7b88ca3e65 : SUCCESS in 3m 47s\n- openstack-tox-py38 https://zuul.opendev.org/t/openstack/build/54f0525d18614a4394bc8dccc349963e : SUCCESS in 5m 31s\n- openstack-tox-docs https://zuul.opendev.org/t/openstack/build/8ce5df6b5b8d40d0b69137239d666da4 : SUCCESS in 5m 53s\n- oslo.cache-functional-tox-py38-etcd3gw https://zuul.opendev.org/t/openstack/build/501319afa2d94613a9a3f9c31d35ba45 : SUCCESS in 5m 20s","accounts_in_message":[],"_revision_number":1},{"id":"b9ff74bda9e6b9f1002010ba4fbbf47ef5f59fb0","author":{"_account_id":27339,"name":"Michal Arbet","email":"michal.arbet@ultimum.io","username":"michalarbet"},"date":"2020-07-21 15:45:47.000000000","message":"Uploaded patch set 2.","accounts_in_message":[],"_revision_number":2},{"id":"c0438b848f696edaa6f86904cd90c07cc74621ec","tag":"autogenerated:zuul:check","author":{"_account_id":22348,"name":"Zuul","username":"zuul","tags":["SERVICE_USER"]},"date":"2020-07-21 17:14:14.000000000","message":"Patch Set 2: Verified+1\n\nBuild succeeded (check pipeline).\n\n- tempest-full-py3 https://zuul.opendev.org/t/openstack/build/bbc87b070dfc4b0e8bb500447a6673b1 : SUCCESS in 1h 17m 10s\n- openstack-tox-lower-constraints https://zuul.opendev.org/t/openstack/build/cd91d1230c9d4eb89d0b37483a3bb4d9 : SUCCESS in 3m 38s\n- openstack-tox-pep8 https://zuul.opendev.org/t/openstack/build/f923404b1ef64562a6fa5afe7cce06ef : SUCCESS in 4m 01s\n- openstack-tox-py36 https://zuul.opendev.org/t/openstack/build/0fc9e5e8dc364887879f10be6425c3b3 : SUCCESS in 4m 12s\n- openstack-tox-py38 https://zuul.opendev.org/t/openstack/build/4fbd8f2d97dc425e8a4ba14bc755f854 : SUCCESS in 4m 01s\n- openstack-tox-docs https://zuul.opendev.org/t/openstack/build/ccb84ffbb13443868e014aef8d3e5fc0 : SUCCESS in 4m 05s\n- build-openstack-releasenotes https://zuul.opendev.org/t/openstack/build/55d8093a8d8b4e41ac1e5bf53ce4518b : SUCCESS in 5m 38s\n- oslo.cache-functional-tox-py38-etcd3gw https://zuul.opendev.org/t/openstack/build/0a3ee509b154484182b4dcd35a343e84 : SUCCESS in 7m 12s","accounts_in_message":[],"_revision_number":2},{"id":"6ef101f2c55b324618a6083ec17e7adcf0527c21","author":{"_account_id":28522,"name":"Hervé Beraud","email":"herveberaud.pro@gmail.com","username":"hberaud"},"date":"2020-07-22 09:59:31.000000000","message":"Patch Set 2:\n\n(2 comments)\n\nHello,\n\nFirst, thanks Michal for these changes and for your differents patch sets.\n\nI\u0027ve some concerns, see my inline comments.\n\nAlso I think we need to add some related tests, you could take example on Moises\u0027s latest changes [1] to introduce your tests.\n\n[1] https://github.com/openstack/oslo.cache/commit/996036acd7b1376f0a378a407697afa28e5b2861","accounts_in_message":[],"_revision_number":2},{"id":"7d11499f1003860a1ba98cd463ea969928a3d4ac","author":{"_account_id":27339,"name":"Michal Arbet","email":"michal.arbet@ultimum.io","username":"michalarbet"},"date":"2020-07-22 15:12:24.000000000","message":"Uploaded patch set 3.","accounts_in_message":[],"_revision_number":3},{"id":"2eb827e15ec8eb6298a2c486d98debfe6556331b","author":{"_account_id":28522,"name":"Hervé Beraud","email":"herveberaud.pro@gmail.com","username":"hberaud"},"date":"2020-07-22 15:23:53.000000000","message":"Patch Set 3: Code-Review-1\n\n(2 comments)","accounts_in_message":[],"_revision_number":3},{"id":"c8c17a7acaecde1b5ee13577e0b1a9205e8fe707","author":{"_account_id":27339,"name":"Michal Arbet","email":"michal.arbet@ultimum.io","username":"michalarbet"},"date":"2020-07-22 15:26:32.000000000","message":"Patch Set 3:\n\n(1 comment)","accounts_in_message":[],"_revision_number":3},{"id":"cb654ced8f9684c5d29fce4f2c50d00a81d98573","author":{"_account_id":27339,"name":"Michal Arbet","email":"michal.arbet@ultimum.io","username":"michalarbet"},"date":"2020-07-22 15:31:37.000000000","message":"Uploaded patch set 4.","accounts_in_message":[],"_revision_number":4},{"id":"450168e002ea9b69d54d16b16bf5f890c7676954","author":{"_account_id":27339,"name":"Michal Arbet","email":"michal.arbet@ultimum.io","username":"michalarbet"},"date":"2020-07-22 15:31:56.000000000","message":"Patch Set 4:\n\n(1 comment)","accounts_in_message":[],"_revision_number":4},{"id":"66f3bf1757dbf078fe25ea29ad4e474f969421cb","author":{"_account_id":28522,"name":"Hervé Beraud","email":"herveberaud.pro@gmail.com","username":"hberaud"},"date":"2020-07-22 16:01:07.000000000","message":"Patch Set 4: Code-Review-1\n\n(1 comment)","accounts_in_message":[],"_revision_number":4},{"id":"446d39df41d351f93a4845d486d615ead42f763a","author":{"_account_id":28522,"name":"Hervé Beraud","email":"herveberaud.pro@gmail.com","username":"hberaud"},"date":"2020-07-22 16:01:48.000000000","message":"Patch Set 4:\n\n(1 comment)","accounts_in_message":[],"_revision_number":4},{"id":"9d0d56723bf6e028648fefcaa301ebef907ed013","author":{"_account_id":28522,"name":"Hervé Beraud","email":"herveberaud.pro@gmail.com","username":"hberaud"},"date":"2020-07-22 16:16:09.000000000","message":"Patch Set 4:\n\n(1 comment)","accounts_in_message":[],"_revision_number":4},{"id":"4436b0e45f9845a5f452877c815ddebf9a4f33aa","author":{"_account_id":28522,"name":"Hervé Beraud","email":"herveberaud.pro@gmail.com","username":"hberaud"},"date":"2020-07-22 16:18:22.000000000","message":"Patch Set 4:\n\n(1 comment)","accounts_in_message":[],"_revision_number":4},{"id":"615d2ee856abcfbcba8b6285a26d785bb1560dd2","author":{"_account_id":28522,"name":"Hervé Beraud","email":"herveberaud.pro@gmail.com","username":"hberaud"},"date":"2020-07-22 16:23:54.000000000","message":"Patch Set 4:\n\n(1 comment)","accounts_in_message":[],"_revision_number":4},{"id":"b70e6bfe00eb263b490d4e27324ef5c8609cbc16","author":{"_account_id":28522,"name":"Hervé Beraud","email":"herveberaud.pro@gmail.com","username":"hberaud"},"date":"2020-07-22 16:29:12.000000000","message":"Patch Set 4:\n\nmy -1 is related to the typo","accounts_in_message":[],"_revision_number":4},{"id":"c6c8910c7956fd8a6294128489d409dce8b58b82","author":{"_account_id":27339,"name":"Michal Arbet","email":"michal.arbet@ultimum.io","username":"michalarbet"},"date":"2020-07-22 16:57:38.000000000","message":"Patch Set 5: Published edit on patch set 4.","accounts_in_message":[],"_revision_number":5},{"id":"597284a5085b4af186d29d2e82f8f190c53f605d","author":{"_account_id":28522,"name":"Hervé Beraud","email":"herveberaud.pro@gmail.com","username":"hberaud"},"date":"2020-07-22 17:10:34.000000000","message":"Patch Set 5: Backport-Candidate-1 Code-Review+1\n\nLGTM, +1 for now by waiting for CI","accounts_in_message":[],"_revision_number":5},{"id":"476b0b78f46f0a8b8d3ee2fca2c2b2c153d20816","tag":"autogenerated:zuul:check","author":{"_account_id":22348,"name":"Zuul","username":"zuul","tags":["SERVICE_USER"]},"date":"2020-07-22 18:31:22.000000000","message":"Patch Set 5: Verified+1\n\nBuild succeeded (check pipeline).\n\n- tempest-full-py3 https://zuul.opendev.org/t/openstack/build/ec2d7b52bdcf4368b1812f339c6a92ab : SUCCESS in 1h 26m 43s\n- openstack-tox-lower-constraints https://zuul.opendev.org/t/openstack/build/e8294563906348a3817192527625967c : SUCCESS in 4m 30s\n- openstack-tox-pep8 https://zuul.opendev.org/t/openstack/build/b68a392c658e4f7c9e69a22cf06d28d8 : SUCCESS in 3m 43s\n- openstack-tox-py36 https://zuul.opendev.org/t/openstack/build/fc60be792e77470483d8a45c04e4e556 : SUCCESS in 4m 01s\n- openstack-tox-py38 https://zuul.opendev.org/t/openstack/build/c2f9cfdb3ec54bd88fb14b8e7e2e4d1a : SUCCESS in 4m 12s\n- openstack-tox-docs https://zuul.opendev.org/t/openstack/build/0b54e8bf57594594b32c28cfa6f3dd39 : SUCCESS in 4m 05s\n- build-openstack-releasenotes https://zuul.opendev.org/t/openstack/build/af89b54cc8534dfeaa3ed1de34e6775a : SUCCESS in 4m 46s\n- oslo.cache-functional-tox-py38-etcd3gw https://zuul.opendev.org/t/openstack/build/63fd189b10bb4ff5aafbb05530acc686 : SUCCESS in 4m 36s","accounts_in_message":[],"_revision_number":5},{"id":"35b1f58fb67c735469f7b76ac0fc48f7aedfd213","author":{"_account_id":27954,"name":"Moisés Guimarães de Medeiros","email":"guimaraes@pm.me","username":"moguimar"},"date":"2020-07-23 08:21:07.000000000","message":"Patch Set 5: Code-Review+1\n\nThanks for the patches, Michal. I\u0027m a bit over the fence with the True/False choice for default value as well.\n\nPoints to consider are:\n\nHow long has this behavior been like this? (cycles/releases)\n\nAren\u0027t we introducing behavior that is not expected for services that are not having the issue?\n\nThis seems to be the last thing to decide, we could bring it up next Monday for the team meeting.","accounts_in_message":[],"_revision_number":5},{"id":"3897bfc854c09395283288f6fe9ffd9faf79a6e2","author":{"_account_id":27339,"name":"Michal Arbet","email":"michal.arbet@ultimum.io","username":"michalarbet"},"date":"2020-07-23 10:50:20.000000000","message":"Patch Set 5:\n\n\u003e Thanks for the patches, Michal. I\u0027m a bit over the fence with the\n \u003e True/False choice for default value as well.\n \u003e \n \u003e Points to consider are:\n \u003e \n \u003e How long has this behavior been like this? (cycles/releases)\n \u003e \n \u003e Aren\u0027t we introducing behavior that is not expected for services\n \u003e that are not having the issue?\n \u003e \n \u003e This seems to be the last thing to decide, we could bring it up\n \u003e next Monday for the team meeting.\n\nHow long has this behavior been like this?\n\u003e\u003e From beginning of oslo.cache option flush_on_reconnect is not defined in a code ..so default value was always used from python-memcache library where it is set to False. So, all releases, cycles was using flush_on_reconnect \u003d False. Unfortunatelly , change of flush_on_reconnect \u003d True was backported also to queens/rocky/stein\n\nAren\u0027t we introducing behavior that is not expected for services\nthat are not having the issue?\n\u003e\u003e New behaviour was introduced in review https://review.opendev.org/#/c/644774/ where flush_on_reconnect was changed to True. This review is only reverting option to default(False) as it was always, but respects previous change and moving this option to oslo.cache config.","accounts_in_message":[],"_revision_number":5},{"id":"cc12855149777746951fc06f7cdba17a280829a4","author":{"_account_id":27339,"name":"Michal Arbet","email":"michal.arbet@ultimum.io","username":"michalarbet"},"date":"2020-07-23 10:52:43.000000000","message":"Patch Set 5:\n\nBtw, I see benefit from this option only when used with keystone, other services can have flush_on_reconnect \u003d False","accounts_in_message":[],"_revision_number":5},{"id":"909a645586396b0c219b563800b1f299256b1374","author":{"_account_id":27954,"name":"Moisés Guimarães de Medeiros","email":"guimaraes@pm.me","username":"moguimar"},"date":"2020-07-23 12:41:30.000000000","message":"Patch Set 5:\n\nAdded @Morgan and @Colleen to reviewers for feedback until Monday.\n\nWe might also need a patch somewhere in Keystone/TripleO setting the option to True.","accounts_in_message":[],"_revision_number":5},{"id":"d6c3fc129dec738b6a7a22d6d788b982040b779e","author":{"_account_id":28522,"name":"Hervé Beraud","email":"herveberaud.pro@gmail.com","username":"hberaud"},"date":"2020-07-24 10:51:39.000000000","message":"Patch Set 5:\n\n\u003e Added @Morgan and @Colleen to reviewers for feedback until Monday.\n \u003e \n \u003e We might also need a patch somewhere in Keystone/TripleO setting\n \u003e the option to True.\n\nGood point!\n\nI agree it could be worth to grab some feedbacks from Morgan and Coleen too.\n\n+1 to discuss a bit about this during the next meeting.","accounts_in_message":[],"_revision_number":5},{"id":"8ba149ac371be76a0528370540ea072e121ed792","author":{"_account_id":27339,"name":"Michal Arbet","email":"michal.arbet@ultimum.io","username":"michalarbet"},"date":"2020-07-27 11:02:21.000000000","message":"Uploaded patch set 6.","accounts_in_message":[],"_revision_number":6},{"id":"4ea765c859a598f22ddaac1007cfb61bf7307360","author":{"_account_id":28522,"name":"Hervé Beraud","email":"herveberaud.pro@gmail.com","username":"hberaud"},"date":"2020-07-27 11:23:17.000000000","message":"Patch Set 6: Code-Review-1\n\n(1 comment)","accounts_in_message":[],"_revision_number":6},{"id":"1ac3bebde8734c759e9f3af30e2d4f2d5151f55f","author":{"_account_id":27339,"name":"Michal Arbet","email":"michal.arbet@ultimum.io","username":"michalarbet"},"date":"2020-07-27 11:42:24.000000000","message":"Uploaded patch set 7.","accounts_in_message":[],"_revision_number":7},{"id":"7fc08b52b037224e40c50e23a5b626180e497b33","author":{"_account_id":27339,"name":"Michal Arbet","email":"michal.arbet@ultimum.io","username":"michalarbet"},"date":"2020-07-27 12:21:17.000000000","message":"Patch Set 7:\n\n(1 comment)","accounts_in_message":[],"_revision_number":7},{"id":"695203d75945da78dfa48c9768fc7afe21c8fdeb","tag":"autogenerated:zuul:check","author":{"_account_id":22348,"name":"Zuul","username":"zuul","tags":["SERVICE_USER"]},"date":"2020-07-27 12:47:19.000000000","message":"Patch Set 7: Verified+1\n\nBuild succeeded (check pipeline).\n\n- tempest-full-py3 https://zuul.opendev.org/t/openstack/build/9b9cb990ee984a829a2de72a35fe358e : SUCCESS in 56m 54s\n- openstack-tox-lower-constraints https://zuul.opendev.org/t/openstack/build/ac4e12b780534315b9635c20f6c1fb42 : SUCCESS in 6m 05s\n- openstack-tox-pep8 https://zuul.opendev.org/t/openstack/build/2d55f06cda234bccb01da687d40987ed : SUCCESS in 5m 23s\n- openstack-tox-py36 https://zuul.opendev.org/t/openstack/build/701bc58813854b4288505cc02a2146b3 : SUCCESS in 5m 26s\n- openstack-tox-py38 https://zuul.opendev.org/t/openstack/build/02f6ba80c4894f7894f94796984adfee : SUCCESS in 6m 06s\n- openstack-tox-docs https://zuul.opendev.org/t/openstack/build/254e465a36b9487a8cf0e7fb75c2b033 : SUCCESS in 5m 27s\n- build-openstack-releasenotes https://zuul.opendev.org/t/openstack/build/d64326a51cb64d3fb97e1760fddef433 : SUCCESS in 6m 04s\n- oslo.cache-functional-tox-py38-etcd3gw https://zuul.opendev.org/t/openstack/build/0e17cc2dfeed4e81a27406c0ab2d7e8a : SUCCESS in 6m 05s","accounts_in_message":[],"_revision_number":7},{"id":"6cbc7ec11cf33b2c54e6c2332b09b3497c39e325","author":{"_account_id":28522,"name":"Hervé Beraud","email":"herveberaud.pro@gmail.com","username":"hberaud"},"date":"2020-07-27 13:25:27.000000000","message":"Patch Set 7: Code-Review+1\n\n+1 for now, waiting for discussion and more feedbacks.","accounts_in_message":[],"_revision_number":7},{"id":"ec59ef49289224226a62da120d79c2eb70549f62","author":{"_account_id":27954,"name":"Moisés Guimarães de Medeiros","email":"guimaraes@pm.me","username":"moguimar"},"date":"2020-07-27 15:20:07.000000000","message":"Patch Set 7: Backport-Candidate+1","accounts_in_message":[],"_revision_number":7},{"id":"433bb8c0ea7de4fd63a827d8760fd6ee599f73e0","author":{"_account_id":27954,"name":"Moisés Guimarães de Medeiros","email":"guimaraes@pm.me","username":"moguimar"},"date":"2020-07-27 15:20:13.000000000","message":"Patch Set 7: -Backport-Candidate Code-Review+1","accounts_in_message":[],"_revision_number":7},{"id":"a1a4c7db18f53bad3c9f9cb37539dfffbeb848c9","author":{"_account_id":11904,"name":"Sean McGinnis","email":"sean.mcginnis@gmail.com","username":"SeanM"},"date":"2020-07-27 15:27:26.000000000","message":"Patch Set 7:\n\n(2 comments)","accounts_in_message":[],"_revision_number":7},{"id":"6c1a9c0c4085dccb656d67d1d8222cf9cb583dbb","author":{"_account_id":6928,"name":"Ben Nemec","email":"openstack@nemebean.com","username":"bnemec"},"date":"2020-07-27 15:28:12.000000000","message":"Patch Set 7:\n\nIf we need a different default for keystone, then my suggestion would be to add a set_defaults like we have here: https://opendev.org/openstack/oslo.policy/src/commit/6296fce9f8e3205506f87dd31d0b18f4b7842000/oslo_policy/opts.py#L121\n\nThat way keystone can opt-in to this behavior by default, and it can stay off by default for other users of the library. Although since I think keystone is one of the main consumers of oslo.cache I wonder how much effect it will have to change this default only in other projects.","accounts_in_message":[],"_revision_number":7},{"id":"75ce9560d58de9c3c03138eabc001dfdda325006","author":{"_account_id":27339,"name":"Michal Arbet","email":"michal.arbet@ultimum.io","username":"michalarbet"},"date":"2020-07-27 15:29:44.000000000","message":"Uploaded patch set 8: Patch Set 7 was rebased.","accounts_in_message":[],"_revision_number":8},{"id":"23f003b7e16350509606e9829fe62fa39f0b5d8e","author":{"_account_id":27339,"name":"Michal Arbet","email":"michal.arbet@ultimum.io","username":"michalarbet"},"date":"2020-07-27 15:30:46.000000000","message":"Patch Set 8:\n\n(2 comments)","accounts_in_message":[],"_revision_number":8},{"id":"85d7d13fb587969a1ef3bc71d9755addf079fa62","author":{"_account_id":5046,"name":"Lance Bragstad","email":"lbragstad@redhat.com","username":"ldbragst"},"date":"2020-07-27 16:08:42.000000000","message":"Patch Set 8:\n\n(5 comments)\n\nI have some questions because I don\u0027t understand how setting this option is resulting in connection issues.\n\nIs there a direct correlation between flushing the cache and network connections?","accounts_in_message":[],"_revision_number":8},{"id":"d82006adb849ad0c2a0c5384cde2da93594c7294","tag":"autogenerated:zuul:check","author":{"_account_id":22348,"name":"Zuul","username":"zuul","tags":["SERVICE_USER"]},"date":"2020-07-27 16:47:57.000000000","message":"Patch Set 8:\n\nBuild succeeded (check pipeline).\n\n- tempest-full-py3 https://zuul.opendev.org/t/openstack/build/723d255fe7744bb681f6d1f6ff84fceb : SUCCESS in 1h 15m 15s\n- openstack-tox-lower-constraints https://zuul.opendev.org/t/openstack/build/06e9e06e9c75446ba721126c85e68d42 : SUCCESS in 4m 36s\n- openstack-tox-pep8 https://zuul.opendev.org/t/openstack/build/f81eaf5b4a5f4e4fb601c978ae153ae7 : SUCCESS in 4m 07s\n- openstack-tox-py36 https://zuul.opendev.org/t/openstack/build/d2438779f2374615a55a0721889a7e7a : SUCCESS in 4m 29s\n- openstack-tox-py38 https://zuul.opendev.org/t/openstack/build/4e760524064b4c098e7fd0caf369ea40 : SUCCESS in 4m 45s\n- openstack-tox-docs https://zuul.opendev.org/t/openstack/build/8e6bf1b2b4ba48d08d48090a40a50d33 : SUCCESS in 6m 39s\n- build-openstack-releasenotes https://zuul.opendev.org/t/openstack/build/37eaf16b4c6b448980d14330231a3e59 : SUCCESS in 4m 38s\n- oslo.cache-functional-tox-py38-etcd3gw https://zuul.opendev.org/t/openstack/build/e03b36f062dc4003a67d82fd5edf6f5d : SUCCESS in 5m 04s","accounts_in_message":[],"_revision_number":8},{"id":"88d2e58406a900e2d5dae122912cae91b8e550b3","author":{"_account_id":27339,"name":"Michal Arbet","email":"michal.arbet@ultimum.io","username":"michalarbet"},"date":"2020-07-27 17:19:26.000000000","message":"Uploaded patch set 9.","accounts_in_message":[],"_revision_number":9},{"id":"41e26a8e3c552564fcd08a743cf3b7e593a287ce","author":{"_account_id":27339,"name":"Michal Arbet","email":"michal.arbet@ultimum.io","username":"michalarbet"},"date":"2020-07-27 17:20:42.000000000","message":"Patch Set 9:\n\n(4 comments)","accounts_in_message":[],"_revision_number":9},{"id":"09a5be24a82072b0f09d71d014d54fd0956aaf1f","author":{"_account_id":27339,"name":"Michal Arbet","email":"michal.arbet@ultimum.io","username":"michalarbet"},"date":"2020-07-27 17:51:12.000000000","message":"Patch Set 9:\n\n(1 comment)","accounts_in_message":[],"_revision_number":9},{"id":"3615c188b85165746e6b1fcf4e8af7d719631e17","tag":"autogenerated:zuul:check","author":{"_account_id":22348,"name":"Zuul","username":"zuul","tags":["SERVICE_USER"]},"date":"2020-07-27 18:21:07.000000000","message":"Patch Set 9: Verified+1\n\nBuild succeeded (check pipeline).\n\n- tempest-full-py3 https://zuul.opendev.org/t/openstack/build/4eec7a1fafdd4f4fa8e49736ff8d86cd : SUCCESS in 58m 32s\n- openstack-tox-lower-constraints https://zuul.opendev.org/t/openstack/build/7f3cecdea1f54d1f80b245eb163c4722 : SUCCESS in 4m 48s\n- openstack-tox-pep8 https://zuul.opendev.org/t/openstack/build/05ab69fdc0244cf99d79e1f3e873b11b : SUCCESS in 4m 09s\n- openstack-tox-py36 https://zuul.opendev.org/t/openstack/build/26d63bab51f14830b050abcadc6c770b : SUCCESS in 3m 44s\n- openstack-tox-py38 https://zuul.opendev.org/t/openstack/build/3adb95a54488417ea02b5fc704833904 : SUCCESS in 5m 28s\n- openstack-tox-docs https://zuul.opendev.org/t/openstack/build/5a92c33a2a5646058f1e747bcdfcee50 : SUCCESS in 3m 42s\n- build-openstack-releasenotes https://zuul.opendev.org/t/openstack/build/8b3f59d0474146e9b4b312d82e7c3243 : SUCCESS in 5m 04s\n- oslo.cache-functional-tox-py38-etcd3gw https://zuul.opendev.org/t/openstack/build/3e1e374aa90f46c4a603023240a0e505 : SUCCESS in 5m 43s","accounts_in_message":[],"_revision_number":9},{"id":"6f9f3093c274e16e89c08ced6f213cd13a26cedc","author":{"_account_id":5046,"name":"Lance Bragstad","email":"lbragstad@redhat.com","username":"ldbragst"},"date":"2020-07-27 20:29:52.000000000","message":"Patch Set 8:\n\n(1 comment)","accounts_in_message":[],"_revision_number":8},{"id":"7cbd0f7c25cd936dbda065f5b0cf9bf9db0a8f60","author":{"_account_id":22072,"name":"Viktor Křivák","email":"viktor.krivak@ultimum.io","username":"viktor-krivak"},"date":"2020-07-28 14:27:24.000000000","message":"Patch Set 8:\n\n\u003e (1 comment)\n\nI think that this flag doesn\u0027t mitigate the problem entirely. If I understand it correctly the original attack vector is:\n- Issue token\n- Somehow stop memcached server that stores this token\n- Revoke token\n- Start memcached server again\n- Token is valid again and can be misused\nBut even with flush_on_reconnect similar issue is still possible:\n- Issue token\n- Somehow stop memcached server that stores this token\n- Wait before someone is use this token again so it is again pushed to the other memcached (token is still valid)\n- Start memcached server\n- Revoke token\n- Stop the original memcached server again\n- Since the other memcache is still holding the old token it is valid again despite the fact that you have flushed the data on the original memcache (that was stopped and started and stopped again)\n\nSo it\u0027s actually doesn\u0027t improve security that much. And add a ton of problem with performance.","accounts_in_message":[],"_revision_number":8},{"id":"001eaa59e324712a71f1f5d0d6ac10f558b9d917","author":{"_account_id":22072,"name":"Viktor Křivák","email":"viktor.krivak@ultimum.io","username":"viktor-krivak"},"date":"2020-07-28 14:31:02.000000000","message":"Patch Set 9:\n\n(1 comment)","accounts_in_message":[],"_revision_number":9},{"id":"1a0cc397f5b9aaee8f3ba4f539312985dc257f54","author":{"_account_id":27339,"name":"Michal Arbet","email":"michal.arbet@ultimum.io","username":"michalarbet"},"date":"2020-08-03 11:20:22.000000000","message":"Patch Set 9:\n\n(1 comment)\n\nHi, it is currently one week from oslo team discussed on irc meeting. So, what is status of this review ? \n\nThanks, Michal","accounts_in_message":[],"_revision_number":9},{"id":"ef8352bb1fd05ca5ddb8e705bf4ba149a40224ea","author":{"_account_id":6928,"name":"Ben Nemec","email":"openstack@nemebean.com","username":"bnemec"},"date":"2020-08-10 15:41:48.000000000","message":"Patch Set 9: Code-Review-1\n\n(2 comments)\n\nAfter discussion in the Oslo meeting today, we\u0027ve agreed to move forward with this change, with the caveat that we want Keystone to be able to opt in to the old behavior since it is security sensitive. Some more comments inline.","accounts_in_message":[],"_revision_number":9},{"id":"5b96a618f2aada8394bcb9cea86edbe1d923d40d","author":{"_account_id":32428,"name":"Tony Liu","email":"tonyliu0592@hotmail.com"},"date":"2020-09-16 15:49:25.000000000","message":"Patch Set 9:\n\nWhat\u0027s the problem we are trying to solve here?\nThe problem I saw is that, with Ussuri, there are thousands connections from neutron-server to memcached, to make memcached irresponsible. Given some analysis, this is because flush_on_reconnect is hardcoded to true. So we are making it optional.\n1) I have a question here, oslo_cache is used by keystone-middleware, which is used by keystone, neutron-server and nova-api. Why do I see thousands connections from neutron-server only, but not from keystone and nova-api? Could there be anything wrong in neutron-server?\n2) If we do need to expose this argument to configuration, we should keep backward compatibility.\n3) In future, we should dig deeper to find out root cause of such connection piling.\n\nThanks!\nTony","accounts_in_message":[],"_revision_number":9},{"id":"ba1b3e6d908cc5fc9d6c8ab2e5f509ba8f1ba6b7","author":{"_account_id":27339,"name":"Michal Arbet","email":"michal.arbet@ultimum.io","username":"michalarbet"},"date":"2020-10-08 17:09:31.000000000","message":"Abandoned","accounts_in_message":[],"_revision_number":9},{"id":"f7586091f858a23f1944e36640cca82ab708095e","author":{"_account_id":28522,"name":"Hervé Beraud","email":"herveberaud.pro@gmail.com","username":"hberaud"},"date":"2021-01-13 11:29:33.000000000","message":"Patch Set 9:\n\nI think that the root cause is a python-memcached issue, I think it\u0027s due to the fact that when python-memcached try to get a socket and when the `flush_on_reconnect` param is passed to python-memcached it will do a `flush_all` [1][2]. However I don\u0027t know if this really something expected by that library or simply something not expected, I\u0027ll try to discuss with the maintainers.\n\nSimilar issues have been reported on the library side [3].\n\nHowever, for now, I think we can continue with this patch to make the `flush_on_reconnect` optional and available for keystone.\n\nI seen many related bugs [4][5] open downstream so this should be addressed.\n\nI\u0027ll try to implement the needed changes as proposed by Ben in is previous comment.\n\n[1] https://github.com/linsomniac/python-memcached/blob/master/memcache.py#L1412,L1413\n[2] https://github.com/linsomniac/python-memcached/blob/master/memcache.py#L1487\n[3] https://github.com/linsomniac/python-memcached/issues/179\n[4] https://bugzilla.redhat.com/show_bug.cgi?id\u003d1893205\n[5] https://bugzilla.redhat.com/show_bug.cgi?id\u003d1915700","accounts_in_message":[],"_revision_number":9},{"id":"0b67873f59a961b84c2c73e6cc00159abadfb308","tag":"autogenerated:gerrit:restore","author":{"_account_id":28522,"name":"Hervé Beraud","email":"herveberaud.pro@gmail.com","username":"hberaud"},"date":"2021-01-13 11:30:07.000000000","message":"Restored","accounts_in_message":[],"_revision_number":9},{"id":"16f285b0312add89d284cb0a722045f2735de7d8","tag":"autogenerated:zuul:check","author":{"_account_id":22348,"name":"Zuul","username":"zuul","tags":["SERVICE_USER"]},"date":"2021-01-13 12:50:14.000000000","message":"Patch Set 9: Verified-1\n\nBuild failed (check pipeline).  For information on how to proceed, see\nhttps://docs.opendev.org/opendev/infra-manual/latest/developers.html#automated-testing\n\n\n- tempest-full-py3 https://zuul.opendev.org/t/openstack/build/d159d16e0a844e3181ca75fdffaf58a7 : SUCCESS in 1h 17m 32s\n- openstack-tox-pep8 https://zuul.opendev.org/t/openstack/build/72503d6bab6b4da197ba160ef7c05c2a : SUCCESS in 4m 25s\n- openstack-tox-py36 https://zuul.opendev.org/t/openstack/build/2e392dded15640918ccb0e420cd1a9fe : SUCCESS in 6m 13s\n- openstack-tox-py38 https://zuul.opendev.org/t/openstack/build/d5ae543ee18e4e02890e16d14d392944 : SUCCESS in 3m 44s\n- openstack-tox-py39 https://zuul.opendev.org/t/openstack/build/b982523a380746d2b6ab458ba2b6629a : SUCCESS in 7m 21s (non-voting)\n- openstack-tox-docs https://zuul.opendev.org/t/openstack/build/2570fc36876a4a12b430835290ea9fc7 : SUCCESS in 4m 45s\n- build-openstack-releasenotes https://zuul.opendev.org/t/openstack/build/e79dfdfb48984a77bc6e5325a23fe347 : SUCCESS in 5m 01s\n- oslo.cache-functional-etcd3gw https://zuul.opendev.org/t/openstack/build/a5e770968b4249aeb0dd2b262ec3f762 : FAILURE in 3m 28s\n- oslo.cache-functional-dogpile.cache.bmemcached https://zuul.opendev.org/t/openstack/build/db81cf61e10046e199dee41dfcd51fb5 : FAILURE in 5m 08s\n- oslo.cache-functional-memcache_pool https://zuul.opendev.org/t/openstack/build/9751a853046e4b7d84d8ef6a1c5c1d96 : FAILURE in 6m 28s","accounts_in_message":[],"_revision_number":9},{"id":"660689faf10c5cb08ea1bb5f29d8a1866b30adcf","tag":"autogenerated:gerrit:newPatchSet","author":{"_account_id":28522,"name":"Hervé Beraud","email":"herveberaud.pro@gmail.com","username":"hberaud"},"date":"2021-01-14 06:32:05.000000000","message":"Uploaded patch set 10.","accounts_in_message":[],"_revision_number":10},{"id":"548c6bf962155c4615b5897b7191176490758056","tag":"autogenerated:zuul:check","author":{"_account_id":22348,"name":"Zuul","username":"zuul","tags":["SERVICE_USER"]},"date":"2021-01-14 08:13:13.000000000","message":"Patch Set 10: Verified-1\n\n(1 comment)\n\nBuild failed (check pipeline).  For information on how to proceed, see\nhttps://docs.opendev.org/opendev/infra-manual/latest/developers.html#automated-testing\n\n\n- tempest-full-py3 https://zuul.opendev.org/t/openstack/build/be82bf0801d74581991e1cb5bd839784 : SUCCESS in 1h 18m 01s\n- openstack-tox-pep8 https://zuul.opendev.org/t/openstack/build/a504f04e5b204593a3929ee2716ce185 : FAILURE in 4m 18s\n- openstack-tox-py36 https://zuul.opendev.org/t/openstack/build/849873e17f794387950a358badd0ef50 : SUCCESS in 3m 28s\n- openstack-tox-py38 https://zuul.opendev.org/t/openstack/build/c1ed8c4890f949bebace3b0f8c82f60a : SUCCESS in 3m 51s\n- openstack-tox-py39 https://zuul.opendev.org/t/openstack/build/fdbc1e0a1d614fec8243d43052c69321 : SUCCESS in 4m 55s (non-voting)\n- openstack-tox-docs https://zuul.opendev.org/t/openstack/build/4fecf3de19204d09b01403964e4ba9f3 : SUCCESS in 4m 58s\n- build-openstack-releasenotes https://zuul.opendev.org/t/openstack/build/bcd4b84805f040ff914c8fcb560b8f31 : SUCCESS in 4m 00s\n- oslo.cache-functional-etcd3gw https://zuul.opendev.org/t/openstack/build/969fd4a735b34910aab947a3eb304f66 : FAILURE in 4m 28s\n- oslo.cache-functional-dogpile.cache.bmemcached https://zuul.opendev.org/t/openstack/build/4487809c74ed438e97b2063be8045cc0 : FAILURE in 4m 00s\n- oslo.cache-functional-memcache_pool https://zuul.opendev.org/t/openstack/build/5f09b0855e394bb68d4fa8a66d3109c6 : FAILURE in 3m 35s","accounts_in_message":[],"_revision_number":10},{"id":"bb08bcd6fbba89b3147c5975182484b76eec747d","tag":"autogenerated:gerrit:newPatchSet","author":{"_account_id":28522,"name":"Hervé Beraud","email":"herveberaud.pro@gmail.com","username":"hberaud"},"date":"2021-01-14 08:56:26.000000000","message":"Uploaded patch set 11.","accounts_in_message":[],"_revision_number":11},{"id":"e8c6ded38803b14d1d57ba7723ff7411d98dcdf5","tag":"autogenerated:zuul:check","author":{"_account_id":22348,"name":"Zuul","username":"zuul","tags":["SERVICE_USER"]},"date":"2021-01-14 10:41:50.000000000","message":"Patch Set 11: Verified-1\n\nBuild failed (check pipeline).  For information on how to proceed, see\nhttps://docs.opendev.org/opendev/infra-manual/latest/developers.html#automated-testing\n\n\n- tempest-full-py3 https://zuul.opendev.org/t/openstack/build/769976f146cb4128b2d54dd375cefe78 : SUCCESS in 1h 39m 54s\n- openstack-tox-pep8 https://zuul.opendev.org/t/openstack/build/165c3a4c8cb64d35a91222fc85e87fb4 : SUCCESS in 3m 59s\n- openstack-tox-py36 https://zuul.opendev.org/t/openstack/build/315a63ca8bc943349c2981d44f03017f : SUCCESS in 4m 10s\n- openstack-tox-py38 https://zuul.opendev.org/t/openstack/build/798bfba70adf4354a521475b12ed4311 : SUCCESS in 3m 45s\n- openstack-tox-py39 https://zuul.opendev.org/t/openstack/build/4c230c56b3934c78bbb104eebfde85ef : SUCCESS in 4m 24s (non-voting)\n- openstack-tox-docs https://zuul.opendev.org/t/openstack/build/ac78463b6b5e45e1a8330be71c37981f : SUCCESS in 7m 02s\n- build-openstack-releasenotes https://zuul.opendev.org/t/openstack/build/aec1f8173e2d4d1c94827bc12fc44f3c : SUCCESS in 3m 14s\n- oslo.cache-functional-etcd3gw https://zuul.opendev.org/t/openstack/build/0241518839d943ea90b5f460face0a5b : FAILURE in 4m 03s\n- oslo.cache-functional-dogpile.cache.bmemcached https://zuul.opendev.org/t/openstack/build/3bf2aaf1c20948a589fba14300129ef5 : FAILURE in 4m 29s\n- oslo.cache-functional-memcache_pool https://zuul.opendev.org/t/openstack/build/1654f3ce185e4c85913f36b5c60bd4b2 : FAILURE in 3m 54s","accounts_in_message":[],"_revision_number":11},{"id":"2f1f264241549c49c8a7da6c7531e3f699a504f3","author":{"_account_id":28522,"name":"Hervé Beraud","email":"herveberaud.pro@gmail.com","username":"hberaud"},"date":"2021-01-18 10:59:05.000000000","message":"Patch Set 11:\n\nrecheck\n\nError don\u0027t seems related to these changes:\n\n```\ntox.exception.MissingSubstitution: MissingSubstitution: PIFPAF_DAEMON\n```","accounts_in_message":[],"_revision_number":11},{"id":"1f3a766dbf52a6aafc540773d4590e07db6f46b7","tag":"autogenerated:zuul:check","author":{"_account_id":22348,"name":"Zuul","username":"zuul","tags":["SERVICE_USER"]},"date":"2021-01-18 12:56:46.000000000","message":"Patch Set 11:\n\nBuild failed (check pipeline).  For information on how to proceed, see\nhttps://docs.opendev.org/opendev/infra-manual/latest/developers.html#automated-testing\n\n\n- tempest-full-py3 https://zuul.opendev.org/t/openstack/build/fa7dd0f99e6d44168d20f55e0299b662 : SUCCESS in 1h 32m 38s\n- openstack-tox-pep8 https://zuul.opendev.org/t/openstack/build/67026a66d8d74d86a0f2f6c70bda7d1b : SUCCESS in 4m 12s\n- openstack-tox-py36 https://zuul.opendev.org/t/openstack/build/b0c38a173c90469c938bf27318d4e4b5 : SUCCESS in 4m 21s\n- openstack-tox-py38 https://zuul.opendev.org/t/openstack/build/f9b79261aafc408bae21a70e5720533e : SUCCESS in 4m 41s\n- openstack-tox-py39 https://zuul.opendev.org/t/openstack/build/1de32387b12a4443b4ab5893197f51dd : SUCCESS in 4m 06s (non-voting)\n- openstack-tox-docs https://zuul.opendev.org/t/openstack/build/ee4da8b61bbe42d2b3d5a617549c2ee0 : SUCCESS in 4m 44s\n- build-openstack-releasenotes https://zuul.opendev.org/t/openstack/build/885719aa71c94111a0bec508c58ca7e8 : SUCCESS in 4m 55s\n- oslo.cache-functional-etcd3gw https://zuul.opendev.org/t/openstack/build/14c30a779ae74441b5563ce2136793b0 : FAILURE in 4m 21s\n- oslo.cache-functional-dogpile.cache.bmemcached https://zuul.opendev.org/t/openstack/build/418751d8da604523aecc84860af2fbf2 : FAILURE in 4m 56s\n- oslo.cache-functional-memcache_pool https://zuul.opendev.org/t/openstack/build/bf864751008741f1863f0514164d44a7 : FAILURE in 4m 22s","accounts_in_message":[],"_revision_number":11},{"id":"5184aa37e769790bcb4eefd0817cd2cc48e78c7c","tag":"autogenerated:gerrit:newPatchSet","author":{"_account_id":28522,"name":"Hervé Beraud","email":"herveberaud.pro@gmail.com","username":"hberaud"},"date":"2021-01-18 13:00:08.000000000","message":"Uploaded patch set 12: Patch Set 11 was rebased.","accounts_in_message":[],"_revision_number":12},{"id":"846eea9a376fab19a6f0cb1d559d96c4b2029ca2","tag":"autogenerated:zuul:check","author":{"_account_id":22348,"name":"Zuul","username":"zuul","tags":["SERVICE_USER"]},"date":"2021-01-18 14:44:01.000000000","message":"Patch Set 12: Verified-1\n\nBuild failed (check pipeline).  For information on how to proceed, see\nhttps://docs.opendev.org/opendev/infra-manual/latest/developers.html#automated-testing\n\n\n- tempest-full-py3 https://zuul.opendev.org/t/openstack/build/94d7c57f5ef242a2aa49171d03cbf4e5 : SUCCESS in 1h 37m 17s\n- openstack-tox-pep8 https://zuul.opendev.org/t/openstack/build/408f34e3ca3040689b0cbf0843fda0d0 : SUCCESS in 3m 33s\n- openstack-tox-py36 https://zuul.opendev.org/t/openstack/build/82b537df5bd54516b8bf6215194799c5 : SUCCESS in 3m 42s\n- openstack-tox-py38 https://zuul.opendev.org/t/openstack/build/5a39ab864f214603895abdcc29b7d62c : SUCCESS in 4m 21s\n- openstack-tox-py39 https://zuul.opendev.org/t/openstack/build/c55b396539f845f2b4bdc22ba404ea28 : SUCCESS in 3m 57s (non-voting)\n- openstack-tox-docs https://zuul.opendev.org/t/openstack/build/557190bde58e48faacde2a6e5bfbe226 : SUCCESS in 5m 07s\n- build-openstack-releasenotes https://zuul.opendev.org/t/openstack/build/67ef6158f79d40cf884c5cb02e9220e3 : SUCCESS in 3m 38s\n- oslo.cache-functional-etcd3gw https://zuul.opendev.org/t/openstack/build/e526d51e7fb643818a1d014f017b36be : FAILURE in 3m 51s\n- oslo.cache-functional-dogpile.cache.bmemcached https://zuul.opendev.org/t/openstack/build/c0e8c612c51f4e9c89b31d071292dd33 : FAILURE in 4m 34s\n- oslo.cache-functional-memcache_pool https://zuul.opendev.org/t/openstack/build/f07c3e08763342c4acb895edeb4fb37b : FAILURE in 5m 31s","accounts_in_message":[],"_revision_number":12},{"id":"6e8ee3393d865c384862111d415a7e7dd3257321","author":{"_account_id":28522,"name":"Hervé Beraud","email":"herveberaud.pro@gmail.com","username":"hberaud"},"date":"2021-01-19 08:43:07.000000000","message":"Patch Set 12:\n\nrecheck","accounts_in_message":[],"_revision_number":12},{"id":"f99eea4430d33884e08265d3f4479199f7ae9111","tag":"autogenerated:zuul:check","author":{"_account_id":22348,"name":"Zuul","username":"zuul","tags":["SERVICE_USER"]},"date":"2021-01-19 09:56:16.000000000","message":"Patch Set 12: Verified+1\n\nBuild succeeded (check pipeline).\n\n- tempest-full-py3 https://zuul.opendev.org/t/openstack/build/bc22123340b04801a3ac5cc7a0fb5d6e : SUCCESS in 1h 11m 16s\n- openstack-tox-pep8 https://zuul.opendev.org/t/openstack/build/89c00200876849c7805261eae461debb : SUCCESS in 3m 34s\n- openstack-tox-py36 https://zuul.opendev.org/t/openstack/build/5414ee839e5943f2abea7167f3b84ddd : SUCCESS in 3m 38s\n- openstack-tox-py38 https://zuul.opendev.org/t/openstack/build/a33bbd56eea648a9ba288caf60b709b8 : SUCCESS in 3m 25s\n- openstack-tox-py39 https://zuul.opendev.org/t/openstack/build/25f19b2da03241e4b7db82a9035cfac1 : SUCCESS in 3m 35s (non-voting)\n- openstack-tox-docs https://zuul.opendev.org/t/openstack/build/35384715c66246cfb08da8dee8de4412 : SUCCESS in 4m 40s\n- build-openstack-releasenotes https://zuul.opendev.org/t/openstack/build/1d997ecceb6e4d3b9fbc6fba86d9a0d1 : SUCCESS in 3m 42s\n- oslo.cache-functional-etcd3gw https://zuul.opendev.org/t/openstack/build/5f46f51fc56443858ad2c06db6309578 : SUCCESS in 4m 07s\n- oslo.cache-functional-dogpile.cache.bmemcached https://zuul.opendev.org/t/openstack/build/6edb760ce16042559516f859dac5fc69 : SUCCESS in 3m 50s\n- oslo.cache-functional-memcache_pool https://zuul.opendev.org/t/openstack/build/bfc64f9f8e974a5399a4fcb6c7fd5c25 : SUCCESS in 4m 11s","accounts_in_message":[],"_revision_number":12},{"id":"22958bb55b6c4256d26e938ce6b63e7fa668bcdf","author":{"_account_id":27954,"name":"Moisés Guimarães de Medeiros","email":"guimaraes@pm.me","username":"moguimar"},"date":"2021-01-25 14:44:46.000000000","message":"Patch Set 12: Code-Review+2","accounts_in_message":[],"_revision_number":12},{"id":"702edb7acb143acb68cd23970167202ebd7c31a5","author":{"_account_id":6928,"name":"Ben Nemec","email":"openstack@nemebean.com","username":"bnemec"},"date":"2021-01-25 16:26:57.000000000","message":"Patch Set 12: Code-Review+2\n\n(2 comments)\n\nI\u0027m good with the code. The comments I left inline could be addressed in a followup.\n\nCan we get/do we have a WIP patch posted to keystone that makes use of this parameter? That\u0027s an important step in this fix so I want to make sure it doesn\u0027t get forgotten.","accounts_in_message":[],"_revision_number":12},{"id":"34ebaf9eb6080dc04156a68871f45ce9776882b0","author":{"_account_id":28522,"name":"Hervé Beraud","email":"herveberaud.pro@gmail.com","username":"hberaud"},"date":"2021-01-25 16:31:01.000000000","message":"Patch Set 12:\n\n(1 comment)\n\n\u003e Patch Set 12: Code-Review+2\n\u003e \n\u003e (2 comments)\n\u003e \n\u003e I\u0027m good with the code. The comments I left inline could be addressed in a followup.\n\u003e \n\u003e Can we get/do we have a WIP patch posted to keystone that makes use of this parameter? That\u0027s an important step in this fix so I want to make sure it doesn\u0027t get forgotten.\n\nIndeed you are right, I\u0027ll propose the patch right now and refer to this one as a \"depends-on\"","accounts_in_message":[],"_revision_number":12},{"id":"4aa3300eb7fd8ec51a700572fe939de75c3ec2e4","author":{"_account_id":28522,"name":"Hervé Beraud","email":"herveberaud.pro@gmail.com","username":"hberaud"},"date":"2021-01-25 18:08:45.000000000","message":"Patch Set 12:\n\nThe keystone patch: https://review.opendev.org/c/openstack/keystone/+/772394","accounts_in_message":[],"_revision_number":12},{"id":"8e1a2ecb784a15e4f621bbb2f97438d1068b77fc","author":{"_account_id":11816,"name":"mike_mp@zzzcomputing.com","display_name":"Mike Bayer","email":"mike_mp@zzzcomputing.com","username":"zzzeek","status":"Red Hat"},"date":"2021-01-28 14:08:06.000000000","message":"Patch Set 12:\n\nI\u0027m fine with this patch but still wondering about the first comment.  Do we actually know *why* this setting would cause many new connections to memcached?   The flush_on_reconnect setting AFAICT does nothing more than send a \"FLUSH ALL\" command to the memcached server, which means, it will expire all the existing keys.  \n\nThe negative side effect we\u0027d expect from this is that clients which rely upon memcached for good performance will now have to go out to the database and get new data.   But nothing intrinsic to that suggests this means many more memcached connections.\n\nThe scenario that *does* make sense for that is:\n\n1. neutron has some caching scenario where new memcached connections are created on a per request basis (this would be a bug, because there is no throttle)\n\n2. when a memcached server receives flush all, neutron now has to spend much more time with its requests as it has to go out to the database to refresh things\n\n3. neutron requests pile up on top of each other (Again, this is a bug in configuration, neutron is configured for too much concurrency and/or the database is not correctly tuned)\n\n4. all the neutron reqs have brand new memcached connections that then overload ....the memcached server?  which also should be configured correctly so that the memcached server doesnt fall over when connections are maxed, it seems we\u0027ve observed lowering thread counts and max connecvtions for memcached also resolves\n\n\noverall flush_on_reconnect is likely not needed for our cases by default but this doesn\u0027t seem like it\u0027s the actual bug we are seeing w/ neutron.","accounts_in_message":[],"_revision_number":12},{"id":"263a94ebc48b9201fbc3accd6888b0fb0f634a34","author":{"_account_id":27339,"name":"Michal Arbet","email":"michal.arbet@ultimum.io","username":"michalarbet"},"date":"2021-02-01 10:25:07.000000000","message":"Patch Set 12:\n\n\u003e Patch Set 12:\n\u003e \n\u003e I\u0027m fine with this patch but still wondering about the first comment.  Do we actually know *why* this setting would cause many new connections to memcached?   The flush_on_reconnect setting AFAICT does nothing more than send a \"FLUSH ALL\" command to the memcached server, which means, it will expire all the existing keys.  \n\u003e \n\u003e The negative side effect we\u0027d expect from this is that clients which rely upon memcached for good performance will now have to go out to the database and get new data.   But nothing intrinsic to that suggests this means many more memcached connections.\n\u003e \n\u003e The scenario that *does* make sense for that is:\n\u003e \n\u003e 1. neutron has some caching scenario where new memcached connections are created on a per request basis (this would be a bug, because there is no throttle)\n\u003e \n\u003e 2. when a memcached server receives flush all, neutron now has to spend much more time with its requests as it has to go out to the database to refresh things\n\u003e \n\u003e 3. neutron requests pile up on top of each other (Again, this is a bug in configuration, neutron is configured for too much concurrency and/or the database is not correctly tuned)\n\u003e \n\u003e 4. all the neutron reqs have brand new memcached connections that then overload ....the memcached server?  which also should be configured correctly so that the memcached server doesnt fall over when connections are maxed, it seems we\u0027ve observed lowering thread counts and max connecvtions for memcached also resolves\n\u003e \n\u003e \n\u003e overall flush_on_reconnect is likely not needed for our cases by default but this doesn\u0027t seem like it\u0027s the actual bug we are seeing w/ neutron.\n\n\nThe idea arose during the openstack tests:\n\n1. While loop openstack token issue in several bashes at once\n2. While loop openstack volume list in several bashes at once\n3. While loop openstack server list in several bashes at once\n.\n.\n.\n\nParallely with above tests I was counting connections to memcached with script.\nFrom these data I created graph, and found that during the individual tests,\nthe connections were raising, only with neutron and keystone were not raising.\n\nSince keystone itself does not use keystonemiddleware, I started looking for something which is configured differently in the neutron\nand found an option in kolla-ansible\u0027s template for neutron \u003d\u003e memcache_use_advanced_pool  \u003d True .\n\nCommit : https://github.com/openstack/kolla-ansible/commit/5a52d8e4a0c5d4c246deb8851ef893df63ee0847\nBug: https://bugs.launchpad.net/keystonemiddleware/+bug/1883659\n\nFrom bug (Pierre Riteau (priteau)): \n\n\u003e\u003e\u003e\u003e\u003e\n\nI am able to make the problem go away with this extra setting in neutron.conf:\n\n[keystone_authtoken]\nmemcache_use_advanced_pool \u003d True\n\nThis is the documentation for this setting:\n\n# (Optional) Use the advanced (eventlet safe) memcached client pool. The\n# advanced pool will only work under python 2.x. (boolean value)\n\nThis description dates from 2016. For now I haven\u0027t seen any issue enabling this setting with Python 3.\n\n\u003c\u003c\u003c\u003c\u003c\n\n\nI then added this option to the affected services and problem dissapeared for all services.\n\nAnyway, I think this patch + keystone patch should be merged as is, I just wanted to \nreact to Mike. Just imagine there is some network problem and clients are sending flush_all very often.\n\nI\u0027m adding also link to graph what I\u0027ve created\n\nhttps://ibb.co/7g2yRGX\n\nLegend :\n\nX -\u003e Cycles (Counter how much loops I\u0027ve iterated to API ^^ in Bash tests)\nY -\u003e Number of connections to memcached (which were raising)\n\n1800_non_f -\u003e 1800 seconds - 15 minutes, with fluentd turned off (logs were sent to 127.0.0.1 which was unreachable because was down)\n1800 -\u003e 1800 seconds - 15 minutes (normal env without remote fluentd logging)\n\nNote : In graph I measured also API calls with fluentd turned off (to simulate situation when\nfluentd is not reachable and when we originally found an issue)\nI suppose that issue is more visible when fluentd and fluentformatter is used from oslo.log\nbecause fluent formatter is adding whole context to log, and these data are from cache (keystone_middleware) so it is raising faster, in our case production was down when fluentd were offline, there are tons of logs...\nHave to say that Fluentd logging just points his finger to a *problem with keystonemiddleware* which is probably very ugly bug.\n\nAll above was tested on clean kolla-ansible (ussuri) with debian/binary (ussuri) images.","accounts_in_message":[],"_revision_number":12},{"id":"53b22a9d39944c01532dfff381885e27ee327b78","author":{"_account_id":27339,"name":"Michal Arbet","email":"michal.arbet@ultimum.io","username":"michalarbet"},"date":"2021-02-01 15:31:39.000000000","message":"Patch Set 12:\n\nI think this is caused by usage of obsolete code which was used in keystonemiddleware long time ago - before oslo.cache support was added to keystonemiddleware as new library (in past time :)).\n\nServices which are using keystonemiddleware should use memcache_use_advanced_pool \u003d True (oslo.cache memcached_pool implementation) instead of obsolete code.\n\nOr better said option memcache_use_advanced_pool should be removed and keystonemiddleware should use oslo.cache implementation by default.\n\nOslo.cache introducing and adding to requirements in :\n https://review.opendev.org/c/openstack/keystonemiddleware/+/268664\n https://review.opendev.org/c/openstack/keystonemiddleware/+/527466/","accounts_in_message":[],"_revision_number":12},{"id":"9931b6ea2fac4bdb641ac8404416b0969dc05042","author":{"_account_id":11816,"name":"mike_mp@zzzcomputing.com","display_name":"Mike Bayer","email":"mike_mp@zzzcomputing.com","username":"zzzeek","status":"Red Hat"},"date":"2021-02-01 15:56:57.000000000","message":"Patch Set 12:\n\nhi and thanks for this reply.\n\n\n\u003e Parallely with above tests I was counting connections to memcached with script.\n\u003e From these data I created graph, and found that during the individual tests,\n\u003e the connections were raising, only with neutron and keystone were not raising.\n\n\nI\u0027m not following this, I thought the issue was that neutron *was* creating too many memcached connections.   Is this after you applied the pooling option or before?\n\n\n\u003e I then added this option to the affected services and problem dissapeared for all services.\n\u003e \n\u003e Anyway, I think this patch + keystone patch should be merged as is, I just wanted to \n\u003e react to Mike. Just imagine there is some network problem and clients are sending flush_all very often.\n\n\nyes, flush_all will erase all the keys. how does that lead to more memcached connections?\n\n\n\n\u003e \n\u003e I\u0027m adding also link to graph what I\u0027ve created\n\u003e \n\u003e https://ibb.co/7g2yRGX\n\u003e \n\u003e Legend :\n\u003e \n\u003e X -\u003e Cycles (Counter how much loops I\u0027ve iterated to API ^^ in Bash tests)\n\u003e Y -\u003e Number of connections to memcached (which were raising)\n\n\nthis is..with a fix applied to neutron?  specifically the memcached_advanced_pool option?  is the test working with changing the flush_all option and seeing what effect that has?  Sorry I just am not getting all the context straight here.","accounts_in_message":[],"_revision_number":12},{"id":"c4bc95ab867fca323332b0c274fd49ad66216a76","author":{"_account_id":27339,"name":"Michal Arbet","email":"michal.arbet@ultimum.io","username":"michalarbet"},"date":"2021-02-01 16:36:39.000000000","message":"Patch Set 12:\n\n\u003e Patch Set 12:\n\u003e \n\u003e hi and thanks for this reply.\n\u003e \n\u003e \n\u003e \u003e Parallely with above tests I was counting connections to memcached with script.\n\u003e \u003e From these data I created graph, and found that during the individual tests,\n\u003e \u003e the connections were raising, only with neutron and keystone were not raising.\n\u003e \n\u003e \n\u003e I\u0027m not following this, I thought the issue was that neutron *was* creating too many memcached connections.   Is this after you applied the pooling option or before?\n\nWell, graph is with  memcache_use_advanced_pool \u003d True , but when it was measured I really didn\u0027t know that neutron has set it to True. I just found it because i was trying to found differencies between services, after I found that kolla-ansible changed in https://review.opendev.org/c/openstack/kolla-ansible/+/746966 \n\n\u003e \n\u003e \n\u003e \u003e I then added this option to the affected services and problem dissapeared for all services.\n\u003e \u003e \n\u003e \u003e Anyway, I think this patch + keystone patch should be merged as is, I just wanted to \n\u003e \u003e react to Mike. Just imagine there is some network problem and clients are sending flush_all very often.\n\u003e \n\u003e \n\u003e yes, flush_all will erase all the keys. how does that lead to more memcached connections?\n\u003e \n\u003e \n\u003e \n\u003e \u003e \n\u003e \u003e I\u0027m adding also link to graph what I\u0027ve created\n\u003e \u003e \n\u003e \u003e https://ibb.co/7g2yRGX\n\u003e \u003e \n\u003e \u003e Legend :\n\u003e \u003e \n\u003e \u003e X -\u003e Cycles (Counter how much loops I\u0027ve iterated to API ^^ in Bash tests)\n\u003e \u003e Y -\u003e Number of connections to memcached (which were raising)\n\u003e \n\u003e \n\u003e this is..with a fix applied to neutron?  specifically the memcached_advanced_pool option?  is the test working with changing the flush_all option and seeing what effect that has?  Sorry I just am not getting all the context straight here.\n\nYes, graph is with memcached_advanced_pool \u003d True, other services with False. Yes, is working with flush_all (but no difference)\n\nHonestly, after amount of time spent on this issue I think that flush_all is not main problem, but keystone_middleware *IS* problem, on the other side don\u0027t think this option should be hardcoded and should be merged as is (also with keystone patch)","accounts_in_message":[],"_revision_number":12},{"id":"478d46e308bf755dbca5a2b5c0dd645533d6e7df","author":{"_account_id":27339,"name":"Michal Arbet","email":"michal.arbet@ultimum.io","username":"michalarbet"},"date":"2021-02-01 16:45:59.000000000","message":"Patch Set 12:\n\n\u003e Patch Set 12:\n\u003e \n\u003e \u003e Patch Set 12:\n\u003e \u003e \n\u003e \u003e hi and thanks for this reply.\n\u003e \u003e \n\u003e \u003e \n\u003e \u003e \u003e Parallely with above tests I was counting connections to memcached with script.\n\u003e \u003e \u003e From these data I created graph, and found that during the individual tests,\n\u003e \u003e \u003e the connections were raising, only with neutron and keystone were not raising.\n\u003e \u003e \n\u003e \u003e \n\u003e \u003e I\u0027m not following this, I thought the issue was that neutron *was* creating too many memcached connections.   Is this after you applied the pooling option or before?\n\u003e \n\u003e Well, graph is with  memcache_use_advanced_pool \u003d True , but when it was measured I really didn\u0027t know that neutron has set it to True. I just found it because i was trying to found differencies between services, after I found that kolla-ansible changed in https://review.opendev.org/c/openstack/kolla-ansible/+/746966 \n\u003e \n\u003e \u003e \n\u003e \u003e \n\u003e \u003e \u003e I then added this option to the affected services and problem dissapeared for all services.\n\u003e \u003e \u003e \n\u003e \u003e \u003e Anyway, I think this patch + keystone patch should be merged as is, I just wanted to \n\u003e \u003e \u003e react to Mike. Just imagine there is some network problem and clients are sending flush_all very often.\n\u003e \u003e \n\u003e \u003e \n\u003e \u003e yes, flush_all will erase all the keys. how does that lead to more memcached connections?\n\u003e \u003e \n\u003e \u003e \n\u003e \u003e \n\u003e \u003e \u003e \n\u003e \u003e \u003e I\u0027m adding also link to graph what I\u0027ve created\n\u003e \u003e \u003e \n\u003e \u003e \u003e https://ibb.co/7g2yRGX\n\u003e \u003e \u003e \n\u003e \u003e \u003e Legend :\n\u003e \u003e \u003e \n\u003e \u003e \u003e X -\u003e Cycles (Counter how much loops I\u0027ve iterated to API ^^ in Bash tests)\n\u003e \u003e \u003e Y -\u003e Number of connections to memcached (which were raising)\n\u003e \u003e \n\u003e \u003e \n\u003e \u003e this is..with a fix applied to neutron?  specifically the memcached_advanced_pool option?  is the test working with changing the flush_all option and seeing what effect that has?  Sorry I just am not getting all the context straight here.\n\u003e \n\u003e Yes, graph is with memcached_advanced_pool \u003d True, other services with False. Yes, is working with flush_all (but no difference)\n\u003e \n\u003e Honestly, after amount of time spent on this issue I think that flush_all is not main problem, but keystone_middleware *IS* problem, on the other side don\u0027t think this option should be hardcoded and should be merged as is (also with keystone patch)\n\nSorry for english :)\n\nShortly said:\n1. This should go as is with keystone patch\n2. Main problem *probably* is in keystone-middleware in function reserve() in _CachePool class which is not eventlet-safe","accounts_in_message":[],"_revision_number":12},{"id":"5bacfa292714990cb21baa3c8ef40a2e80c9c966","author":{"_account_id":28522,"name":"Hervé Beraud","email":"herveberaud.pro@gmail.com","username":"hberaud"},"date":"2021-02-01 17:59:56.000000000","message":"Patch Set 12:\n\n\u003e Patch Set 12:\n\u003e \n\u003e \u003e Patch Set 12:\n\u003e \u003e \n\u003e \u003e \u003e Patch Set 12:\n\u003e \u003e \u003e \n\u003e \u003e \u003e hi and thanks for this reply.\n\u003e \u003e \u003e \n\u003e \u003e \u003e \n\u003e \u003e \u003e \u003e Parallely with above tests I was counting connections to memcached with script.\n\u003e \u003e \u003e \u003e From these data I created graph, and found that during the individual tests,\n\u003e \u003e \u003e \u003e the connections were raising, only with neutron and keystone were not raising.\n\u003e \u003e \u003e \n\u003e \u003e \u003e \n\u003e \u003e \u003e I\u0027m not following this, I thought the issue was that neutron *was* creating too many memcached connections.   Is this after you applied the pooling option or before?\n\u003e \u003e \n\u003e \u003e Well, graph is with  memcache_use_advanced_pool \u003d True , but when it was measured I really didn\u0027t know that neutron has set it to True. I just found it because i was trying to found differencies between services, after I found that kolla-ansible changed in https://review.opendev.org/c/openstack/kolla-ansible/+/746966 \n\u003e \u003e \n\u003e \u003e \u003e \n\u003e \u003e \u003e \n\u003e \u003e \u003e \u003e I then added this option to the affected services and problem dissapeared for all services.\n\u003e \u003e \u003e \u003e \n\u003e \u003e \u003e \u003e Anyway, I think this patch + keystone patch should be merged as is, I just wanted to \n\u003e \u003e \u003e \u003e react to Mike. Just imagine there is some network problem and clients are sending flush_all very often.\n\u003e \u003e \u003e \n\u003e \u003e \u003e \n\u003e \u003e \u003e yes, flush_all will erase all the keys. how does that lead to more memcached connections?\n\u003e \u003e \u003e \n\u003e \u003e \u003e \n\u003e \u003e \u003e \n\u003e \u003e \u003e \u003e \n\u003e \u003e \u003e \u003e I\u0027m adding also link to graph what I\u0027ve created\n\u003e \u003e \u003e \u003e \n\u003e \u003e \u003e \u003e https://ibb.co/7g2yRGX\n\u003e \u003e \u003e \u003e \n\u003e \u003e \u003e \u003e Legend :\n\u003e \u003e \u003e \u003e \n\u003e \u003e \u003e \u003e X -\u003e Cycles (Counter how much loops I\u0027ve iterated to API ^^ in Bash tests)\n\u003e \u003e \u003e \u003e Y -\u003e Number of connections to memcached (which were raising)\n\u003e \u003e \u003e \n\u003e \u003e \u003e \n\u003e \u003e \u003e this is..with a fix applied to neutron?  specifically the memcached_advanced_pool option?  is the test working with changing the flush_all option and seeing what effect that has?  Sorry I just am not getting all the context straight here.\n\u003e \u003e \n\u003e \u003e Yes, graph is with memcached_advanced_pool \u003d True, other services with False. Yes, is working with flush_all (but no difference)\n\u003e \u003e \n\u003e \u003e Honestly, after amount of time spent on this issue I think that flush_all is not main problem, but keystone_middleware *IS* problem, on the other side don\u0027t think this option should be hardcoded and should be merged as is (also with keystone patch)\n\u003e \n\u003e Sorry for english :)\n\u003e \n\u003e Shortly said:\n\u003e 1. This should go as is with keystone patch\n\u003e 2. Main problem *probably* is in keystone-middleware in function reserve() in _CachePool class which is not eventlet-safe\n\nI agree, I don\u0027t know the reason behind the duplicated code (cache pool between keystone and oslo), but I think we should use only the oslo.cache module here.\n\nIf keystonemiddleware its implementation isn\u0027t eventlet safe it could explain why the connections continue to grow up.\n\nAlways concerning keystonemiddleware\u0027s implementation, I also I wonder how dead hosts are handled [1] for the connection pool within keystone\u0027s implementation, I didn\u0027t see related code, so, definitelly, I think we should give up this code and prefer oslo.cache\u0027s implementation.\n\nI also noticed duplicated memcached options, that are redefined in keystone and where a similar options are also given in oslo.cache, I don\u0027t know the reason behind this approach, but we (Ben and myself) have been a little confused by the whole conf file though [2][3]. Does Keystone not use the oslo.cache opts? If not, why?\n\nConcerning this patch, I think we could continue with that, make this feature optional can\u0027t hurt, however if flushing isn\u0027t our issue I think we should activate it by default as it help us to remove stale data. Thoughts?\n\n[1] https://opendev.org/openstack/keystonemiddleware/src/branch/master/keystonemiddleware/auth_token/_cache.py#L88-L105\n[2] https://review.opendev.org/c/openstack/keystone/+/772394\n[3] https://opendev.org/openstack/keystone/src/branch/master/keystone/conf/memcache.py","accounts_in_message":[],"_revision_number":12},{"id":"fa4cc46720a512c41b781aed0de134832ce29e43","author":{"_account_id":15334,"name":"Stephen Finucane","display_name":"stephenfin","email":"stephenfin@redhat.com","username":"sfinucan"},"date":"2021-02-03 11:21:37.000000000","message":"Patch Set 12: Code-Review+2 Workflow+1\n\n(2 comments)","accounts_in_message":[],"_revision_number":12},{"id":"69f401db70b923cc1e187bcddf9c1b568b0511e3","tag":"autogenerated:zuul:gate","author":{"_account_id":22348,"name":"Zuul","username":"zuul","tags":["SERVICE_USER"]},"date":"2021-02-03 11:22:04.000000000","message":"Patch Set 12: -Verified\n\nStarting gate jobs.","accounts_in_message":[],"_revision_number":12},{"id":"2687ce302f5b51f783a9e16e7d53e5919dc29cdb","tag":"autogenerated:zuul:gate","author":{"_account_id":22348,"name":"Zuul","username":"zuul","tags":["SERVICE_USER"]},"date":"2021-02-03 12:27:24.000000000","message":"Patch Set 12: Verified+2\n\nBuild succeeded (gate pipeline).\n\n- tempest-full-py3 https://zuul.opendev.org/t/openstack/build/9516745ddd0f4e67a8061df9981c22bd : SUCCESS in 1h 03m 33s\n- openstack-tox-pep8 https://zuul.opendev.org/t/openstack/build/fa08a182941849709d442aa64d4b2a8b : SUCCESS in 3m 16s\n- openstack-tox-py36 https://zuul.opendev.org/t/openstack/build/db30fd825040407b9aea2ee6da19bc62 : SUCCESS in 4m 41s\n- openstack-tox-py38 https://zuul.opendev.org/t/openstack/build/6b85af0391054d8c8b79ea751e19ec49 : SUCCESS in 3m 27s\n- openstack-tox-docs https://zuul.opendev.org/t/openstack/build/d16bf12e2e254515bec0c3afbd87c1cf : SUCCESS in 4m 58s\n- build-openstack-releasenotes https://zuul.opendev.org/t/openstack/build/892c0ca2672f4f90a6e6a1f6b64afcbb : SUCCESS in 5m 17s","accounts_in_message":[],"_revision_number":12},{"id":"09950bec17cda9d6ab7a37105e1df131d63ef64e","tag":"autogenerated:gerrit:merged","author":{"_account_id":22348,"name":"Zuul","username":"zuul","tags":["SERVICE_USER"]},"date":"2021-02-03 12:27:25.000000000","message":"Change has been successfully merged by Zuul","accounts_in_message":[],"_revision_number":12},{"id":"337dc0cb7ec762edf417160f1a992c7900b37470","tag":"autogenerated:zuul:promote","author":{"_account_id":22348,"name":"Zuul","username":"zuul","tags":["SERVICE_USER"]},"date":"2021-02-03 12:29:10.000000000","message":"Patch Set 12:\n\nBuild succeeded (promote pipeline).\n\n- promote-openstack-tox-docs https://zuul.opendev.org/t/openstack/build/1d424d40e2b34489b0ca5dbd182206fd : SUCCESS in 53s\n- promote-openstack-releasenotes https://zuul.opendev.org/t/openstack/build/b620c8ceec5248548a343e7d2f7da155 : SUCCESS in 59s","accounts_in_message":[],"_revision_number":12}],"current_revision_number":12,"current_revision":"a437b219acbf3ad092ecfb6f293a3eee133ec755","revisions":{"420ed5c5206b4da68350377334fb2db3877d124f":{"kind":"REWORK","_number":1,"created":"2020-07-21 13:42:33.000000000","uploader":{"_account_id":27339,"name":"Michal Arbet","email":"michal.arbet@ultimum.io","username":"michalarbet"},"ref":"refs/changes/93/742193/1","fetch":{"anonymous http":{"url":"https://review.opendev.org/openstack/oslo.cache","ref":"refs/changes/93/742193/1","commands":{"Checkout":"git fetch https://review.opendev.org/openstack/oslo.cache refs/changes/93/742193/1 \u0026\u0026 git checkout FETCH_HEAD","Cherry Pick":"git fetch https://review.opendev.org/openstack/oslo.cache refs/changes/93/742193/1 \u0026\u0026 git cherry-pick FETCH_HEAD","Format Patch":"git fetch https://review.opendev.org/openstack/oslo.cache refs/changes/93/742193/1 \u0026\u0026 git format-patch -1 --stdout FETCH_HEAD","Pull":"git pull https://review.opendev.org/openstack/oslo.cache refs/changes/93/742193/1"}}},"commit":{"parents":[{"commit":"1fe3c4638cabe988c63851b0acd56d32685bdbd9","subject":"Merge \"Introducing usage of pre-commit\"","web_links":[{"name":"gitea","tooltip":"Open in GitWeb","url":"https://opendev.org/openstack/oslo.cache/commit/1fe3c4638cabe988c63851b0acd56d32685bdbd9"}]}],"author":{"name":"Michal Arbet","email":"michal.arbet@ultimum.io","date":"2020-07-21 13:42:17.000000000","tz":120},"committer":{"name":"Michal Arbet","email":"michal.arbet@ultimum.io","date":"2020-07-21 13:42:17.000000000","tz":120},"subject":"Do not hardcode flush_on_reconnect, move to oslo.cache config","message":"Do not hardcode flush_on_reconnect, move to oslo.cache config\n\nParam flush_on_reconnect is very risky to use on production\ndeployments. It can cause exponential raising of connections\nto memcached servers. Moreover this option makes sense only\nin keystone\u0027s oslo.cache config.\n\nThis patch is moving flush_on_reconnect from code to oslo.cache\nconfig block to be configurable.\n\nChange-Id: I8e6826bfb2c85e7ceed03e1667bd6a06b3dff469\nCloses-Bug: #1888394\n","web_links":[{"name":"gitea","tooltip":"Open in GitWeb","url":"https://opendev.org/openstack/oslo.cache/commit/420ed5c5206b4da68350377334fb2db3877d124f"}],"resolve_conflicts_web_links":[{"name":"gitea","tooltip":"Open in GitWeb","url":"https://opendev.org/openstack/oslo.cache/commit/420ed5c5206b4da68350377334fb2db3877d124f"}]},"branch":"refs/heads/master"},"e43817046cd2174be28c4f02c7721890bc247dd3":{"kind":"REWORK","_number":2,"created":"2020-07-21 15:45:47.000000000","uploader":{"_account_id":27339,"name":"Michal Arbet","email":"michal.arbet@ultimum.io","username":"michalarbet"},"ref":"refs/changes/93/742193/2","fetch":{"anonymous http":{"url":"https://review.opendev.org/openstack/oslo.cache","ref":"refs/changes/93/742193/2","commands":{"Checkout":"git fetch https://review.opendev.org/openstack/oslo.cache refs/changes/93/742193/2 \u0026\u0026 git checkout FETCH_HEAD","Cherry Pick":"git fetch https://review.opendev.org/openstack/oslo.cache refs/changes/93/742193/2 \u0026\u0026 git cherry-pick FETCH_HEAD","Format Patch":"git fetch https://review.opendev.org/openstack/oslo.cache refs/changes/93/742193/2 \u0026\u0026 git format-patch -1 --stdout FETCH_HEAD","Pull":"git pull https://review.opendev.org/openstack/oslo.cache refs/changes/93/742193/2"}}},"commit":{"parents":[{"commit":"1fe3c4638cabe988c63851b0acd56d32685bdbd9","subject":"Merge \"Introducing usage of pre-commit\"","web_links":[{"name":"gitea","tooltip":"Open in GitWeb","url":"https://opendev.org/openstack/oslo.cache/commit/1fe3c4638cabe988c63851b0acd56d32685bdbd9"}]}],"author":{"name":"Michal Arbet","email":"michal.arbet@ultimum.io","date":"2020-07-21 13:42:17.000000000","tz":120},"committer":{"name":"Michal Arbet","email":"michal.arbet@ultimum.io","date":"2020-07-21 15:45:03.000000000","tz":120},"subject":"Do not hardcode flush_on_reconnect, move to oslo.cache config","message":"Do not hardcode flush_on_reconnect, move to oslo.cache config\n\nParam flush_on_reconnect is very risky to use on production\ndeployments. It can cause exponential raising of connections\nto memcached servers. Moreover this option makes sense only\nin keystone\u0027s oslo.cache config.\n\nThis patch is moving flush_on_reconnect from code to oslo.cache\nconfig block to be configurable.\n\nChange-Id: I8e6826bfb2c85e7ceed03e1667bd6a06b3dff469\nCloses-Bug: #1888394\n","web_links":[{"name":"gitea","tooltip":"Open in GitWeb","url":"https://opendev.org/openstack/oslo.cache/commit/e43817046cd2174be28c4f02c7721890bc247dd3"}],"resolve_conflicts_web_links":[{"name":"gitea","tooltip":"Open in GitWeb","url":"https://opendev.org/openstack/oslo.cache/commit/e43817046cd2174be28c4f02c7721890bc247dd3"}]},"branch":"refs/heads/master"},"cadee925dceb54c1f93c36f37eb2511f48e757b1":{"kind":"REWORK","_number":3,"created":"2020-07-22 15:12:24.000000000","uploader":{"_account_id":27339,"name":"Michal Arbet","email":"michal.arbet@ultimum.io","username":"michalarbet"},"ref":"refs/changes/93/742193/3","fetch":{"anonymous http":{"url":"https://review.opendev.org/openstack/oslo.cache","ref":"refs/changes/93/742193/3","commands":{"Checkout":"git fetch https://review.opendev.org/openstack/oslo.cache refs/changes/93/742193/3 \u0026\u0026 git checkout FETCH_HEAD","Cherry Pick":"git fetch https://review.opendev.org/openstack/oslo.cache refs/changes/93/742193/3 \u0026\u0026 git cherry-pick FETCH_HEAD","Format Patch":"git fetch https://review.opendev.org/openstack/oslo.cache refs/changes/93/742193/3 \u0026\u0026 git format-patch -1 --stdout FETCH_HEAD","Pull":"git pull https://review.opendev.org/openstack/oslo.cache refs/changes/93/742193/3"}}},"commit":{"parents":[{"commit":"1fe3c4638cabe988c63851b0acd56d32685bdbd9","subject":"Merge \"Introducing usage of pre-commit\"","web_links":[{"name":"gitea","tooltip":"Open in GitWeb","url":"https://opendev.org/openstack/oslo.cache/commit/1fe3c4638cabe988c63851b0acd56d32685bdbd9"}]}],"author":{"name":"Michal Arbet","email":"michal.arbet@ultimum.io","date":"2020-07-21 13:42:17.000000000","tz":120},"committer":{"name":"Michal Arbet","email":"michal.arbet@ultimum.io","date":"2020-07-22 15:11:37.000000000","tz":120},"subject":"Do not hardcode flush_on_reconnect, move to oslo.cache config","message":"Do not hardcode flush_on_reconnect, move to oslo.cache config\n\nParam flush_on_reconnect is very risky to use on production\ndeployments. It can cause exponential raising of connections\nto memcached servers. Moreover this option makes sense only\nin keystone\u0027s oslo.cache config.\n\nThis patch is moving flush_on_reconnect from code to oslo.cache\nconfig block to be configurable.\n\nChange-Id: I8e6826bfb2c85e7ceed03e1667bd6a06b3dff469\nCloses-Bug: #1888394\n","web_links":[{"name":"gitea","tooltip":"Open in GitWeb","url":"https://opendev.org/openstack/oslo.cache/commit/cadee925dceb54c1f93c36f37eb2511f48e757b1"}],"resolve_conflicts_web_links":[{"name":"gitea","tooltip":"Open in GitWeb","url":"https://opendev.org/openstack/oslo.cache/commit/cadee925dceb54c1f93c36f37eb2511f48e757b1"}]},"branch":"refs/heads/master"},"f9623853c3e38900308b3b4155fbb44ad9fbc4fb":{"kind":"REWORK","_number":4,"created":"2020-07-22 15:31:37.000000000","uploader":{"_account_id":27339,"name":"Michal Arbet","email":"michal.arbet@ultimum.io","username":"michalarbet"},"ref":"refs/changes/93/742193/4","fetch":{"anonymous http":{"url":"https://review.opendev.org/openstack/oslo.cache","ref":"refs/changes/93/742193/4","commands":{"Checkout":"git fetch https://review.opendev.org/openstack/oslo.cache refs/changes/93/742193/4 \u0026\u0026 git checkout FETCH_HEAD","Cherry Pick":"git fetch https://review.opendev.org/openstack/oslo.cache refs/changes/93/742193/4 \u0026\u0026 git cherry-pick FETCH_HEAD","Format Patch":"git fetch https://review.opendev.org/openstack/oslo.cache refs/changes/93/742193/4 \u0026\u0026 git format-patch -1 --stdout FETCH_HEAD","Pull":"git pull https://review.opendev.org/openstack/oslo.cache refs/changes/93/742193/4"}}},"commit":{"parents":[{"commit":"1fe3c4638cabe988c63851b0acd56d32685bdbd9","subject":"Merge \"Introducing usage of pre-commit\"","web_links":[{"name":"gitea","tooltip":"Open in GitWeb","url":"https://opendev.org/openstack/oslo.cache/commit/1fe3c4638cabe988c63851b0acd56d32685bdbd9"}]}],"author":{"name":"Michal Arbet","email":"michal.arbet@ultimum.io","date":"2020-07-21 13:42:17.000000000","tz":120},"committer":{"name":"Michal Arbet","email":"michal.arbet@ultimum.io","date":"2020-07-22 15:31:30.000000000","tz":120},"subject":"Do not hardcode flush_on_reconnect, move to oslo.cache config","message":"Do not hardcode flush_on_reconnect, move to oslo.cache config\n\nParam flush_on_reconnect is very risky to use on production\ndeployments. It can cause exponential raising of connections\nto memcached servers. Moreover this option makes sense only\nin keystone\u0027s oslo.cache config.\n\nThis patch is moving flush_on_reconnect from code to oslo.cache\nconfig block to be configurable.\n\nChange-Id: I8e6826bfb2c85e7ceed03e1667bd6a06b3dff469\nCloses-Bug: #1888394\n","web_links":[{"name":"gitea","tooltip":"Open in GitWeb","url":"https://opendev.org/openstack/oslo.cache/commit/f9623853c3e38900308b3b4155fbb44ad9fbc4fb"}],"resolve_conflicts_web_links":[{"name":"gitea","tooltip":"Open in GitWeb","url":"https://opendev.org/openstack/oslo.cache/commit/f9623853c3e38900308b3b4155fbb44ad9fbc4fb"}]},"branch":"refs/heads/master"},"624acb14c31e7c2dffb931c0fcaf715ea19c163d":{"kind":"REWORK","_number":5,"created":"2020-07-22 16:57:38.000000000","uploader":{"_account_id":27339,"name":"Michal Arbet","email":"michal.arbet@ultimum.io","username":"michalarbet"},"ref":"refs/changes/93/742193/5","fetch":{"anonymous http":{"url":"https://review.opendev.org/openstack/oslo.cache","ref":"refs/changes/93/742193/5","commands":{"Checkout":"git fetch https://review.opendev.org/openstack/oslo.cache refs/changes/93/742193/5 \u0026\u0026 git checkout FETCH_HEAD","Cherry Pick":"git fetch https://review.opendev.org/openstack/oslo.cache refs/changes/93/742193/5 \u0026\u0026 git cherry-pick FETCH_HEAD","Format Patch":"git fetch https://review.opendev.org/openstack/oslo.cache refs/changes/93/742193/5 \u0026\u0026 git format-patch -1 --stdout FETCH_HEAD","Pull":"git pull https://review.opendev.org/openstack/oslo.cache refs/changes/93/742193/5"}}},"commit":{"parents":[{"commit":"1fe3c4638cabe988c63851b0acd56d32685bdbd9","subject":"Merge \"Introducing usage of pre-commit\"","web_links":[{"name":"gitea","tooltip":"Open in GitWeb","url":"https://opendev.org/openstack/oslo.cache/commit/1fe3c4638cabe988c63851b0acd56d32685bdbd9"}]}],"author":{"name":"Michal Arbet","email":"michal.arbet@ultimum.io","date":"2020-07-21 13:42:17.000000000","tz":120},"committer":{"name":"Michal Arbet","email":"michal.arbet@ultimum.io","date":"2020-07-22 16:56:47.000000000","tz":0},"subject":"Do not hardcode flush_on_reconnect, move to oslo.cache config","message":"Do not hardcode flush_on_reconnect, move to oslo.cache config\n\nParam flush_on_reconnect is very risky to use on production\ndeployments. It can cause exponential raising of connections\nto memcached servers. Moreover this option makes sense only\nin keystone\u0027s oslo.cache config.\n\nThis patch is moving flush_on_reconnect from code to oslo.cache\nconfig block to be configurable.\n\nChange-Id: I8e6826bfb2c85e7ceed03e1667bd6a06b3dff469\nCloses-Bug: #1888394\n","web_links":[{"name":"gitea","tooltip":"Open in GitWeb","url":"https://opendev.org/openstack/oslo.cache/commit/624acb14c31e7c2dffb931c0fcaf715ea19c163d"}],"resolve_conflicts_web_links":[{"name":"gitea","tooltip":"Open in GitWeb","url":"https://opendev.org/openstack/oslo.cache/commit/624acb14c31e7c2dffb931c0fcaf715ea19c163d"}]},"branch":"refs/heads/master"},"a4b1f24e45a58f7f705f31c3b641ac4bb64e5b23":{"kind":"REWORK","_number":6,"created":"2020-07-27 11:02:21.000000000","uploader":{"_account_id":27339,"name":"Michal Arbet","email":"michal.arbet@ultimum.io","username":"michalarbet"},"ref":"refs/changes/93/742193/6","fetch":{"anonymous http":{"url":"https://review.opendev.org/openstack/oslo.cache","ref":"refs/changes/93/742193/6","commands":{"Checkout":"git fetch https://review.opendev.org/openstack/oslo.cache refs/changes/93/742193/6 \u0026\u0026 git checkout FETCH_HEAD","Cherry Pick":"git fetch https://review.opendev.org/openstack/oslo.cache refs/changes/93/742193/6 \u0026\u0026 git cherry-pick FETCH_HEAD","Format Patch":"git fetch https://review.opendev.org/openstack/oslo.cache refs/changes/93/742193/6 \u0026\u0026 git format-patch -1 --stdout FETCH_HEAD","Pull":"git pull https://review.opendev.org/openstack/oslo.cache refs/changes/93/742193/6"}}},"commit":{"parents":[{"commit":"1fe3c4638cabe988c63851b0acd56d32685bdbd9","subject":"Merge \"Introducing usage of pre-commit\"","web_links":[{"name":"gitea","tooltip":"Open in GitWeb","url":"https://opendev.org/openstack/oslo.cache/commit/1fe3c4638cabe988c63851b0acd56d32685bdbd9"}]}],"author":{"name":"Michal Arbet","email":"michal.arbet@ultimum.io","date":"2020-07-21 13:42:17.000000000","tz":120},"committer":{"name":"Michal Arbet","email":"michal.arbet@ultimum.io","date":"2020-07-27 11:01:58.000000000","tz":120},"subject":"Do not hardcode flush_on_reconnect, move to oslo.cache config","message":"Do not hardcode flush_on_reconnect, move to oslo.cache config\n\nParam flush_on_reconnect is very risky to use on production\ndeployments. It can cause exponential raising of connections\nto memcached servers. Moreover this option makes sense only\nin keystone\u0027s oslo.cache config.\n\nThis patch is moving flush_on_reconnect from code to oslo.cache\nconfig block to be configurable.\n\nChange-Id: I8e6826bfb2c85e7ceed03e1667bd6a06b3dff469\nCloses-Bug: #1888394\n","web_links":[{"name":"gitea","tooltip":"Open in GitWeb","url":"https://opendev.org/openstack/oslo.cache/commit/a4b1f24e45a58f7f705f31c3b641ac4bb64e5b23"}],"resolve_conflicts_web_links":[{"name":"gitea","tooltip":"Open in GitWeb","url":"https://opendev.org/openstack/oslo.cache/commit/a4b1f24e45a58f7f705f31c3b641ac4bb64e5b23"}]},"branch":"refs/heads/master"},"bd229ad7bc64f0b50611513f69d4e6b700147307":{"kind":"REWORK","_number":7,"created":"2020-07-27 11:42:24.000000000","uploader":{"_account_id":27339,"name":"Michal Arbet","email":"michal.arbet@ultimum.io","username":"michalarbet"},"ref":"refs/changes/93/742193/7","fetch":{"anonymous http":{"url":"https://review.opendev.org/openstack/oslo.cache","ref":"refs/changes/93/742193/7","commands":{"Checkout":"git fetch https://review.opendev.org/openstack/oslo.cache refs/changes/93/742193/7 \u0026\u0026 git checkout FETCH_HEAD","Cherry Pick":"git fetch https://review.opendev.org/openstack/oslo.cache refs/changes/93/742193/7 \u0026\u0026 git cherry-pick FETCH_HEAD","Format Patch":"git fetch https://review.opendev.org/openstack/oslo.cache refs/changes/93/742193/7 \u0026\u0026 git format-patch -1 --stdout FETCH_HEAD","Pull":"git pull https://review.opendev.org/openstack/oslo.cache refs/changes/93/742193/7"}}},"commit":{"parents":[{"commit":"1fe3c4638cabe988c63851b0acd56d32685bdbd9","subject":"Merge \"Introducing usage of pre-commit\"","web_links":[{"name":"gitea","tooltip":"Open in GitWeb","url":"https://opendev.org/openstack/oslo.cache/commit/1fe3c4638cabe988c63851b0acd56d32685bdbd9"}]}],"author":{"name":"Michal Arbet","email":"michal.arbet@ultimum.io","date":"2020-07-21 13:42:17.000000000","tz":120},"committer":{"name":"Michal Arbet","email":"michal.arbet@ultimum.io","date":"2020-07-27 11:42:15.000000000","tz":120},"subject":"Do not hardcode flush_on_reconnect, move to oslo.cache config","message":"Do not hardcode flush_on_reconnect, move to oslo.cache config\n\nParam flush_on_reconnect is very risky to use on production\ndeployments. It can cause exponential raising of connections\nto memcached servers. Moreover this option makes sense only\nin keystone\u0027s oslo.cache config.\n\nThis patch is moving flush_on_reconnect from code to oslo.cache\nconfig block to be configurable.\n\nChange-Id: I8e6826bfb2c85e7ceed03e1667bd6a06b3dff469\nCloses-Bug: #1888394\n","web_links":[{"name":"gitea","tooltip":"Open in GitWeb","url":"https://opendev.org/openstack/oslo.cache/commit/bd229ad7bc64f0b50611513f69d4e6b700147307"}],"resolve_conflicts_web_links":[{"name":"gitea","tooltip":"Open in GitWeb","url":"https://opendev.org/openstack/oslo.cache/commit/bd229ad7bc64f0b50611513f69d4e6b700147307"}]},"branch":"refs/heads/master"},"54dcb281ed343d6c2db0cf3a0ca04e7b59deb290":{"kind":"NO_CHANGE","_number":8,"created":"2020-07-27 15:29:44.000000000","uploader":{"_account_id":27339,"name":"Michal Arbet","email":"michal.arbet@ultimum.io","username":"michalarbet"},"ref":"refs/changes/93/742193/8","fetch":{"anonymous http":{"url":"https://review.opendev.org/openstack/oslo.cache","ref":"refs/changes/93/742193/8","commands":{"Checkout":"git fetch https://review.opendev.org/openstack/oslo.cache refs/changes/93/742193/8 \u0026\u0026 git checkout FETCH_HEAD","Cherry Pick":"git fetch https://review.opendev.org/openstack/oslo.cache refs/changes/93/742193/8 \u0026\u0026 git cherry-pick FETCH_HEAD","Format Patch":"git fetch https://review.opendev.org/openstack/oslo.cache refs/changes/93/742193/8 \u0026\u0026 git format-patch -1 --stdout FETCH_HEAD","Pull":"git pull https://review.opendev.org/openstack/oslo.cache refs/changes/93/742193/8"}}},"commit":{"parents":[{"commit":"1fe3c4638cabe988c63851b0acd56d32685bdbd9","subject":"Merge \"Introducing usage of pre-commit\"","web_links":[{"name":"gitea","tooltip":"Open in GitWeb","url":"https://opendev.org/openstack/oslo.cache/commit/1fe3c4638cabe988c63851b0acd56d32685bdbd9"}]}],"author":{"name":"Michal Arbet","email":"michal.arbet@ultimum.io","date":"2020-07-21 13:42:17.000000000","tz":120},"committer":{"name":"Michal Arbet","email":"michal.arbet@ultimum.io","date":"2020-07-27 15:29:34.000000000","tz":120},"subject":"Do not hardcode flush_on_reconnect, move to oslo.cache config","message":"Do not hardcode flush_on_reconnect, move to oslo.cache config\n\nParam flush_on_reconnect is very risky to use on production\ndeployments. It can cause exponential raising of connections\nto memcached servers. Moreover this option makes sense only\nin keystone\u0027s oslo.cache config.\n\nThis patch is moving flush_on_reconnect from code to oslo.cache\nconfig block to be configurable.\n\nChange-Id: I8e6826bfb2c85e7ceed03e1667bd6a06b3dff469\nCloses-Bug: #1888394\n","web_links":[{"name":"gitea","tooltip":"Open in GitWeb","url":"https://opendev.org/openstack/oslo.cache/commit/54dcb281ed343d6c2db0cf3a0ca04e7b59deb290"}],"resolve_conflicts_web_links":[{"name":"gitea","tooltip":"Open in GitWeb","url":"https://opendev.org/openstack/oslo.cache/commit/54dcb281ed343d6c2db0cf3a0ca04e7b59deb290"}]},"branch":"refs/heads/master"},"d9a9a0ef53dc4cdb8e0541d86f5881040c1ff91a":{"kind":"REWORK","_number":9,"created":"2020-07-27 17:19:26.000000000","uploader":{"_account_id":27339,"name":"Michal Arbet","email":"michal.arbet@ultimum.io","username":"michalarbet"},"ref":"refs/changes/93/742193/9","fetch":{"anonymous http":{"url":"https://review.opendev.org/openstack/oslo.cache","ref":"refs/changes/93/742193/9","commands":{"Checkout":"git fetch https://review.opendev.org/openstack/oslo.cache refs/changes/93/742193/9 \u0026\u0026 git checkout FETCH_HEAD","Cherry Pick":"git fetch https://review.opendev.org/openstack/oslo.cache refs/changes/93/742193/9 \u0026\u0026 git cherry-pick FETCH_HEAD","Format Patch":"git fetch https://review.opendev.org/openstack/oslo.cache refs/changes/93/742193/9 \u0026\u0026 git format-patch -1 --stdout FETCH_HEAD","Pull":"git pull https://review.opendev.org/openstack/oslo.cache refs/changes/93/742193/9"}}},"commit":{"parents":[{"commit":"1fe3c4638cabe988c63851b0acd56d32685bdbd9","subject":"Merge \"Introducing usage of pre-commit\"","web_links":[{"name":"gitea","tooltip":"Open in GitWeb","url":"https://opendev.org/openstack/oslo.cache/commit/1fe3c4638cabe988c63851b0acd56d32685bdbd9"}]}],"author":{"name":"Michal Arbet","email":"michal.arbet@ultimum.io","date":"2020-07-21 13:42:17.000000000","tz":120},"committer":{"name":"Michal Arbet","email":"michal.arbet@ultimum.io","date":"2020-07-27 17:19:19.000000000","tz":120},"subject":"Do not hardcode flush_on_reconnect, move to oslo.cache config","message":"Do not hardcode flush_on_reconnect, move to oslo.cache config\n\nParam flush_on_reconnect is very risky to use on production\ndeployments. It can cause exponential raising of connections\nto memcached servers. Moreover this option makes sense only\nin keystone\u0027s oslo.cache config.\n\nThis patch is moving flush_on_reconnect from code to oslo.cache\nconfig block to be configurable.\n\nChange-Id: I8e6826bfb2c85e7ceed03e1667bd6a06b3dff469\nCloses-Bug: #1888394\n","web_links":[{"name":"gitea","tooltip":"Open in GitWeb","url":"https://opendev.org/openstack/oslo.cache/commit/d9a9a0ef53dc4cdb8e0541d86f5881040c1ff91a"}],"resolve_conflicts_web_links":[{"name":"gitea","tooltip":"Open in GitWeb","url":"https://opendev.org/openstack/oslo.cache/commit/d9a9a0ef53dc4cdb8e0541d86f5881040c1ff91a"}]},"branch":"refs/heads/master"},"adff7fd47cce6c54dba2cbb464abe04e8c9be937":{"kind":"REWORK","_number":10,"created":"2021-01-14 06:32:05.000000000","uploader":{"_account_id":28522,"name":"Hervé Beraud","email":"herveberaud.pro@gmail.com","username":"hberaud"},"ref":"refs/changes/93/742193/10","fetch":{"anonymous http":{"url":"https://review.opendev.org/openstack/oslo.cache","ref":"refs/changes/93/742193/10","commands":{"Checkout":"git fetch https://review.opendev.org/openstack/oslo.cache refs/changes/93/742193/10 \u0026\u0026 git checkout FETCH_HEAD","Cherry Pick":"git fetch https://review.opendev.org/openstack/oslo.cache refs/changes/93/742193/10 \u0026\u0026 git cherry-pick FETCH_HEAD","Format Patch":"git fetch https://review.opendev.org/openstack/oslo.cache refs/changes/93/742193/10 \u0026\u0026 git format-patch -1 --stdout FETCH_HEAD","Pull":"git pull https://review.opendev.org/openstack/oslo.cache refs/changes/93/742193/10"}}},"commit":{"parents":[{"commit":"1fe3c4638cabe988c63851b0acd56d32685bdbd9","subject":"Merge \"Introducing usage of pre-commit\"","web_links":[{"name":"gitea","tooltip":"Open in GitWeb","url":"https://opendev.org/openstack/oslo.cache/commit/1fe3c4638cabe988c63851b0acd56d32685bdbd9"}]}],"author":{"name":"Michal Arbet","email":"michal.arbet@ultimum.io","date":"2020-07-21 13:42:17.000000000","tz":120},"committer":{"name":"Hervé Beraud","email":"hberaud@redhat.com","date":"2021-01-14 06:30:17.000000000","tz":60},"subject":"Do not hardcode flush_on_reconnect, move to oslo.cache config","message":"Do not hardcode flush_on_reconnect, move to oslo.cache config\n\nParam flush_on_reconnect is very risky to use on production\ndeployments. It can cause exponential raising of connections\nto memcached servers. Moreover this option makes sense only\nin keystone\u0027s oslo.cache config.\n\nThis patch is moving flush_on_reconnect from code to oslo.cache\nconfig block to be configurable.\n\nCo-Authored-By: Hervé Beraud \u003chberaud@redhat.com\u003e\nChange-Id: I8e6826bfb2c85e7ceed03e1667bd6a06b3dff469\nCloses-Bug: #1888394\n","web_links":[{"name":"gitea","tooltip":"Open in GitWeb","url":"https://opendev.org/openstack/oslo.cache/commit/adff7fd47cce6c54dba2cbb464abe04e8c9be937"}],"resolve_conflicts_web_links":[{"name":"gitea","tooltip":"Open in GitWeb","url":"https://opendev.org/openstack/oslo.cache/commit/adff7fd47cce6c54dba2cbb464abe04e8c9be937"}]},"branch":"refs/heads/master"},"7791ed01c8afe6bfd2082887f5c19e87b0b4f03a":{"kind":"REWORK","_number":11,"created":"2021-01-14 08:56:26.000000000","uploader":{"_account_id":28522,"name":"Hervé Beraud","email":"herveberaud.pro@gmail.com","username":"hberaud"},"ref":"refs/changes/93/742193/11","fetch":{"anonymous http":{"url":"https://review.opendev.org/openstack/oslo.cache","ref":"refs/changes/93/742193/11","commands":{"Checkout":"git fetch https://review.opendev.org/openstack/oslo.cache refs/changes/93/742193/11 \u0026\u0026 git checkout FETCH_HEAD","Cherry Pick":"git fetch https://review.opendev.org/openstack/oslo.cache refs/changes/93/742193/11 \u0026\u0026 git cherry-pick FETCH_HEAD","Format Patch":"git fetch https://review.opendev.org/openstack/oslo.cache refs/changes/93/742193/11 \u0026\u0026 git format-patch -1 --stdout FETCH_HEAD","Pull":"git pull https://review.opendev.org/openstack/oslo.cache refs/changes/93/742193/11"}}},"commit":{"parents":[{"commit":"1fe3c4638cabe988c63851b0acd56d32685bdbd9","subject":"Merge \"Introducing usage of pre-commit\"","web_links":[{"name":"gitea","tooltip":"Open in GitWeb","url":"https://opendev.org/openstack/oslo.cache/commit/1fe3c4638cabe988c63851b0acd56d32685bdbd9"}]}],"author":{"name":"Michal Arbet","email":"michal.arbet@ultimum.io","date":"2020-07-21 13:42:17.000000000","tz":120},"committer":{"name":"Hervé Beraud","email":"hberaud@redhat.com","date":"2021-01-14 08:56:16.000000000","tz":60},"subject":"Do not hardcode flush_on_reconnect, move to oslo.cache config","message":"Do not hardcode flush_on_reconnect, move to oslo.cache config\n\nParam flush_on_reconnect is very risky to use on production\ndeployments. It can cause exponential raising of connections\nto memcached servers. Moreover this option makes sense only\nin keystone\u0027s oslo.cache config.\n\nThis patch is moving flush_on_reconnect from code to oslo.cache\nconfig block to be configurable.\n\nCo-Authored-By: Hervé Beraud \u003chberaud@redhat.com\u003e\nChange-Id: I8e6826bfb2c85e7ceed03e1667bd6a06b3dff469\nCloses-Bug: #1888394\n","web_links":[{"name":"gitea","tooltip":"Open in GitWeb","url":"https://opendev.org/openstack/oslo.cache/commit/7791ed01c8afe6bfd2082887f5c19e87b0b4f03a"}],"resolve_conflicts_web_links":[{"name":"gitea","tooltip":"Open in GitWeb","url":"https://opendev.org/openstack/oslo.cache/commit/7791ed01c8afe6bfd2082887f5c19e87b0b4f03a"}]},"branch":"refs/heads/master"},"a437b219acbf3ad092ecfb6f293a3eee133ec755":{"kind":"TRIVIAL_REBASE","_number":12,"created":"2021-01-18 13:00:08.000000000","uploader":{"_account_id":28522,"name":"Hervé Beraud","email":"herveberaud.pro@gmail.com","username":"hberaud"},"ref":"refs/changes/93/742193/12","fetch":{"anonymous http":{"url":"https://review.opendev.org/openstack/oslo.cache","ref":"refs/changes/93/742193/12","commands":{"Checkout":"git fetch https://review.opendev.org/openstack/oslo.cache refs/changes/93/742193/12 \u0026\u0026 git checkout FETCH_HEAD","Cherry Pick":"git fetch https://review.opendev.org/openstack/oslo.cache refs/changes/93/742193/12 \u0026\u0026 git cherry-pick FETCH_HEAD","Format Patch":"git fetch https://review.opendev.org/openstack/oslo.cache refs/changes/93/742193/12 \u0026\u0026 git format-patch -1 --stdout FETCH_HEAD","Pull":"git pull https://review.opendev.org/openstack/oslo.cache refs/changes/93/742193/12"}}},"commit":{"parents":[{"commit":"bc9c70fd666f83feff77a3ac87ab82f9b450edac","subject":"Dropping lower constraints testing","web_links":[{"name":"gitea","tooltip":"Open in GitWeb","url":"https://opendev.org/openstack/oslo.cache/commit/bc9c70fd666f83feff77a3ac87ab82f9b450edac"}]}],"author":{"name":"Michal Arbet","email":"michal.arbet@ultimum.io","date":"2020-07-21 13:42:17.000000000","tz":120},"committer":{"name":"Hervé Beraud","email":"hberaud@redhat.com","date":"2021-01-18 12:59:55.000000000","tz":60},"subject":"Do not hardcode flush_on_reconnect, move to oslo.cache config","message":"Do not hardcode flush_on_reconnect, move to oslo.cache config\n\nParam flush_on_reconnect is very risky to use on production\ndeployments. It can cause exponential raising of connections\nto memcached servers. Moreover this option makes sense only\nin keystone\u0027s oslo.cache config.\n\nThis patch is moving flush_on_reconnect from code to oslo.cache\nconfig block to be configurable.\n\nCo-Authored-By: Hervé Beraud \u003chberaud@redhat.com\u003e\nChange-Id: I8e6826bfb2c85e7ceed03e1667bd6a06b3dff469\nCloses-Bug: #1888394\n","web_links":[{"name":"gitea","tooltip":"Open in GitWeb","url":"https://opendev.org/openstack/oslo.cache/commit/a437b219acbf3ad092ecfb6f293a3eee133ec755"}],"resolve_conflicts_web_links":[{"name":"gitea","tooltip":"Open in GitWeb","url":"https://opendev.org/openstack/oslo.cache/commit/a437b219acbf3ad092ecfb6f293a3eee133ec755"}]},"branch":"refs/heads/master"}},"requirements":[],"submit_records":[{"status":"CLOSED","labels":[{"label":"Verified","status":"OK","applied_by":{"_account_id":22348,"name":"Zuul","username":"zuul","tags":["SERVICE_USER"]}},{"label":"Code-Review","status":"OK","applied_by":{"_account_id":15334,"name":"Stephen Finucane","display_name":"stephenfin","email":"stephenfin@redhat.com","username":"sfinucan"}},{"label":"Workflow","status":"OK","applied_by":{"_account_id":15334,"name":"Stephen Finucane","display_name":"stephenfin","email":"stephenfin@redhat.com","username":"sfinucan"}},{"label":"Backport-Candidate","status":"MAY"}]}],"submit_requirements":[{"name":"Verified","status":"SATISFIED","is_legacy":true,"submittability_expression_result":{"expression":"label:Verified\u003dCustom-Rule","fulfilled":true,"status":"PASS","passing_atoms":["label:Verified\u003dCustom-Rule"],"failing_atoms":[],"atom_explanations":{}}},{"name":"Workflow","status":"SATISFIED","is_legacy":true,"submittability_expression_result":{"expression":"label:Workflow\u003dCustom-Rule","fulfilled":true,"status":"PASS","passing_atoms":["label:Workflow\u003dCustom-Rule"],"failing_atoms":[],"atom_explanations":{}}},{"name":"Code-Review","status":"SATISFIED","is_legacy":true,"submittability_expression_result":{"expression":"label:Code-Review\u003dCustom-Rule","fulfilled":true,"status":"PASS","passing_atoms":["label:Code-Review\u003dCustom-Rule"],"failing_atoms":[],"atom_explanations":{}}}]}
