)]}'
{".zuul.yaml":[{"author":{"_account_id":6873,"name":"Matt Riedemann","email":"mriedem.os@gmail.com","username":"mriedem"},"change_message_id":"e84ac48a3a0070d0ffd5a6c82f59a02afcc39fd7","unresolved":false,"context_lines":[{"line_number":18,"context_line":"      devstack_local_conf:"},{"line_number":19,"context_line":"        post-config:"},{"line_number":20,"context_line":"          $NOVA_CONF:"},{"line_number":21,"context_line":"            cache:"},{"line_number":22,"context_line":"              memcache_servers: \"localhost:11211\""},{"line_number":23,"context_line":"              backend: \"dogpile.cache.memcached\""},{"line_number":24,"context_line":"              enabled: \"True\""}],"source_content_type":"text/x-yaml","patch_set":1,"id":"7faddb67_fa7546fb","line":21,"updated":"2019-07-25 21:28:07.000000000","message":"Well this is surprisingly easy. How about multinode jobs? Oh I guess that\u0027s automatic via inheritance of the job vars:\n\nhttp://logs.openstack.org/15/672715/1/check/tempest-multinode-full-py3/6d1724e/controller/logs/etc/nova/nova_conf.txt.gz\n\nAnd is memcached already setup by devstack by default on all nodes? I guess so since it looks like keystone_authtoken is configured for it:\n\nhttp://logs.openstack.org/15/672715/1/check/tempest-full-py3/4737cce/controller/logs/etc/nova/nova_conf.txt.gz\n\n[keystone_authtoken]\nmemcached_servers \u003d localhost:11211\n\n[cache]\nenabled \u003d True\nmemcache_servers \u003d localhost:11211\nbackend \u003d dogpile.cache.memcached","commit_id":"494d0ded8fb283c4f0156790998ee904db29d5c0"},{"author":{"_account_id":6873,"name":"Matt Riedemann","email":"mriedem.os@gmail.com","username":"mriedem"},"change_message_id":"5f3ac533a08baa530273c2c9133b13d8934068b7","unresolved":false,"context_lines":[{"line_number":19,"context_line":"        post-config:"},{"line_number":20,"context_line":"          $NOVA_CONF:"},{"line_number":21,"context_line":"            cache:"},{"line_number":22,"context_line":"              memcache_servers: \"localhost:11211\""},{"line_number":23,"context_line":"              backend: \"dogpile.cache.memcached\""},{"line_number":24,"context_line":"              enabled: \"True\""},{"line_number":25,"context_line":"        test-config:"}],"source_content_type":"text/x-yaml","patch_set":1,"id":"7faddb67_ba94aee9","line":22,"range":{"start_line":22,"start_character":43,"end_line":22,"end_character":48},"updated":"2019-07-25 21:32:06.000000000","message":"This does kind of seem like something we shouldn\u0027t be hard-coded in tempest job config but rather pass a variable to devstack via localrc telling it to enable caching for nova and then could use whatever it\u0027s using for keystone, but that\u0027s a nit.","commit_id":"494d0ded8fb283c4f0156790998ee904db29d5c0"},{"author":{"_account_id":11604,"name":"sean mooney","email":"smooney@redhat.com","username":"sean-k-mooney"},"change_message_id":"4c46c9dc9a7c605cf3bcc4d62038ccaed9d10dec","unresolved":false,"context_lines":[{"line_number":19,"context_line":"        post-config:"},{"line_number":20,"context_line":"          $NOVA_CONF:"},{"line_number":21,"context_line":"            cache:"},{"line_number":22,"context_line":"              memcache_servers: \"localhost:11211\""},{"line_number":23,"context_line":"              backend: \"dogpile.cache.memcached\""},{"line_number":24,"context_line":"              enabled: \"True\""},{"line_number":25,"context_line":"        test-config:"}],"source_content_type":"text/x-yaml","patch_set":1,"id":"7faddb67_5c4c9775","line":22,"range":{"start_line":22,"start_character":43,"end_line":22,"end_character":48},"in_reply_to":"7faddb67_1a29429a","updated":"2019-07-25 21:52:46.000000000","message":"we install memcache by defualt for keystone\nhttps://github.com/openstack/devstack/blob/master/files/rpms/keystone\n\nand swift\nhttps://github.com/openstack/devstack/blob/master/files/rpms/swift\n\non the contolers where they run.\n\ni dont know if we install it on all host but if you are doing multi node devstack by hand like i normally do we would just sub out localhost for SERVICE_HOST which will work on all nodes and defulat to localhost on a singel node deployment.\n\nhttps://github.com/openstack/devstack/blob/master/stackrc#L876-L890\n\nand i guess a defualt memcached port too but ya this would be generally useful for devstack.\n\ni suspect that infra precache memcached in the gate images so its on all hosts.","commit_id":"494d0ded8fb283c4f0156790998ee904db29d5c0"},{"author":{"_account_id":11604,"name":"sean mooney","email":"smooney@redhat.com","username":"sean-k-mooney"},"change_message_id":"82065f84781f35521ea1dccf2da0112985f25433","unresolved":false,"context_lines":[{"line_number":19,"context_line":"        post-config:"},{"line_number":20,"context_line":"          $NOVA_CONF:"},{"line_number":21,"context_line":"            cache:"},{"line_number":22,"context_line":"              memcache_servers: \"localhost:11211\""},{"line_number":23,"context_line":"              backend: \"dogpile.cache.memcached\""},{"line_number":24,"context_line":"              enabled: \"True\""},{"line_number":25,"context_line":"        test-config:"}],"source_content_type":"text/x-yaml","patch_set":1,"id":"7faddb67_7c863368","line":22,"range":{"start_line":22,"start_character":43,"end_line":22,"end_character":48},"in_reply_to":"7faddb67_5c4c9775","updated":"2019-07-25 22:00:39.000000000","message":"11211 is the defualt memcached port by the way \n\nhttps://github.com/memcached/memcached/wiki/ConfiguringServer#networking","commit_id":"494d0ded8fb283c4f0156790998ee904db29d5c0"},{"author":{"_account_id":11975,"name":"Slawek Kaplonski","email":"skaplons@redhat.com","username":"slaweq"},"change_message_id":"57c12308429a627b34c9585129b6f39154034884","unresolved":false,"context_lines":[{"line_number":19,"context_line":"        post-config:"},{"line_number":20,"context_line":"          $NOVA_CONF:"},{"line_number":21,"context_line":"            cache:"},{"line_number":22,"context_line":"              memcache_servers: \"localhost:11211\""},{"line_number":23,"context_line":"              backend: \"dogpile.cache.memcached\""},{"line_number":24,"context_line":"              enabled: \"True\""},{"line_number":25,"context_line":"        test-config:"}],"source_content_type":"text/x-yaml","patch_set":1,"id":"7faddb67_458596ce","line":22,"range":{"start_line":22,"start_character":43,"end_line":22,"end_character":48},"in_reply_to":"7faddb67_7c863368","updated":"2019-07-26 10:21:01.000000000","message":"Ok, I can work on such patch for devstack to enable this caching in more user friendly way but if that is possible I would for now to go with this one and later do it in better way in devstack.\n\nI even opened bug for devstack about it https://bugs.launchpad.net/devstack/+bug/1837980 added todo here to change it when devstack will support that.\n\nWill it be ok for You?","commit_id":"494d0ded8fb283c4f0156790998ee904db29d5c0"},{"author":{"_account_id":4146,"name":"Clark Boylan","email":"cboylan@sapwetik.org","username":"cboylan"},"change_message_id":"1b31886527f1c6f642d0bcbd91ae15b383497267","unresolved":false,"context_lines":[{"line_number":19,"context_line":"        post-config:"},{"line_number":20,"context_line":"          $NOVA_CONF:"},{"line_number":21,"context_line":"            cache:"},{"line_number":22,"context_line":"              memcache_servers: \"localhost:11211\""},{"line_number":23,"context_line":"              backend: \"dogpile.cache.memcached\""},{"line_number":24,"context_line":"              enabled: \"True\""},{"line_number":25,"context_line":"        test-config:"}],"source_content_type":"text/x-yaml","patch_set":1,"id":"7faddb67_1a29429a","line":22,"range":{"start_line":22,"start_character":43,"end_line":22,"end_character":48},"in_reply_to":"7faddb67_ba94aee9","updated":"2019-07-25 21:44:02.000000000","message":"I would review and +2 that change in devstack :) In particular if memcache is already running then we could configure this for nova without any config by default too. Gives users a friendlier experience.","commit_id":"494d0ded8fb283c4f0156790998ee904db29d5c0"}],"/COMMIT_MSG":[{"author":{"_account_id":6873,"name":"Matt Riedemann","email":"mriedem.os@gmail.com","username":"mriedem"},"change_message_id":"e84ac48a3a0070d0ffd5a6c82f59a02afcc39fd7","unresolved":false,"context_lines":[{"line_number":25,"context_line":"for each worker so it will be cached always during processing"},{"line_number":26,"context_line":"instance-id request. And this request for instance-id is retried"},{"line_number":27,"context_line":"by CirrOS image in case of failure so this shoudn\u0027t cause problems."},{"line_number":28,"context_line":""},{"line_number":29,"context_line":"Change-Id: I8efedb8c5ae073bd2cb4034369bfcff4b67eb868"},{"line_number":30,"context_line":"Close-bug: #1836642"}],"source_content_type":"text/x-gerrit-commit-message","patch_set":1,"id":"7faddb67_dac96a25","line":28,"updated":"2019-07-25 21:28:07.000000000","message":"Yup, it works:\n\nhttp://logs.openstack.org/15/672715/1/check/tempest-multinode-full-py3/6d1724e/controller/logs/screen-n-api-meta.txt.gz#_Jul_25_19_06_51_867323\n\nJul 25 19:06:51.867323 ubuntu-bionic-rax-ord-0009341455 devstack@n-api-meta.service[25333]: DEBUG nova.api.metadata.handler [None req-c525ef28-fecf-422b-b19c-30780b81d7ba None None] Using cached metadata for instance a2857703-5af5-478b-ac4c-a367bb775039 {{(pid\u003d25335) get_metadata_by_instance_id /opt/stack/nova/nova/api/metadata/handler.py:79}}","commit_id":"494d0ded8fb283c4f0156790998ee904db29d5c0"},{"author":{"_account_id":6873,"name":"Matt Riedemann","email":"mriedem.os@gmail.com","username":"mriedem"},"change_message_id":"e84ac48a3a0070d0ffd5a6c82f59a02afcc39fd7","unresolved":false,"context_lines":[{"line_number":27,"context_line":"by CirrOS image in case of failure so this shoudn\u0027t cause problems."},{"line_number":28,"context_line":""},{"line_number":29,"context_line":"Change-Id: I8efedb8c5ae073bd2cb4034369bfcff4b67eb868"},{"line_number":30,"context_line":"Close-bug: #1836642"}],"source_content_type":"text/x-gerrit-commit-message","patch_set":1,"id":"7faddb67_9a8532d2","line":30,"updated":"2019-07-25 21:28:07.000000000","message":"Bug 1808010 would also be related to this. They are probably duplicates.","commit_id":"494d0ded8fb283c4f0156790998ee904db29d5c0"},{"author":{"_account_id":5689,"name":"Masayuki Igawa","email":"masayuki@igawa.io","username":"igawa"},"change_message_id":"25bbac7ef171ac3d905107194cd0b0dec154be75","unresolved":false,"context_lines":[{"line_number":7,"context_line":"Use memcached based cache in nova in all devstack-tempest jobs"},{"line_number":8,"context_line":""},{"line_number":9,"context_line":"Using memcached based cache for metadata api to cache data for"},{"line_number":10,"context_line":"instance always after firts metadata request is send can solve"},{"line_number":11,"context_line":"problems with unable to ssh to instance due to not configured"},{"line_number":12,"context_line":"public-key on instance."},{"line_number":13,"context_line":""}],"source_content_type":"text/x-gerrit-commit-message","patch_set":2,"id":"7faddb67_88714c0f","line":10,"updated":"2019-07-29 09:25:23.000000000","message":"first?","commit_id":"eb0a2cc5f240d52efa3a58c5a1ba8821bae3147e"},{"author":{"_account_id":8449,"name":"Marios Andreou","email":"marios.andreou@gmail.com","username":"marios"},"change_message_id":"02f05794a4521025e0d2cb97d346735e70ed7faf","unresolved":false,"context_lines":[{"line_number":28,"context_line":""},{"line_number":29,"context_line":"Change-Id: I8efedb8c5ae073bd2cb4034369bfcff4b67eb868"},{"line_number":30,"context_line":"Closes-bug: #1836642"},{"line_number":31,"context_line":"Closes-bug: #1808010"}],"source_content_type":"text/x-gerrit-commit-message","patch_set":2,"id":"7faddb67_c8cae46c","line":31,"updated":"2019-07-29 09:06:20.000000000","message":"heh so i found even _more_ bugs related to this :D including the one you have in your TODO comment ... added these to the tripleo one at https://bugs.launchpad.net/tripleo/+bug/1836046","commit_id":"eb0a2cc5f240d52efa3a58c5a1ba8821bae3147e"},{"author":{"_account_id":11975,"name":"Slawek Kaplonski","email":"skaplons@redhat.com","username":"slaweq"},"change_message_id":"5128d4e1187434458dd46cd109e6972f6215a5c2","unresolved":false,"context_lines":[{"line_number":28,"context_line":""},{"line_number":29,"context_line":"Change-Id: I8efedb8c5ae073bd2cb4034369bfcff4b67eb868"},{"line_number":30,"context_line":"Closes-bug: #1836642"},{"line_number":31,"context_line":"Closes-bug: #1808010"}],"source_content_type":"text/x-gerrit-commit-message","patch_set":2,"id":"7faddb67_488d942f","line":31,"in_reply_to":"7faddb67_c8cae46c","updated":"2019-07-29 09:09:14.000000000","message":"But I didn\u0027t add it here because this patch will not solve problem in TripleO based jobs, there will be needed similar patch to TripleO jobs too to do same thing probably.","commit_id":"eb0a2cc5f240d52efa3a58c5a1ba8821bae3147e"}]}
