)]}'
{"/COMMIT_MSG":[{"author":{"_account_id":4690,"name":"melanie witt","display_name":"melwitt","email":"melwittt@gmail.com","username":"melwitt"},"change_message_id":"9ed7d92879fce4ee2c7e6feac05abcd473c72bd1","unresolved":false,"context_lines":[{"line_number":6,"context_line":""},{"line_number":7,"context_line":"WIP: Try setting the bluestore block size"},{"line_number":8,"context_line":""},{"line_number":9,"context_line":"It looks like the default bluestore that came with nautilus"},{"line_number":10,"context_line":"is limiting us to a for-testing size of 10G regardless of what the"},{"line_number":11,"context_line":"actual ceph volume is supposed to be. Try setting it to the size"},{"line_number":12,"context_line":"we want (minus some room for overhead)."}],"source_content_type":"text/x-gerrit-commit-message","patch_set":1,"id":"9f560f44_e0ef18ae","line":9,"range":{"start_line":9,"start_character":51,"end_line":9,"end_character":59},"updated":"2020-07-24 18:27:54.000000000","message":"As sean-k-mooney noted [1], there was a change in the past to bump the default to 100G and that was backported to nautilus BUT we are pulling tag 14.2.2 which does not contain the backport.\n\n[1] http://eavesdrop.openstack.org/irclogs/%23openstack-nova/%23openstack-nova.2020-07-24.log.html#t2020-07-24T17:43:57","commit_id":"8f61f93c64fe2ec1eebd5f4b84575329f25987f6"},{"author":{"_account_id":4393,"name":"Dan Smith","email":"dms@danplanet.com","username":"danms"},"change_message_id":"545252571ccc9c8d729582a24740ea4583e14a07","unresolved":false,"context_lines":[{"line_number":6,"context_line":""},{"line_number":7,"context_line":"WIP: Try setting the bluestore block size"},{"line_number":8,"context_line":""},{"line_number":9,"context_line":"It looks like the default bluestore that came with nautilus"},{"line_number":10,"context_line":"is limiting us to a for-testing size of 10G regardless of what the"},{"line_number":11,"context_line":"actual ceph volume is supposed to be. Try setting it to the size"},{"line_number":12,"context_line":"we want (minus some room for overhead)."}],"source_content_type":"text/x-gerrit-commit-message","patch_set":1,"id":"9f560f44_60ee6889","line":9,"range":{"start_line":9,"start_character":51,"end_line":9,"end_character":59},"in_reply_to":"9f560f44_e0ef18ae","updated":"2020-07-24 18:34:09.000000000","message":"I was referring to the change from xfs to blue that came with nautilus. However, we\u0027re doing this on a loop mount of 24G, on a filesystem of no more than 80G. Assuming it\u0027s not creating a sparse file, we probably need to set this ourselves, even if it\u0027s smaller vs. larger in the future :)","commit_id":"8f61f93c64fe2ec1eebd5f4b84575329f25987f6"}],"devstack/lib/ceph":[{"author":{"_account_id":15334,"name":"Stephen Finucane","display_name":"stephenfin","email":"stephenfin@redhat.com","username":"sfinucan"},"change_message_id":"cb7fb3bc16ce8427921bd6050ff56f3b5aca884f","unresolved":false,"context_lines":[{"line_number":429,"context_line":""},{"line_number":430,"context_line":"    local gigs"},{"line_number":431,"context_line":"    gigs\u003d$(echo $CEPH_LOOPBACK_DISK_SIZE | grep -o \u0027^[0-9]*\u0027)"},{"line_number":432,"context_line":"    iniset -sudo ${CEPH_CONF_FILE} global \"bluestore_block_size\" $((($gigs - 4) \u003c\u003c 30))"},{"line_number":433,"context_line":""},{"line_number":434,"context_line":"    # bootstrap the ceph monitor"},{"line_number":435,"context_line":"    sudo ceph-mon -c ${CEPH_CONF_FILE} --mkfs -i $(hostname) \\"}],"source_content_type":"application/x-shellscript","patch_set":3,"id":"9f560f44_32e31436","line":432,"range":{"start_line":432,"start_character":68,"end_line":432,"end_character":79},"updated":"2020-07-27 10:23:14.000000000","message":"I guess it\u0027s safe to assume that this will always be \u003e 0 (or that ceph will simply crash and burn in a helpful manner if not)","commit_id":"5a0a36f1591d71c0a95a6f04e80d5c503c958646"},{"author":{"_account_id":11604,"name":"sean mooney","email":"smooney@redhat.com","username":"sean-k-mooney"},"change_message_id":"d53999fb118b548aa543d9c2794974a80c603dc9","unresolved":false,"context_lines":[{"line_number":429,"context_line":""},{"line_number":430,"context_line":"    local gigs"},{"line_number":431,"context_line":"    gigs\u003d$(echo $CEPH_LOOPBACK_DISK_SIZE | grep -o \u0027^[0-9]*\u0027)"},{"line_number":432,"context_line":"    iniset -sudo ${CEPH_CONF_FILE} global \"bluestore_block_size\" $((($gigs - 4) \u003c\u003c 30))"},{"line_number":433,"context_line":""},{"line_number":434,"context_line":"    # bootstrap the ceph monitor"},{"line_number":435,"context_line":"    sudo ceph-mon -c ${CEPH_CONF_FILE} --mkfs -i $(hostname) \\"}],"source_content_type":"application/x-shellscript","patch_set":3,"id":"9f560f44_92dca019","line":432,"range":{"start_line":432,"start_character":68,"end_line":432,"end_character":79},"in_reply_to":"9f560f44_32e31436","updated":"2020-07-27 10:46:42.000000000","message":"devstacks defaul will make CEPH_LOOPBACK_DISK_SIZE 24G\nand the plugins default would make it 8G even if you used it without devstack so ya i think we can just assume that if you set those to 4 or less that ceph will be unhappy.\n\ni dont really think its worth overcomplicating this since the pluging is just for testing anyway.\n\nno one is running devstack with ceph in production outside of ci.","commit_id":"5a0a36f1591d71c0a95a6f04e80d5c503c958646"}]}
