Orit,

Here is what i've done just now:

root@arh-ibstorage1-ib:~# service [email protected] stop

(the above command was ran on both radosgw servers). Checked with ps and no 
radosgw services were running. After that I've done:



root@arh-ibstorage1-ib:~# ./ceph-zones-fix.sh
+ RADOSGW_ADMIN=radosgw-admin
+ echo Exercise initialization code
Exercise initialization code
+ radosgw-admin user info --uid=foo
could not fetch user info: no user info saved
+ echo Get default zonegroup
Get default zonegroup
+ radosgw-admin zonegroup get --rgw-zonegroup=default
+ sed s/"id":.*/"id": "default",/g
+ sed s/"master_zone.*/"master_zone": "default",/g
+ echo Get default zone
Get default zone
+ radosgw-admin zone get --zone-id=default
+ echo Creating realm
Creating realm
+ radosgw-admin realm create --rgw-realm=london-ldex
ERROR: couldn't create realm london-ldex: (17) File exists
2016-11-10 13:44:48.872839 7f87a13d9a00  0 ERROR creating new realm object 
london-ldex: (17) File exists
+ echo Creating default zonegroup
Creating default zonegroup
+ radosgw-admin zonegroup set --rgw-zonegroup=default
{
    "id": "default",
    "name": "default",
    "api_name": "",
    "is_master": "true",
    "endpoints": [],
    "hostnames": [],
    "hostnames_s3website": [],
    "master_zone": "default",
    "zones": [
        {
            "id": "default",
            "name": "default",
            "endpoints": [],
            "log_meta": "false",
            "log_data": "false",
            "bucket_index_max_shards": 0,
            "read_only": "false"
        }
    ],
    "placement_targets": [
        {
            "name": "default-placement",
            "tags": []
        }
    ],
    "default_placement": "default-placement",
    "realm_id": "5b41b1b2-0f92-463d-b582-07552f83e66c"
}
+ echo Creating default zone
Creating default zone
+ radosgw-admin zone set --rgw-zone=default
zone id default{
    "id": "default",
    "name": "default",
    "domain_root": ".rgw",
    "control_pool": ".rgw.control",
    "gc_pool": ".rgw.gc",
    "log_pool": ".log",
    "intent_log_pool": ".intent-log",
    "usage_log_pool": ".usage",
    "user_keys_pool": ".users",
    "user_email_pool": ".users.email",
    "user_swift_pool": ".users.swift",
    "user_uid_pool": ".users.uid",
    "system_key": {
        "access_key": "",
        "secret_key": ""
    },  
    "placement_pools": [
        {
            "key": "default-placement",
            "val": {
                "index_pool": ".rgw.buckets.index",
                "data_pool": ".rgw.buckets",
                "data_extra_pool": "default.rgw.buckets.non-ec",
                "index_type": 0
            }   
        }   
    ],  
    "metadata_heap": ".rgw.meta",
    "realm_id": "5b41b1b2-0f92-463d-b582-07552f83e66c"
}   
+ echo Setting default zonegroup to 'default'
Setting default zonegroup to 'default'
+ radosgw-admin zonegroup default --rgw-zonegroup=default
+ echo Setting default zone to 'default'
Setting default zone to 'default'
+ radosgw-admin zone default --rgw-zone=default






After that i've ran the following to make sure the details have been updated:




root@arh-ibstorage1-ib:~# radosgw-admin zone get --rgw-zone=default
{
    "id": "default",
    "name": "default",
    "domain_root": ".rgw",
    "control_pool": ".rgw.control",
    "gc_pool": ".rgw.gc",
    "log_pool": ".log",
    "intent_log_pool": ".intent-log",
    "usage_log_pool": ".usage",
    "user_keys_pool": ".users",
    "user_email_pool": ".users.email",
    "user_swift_pool": ".users.swift",
    "user_uid_pool": ".users.uid",
    "system_key": {
        "access_key": "",
        "secret_key": ""
    },
    "placement_pools": [
        {   
            "key": "default-placement",
            "val": {
                "index_pool": ".rgw.buckets.index",
                "data_pool": ".rgw.buckets",
                "data_extra_pool": "default.rgw.buckets.non-ec",
                "index_type": 0
            }
        }
    ],
    "metadata_heap": ".rgw.meta",
    "realm_id": "5b41b1b2-0f92-463d-b582-07552f83e66c"
}



root@arh-ibstorage1-ib:~# radosgw-admin zonegroup get --rgw-zonegroup=default
{
    "id": "default",
    "name": "default",
    "api_name": "",
    "is_master": "true",
    "endpoints": [],
    "hostnames": [],
    "hostnames_s3website": [],
    "master_zone": "default",
    "zones": [
        {
            "id": "default",
            "name": "default",
            "endpoints": [],
            "log_meta": "false",
            "log_data": "false",
            "bucket_index_max_shards": 0,
            "read_only": "false"
        }
    ],
    "placement_targets": [
        {
            "name": "default-placement",
            "tags": []
        }
    ],
    "default_placement": "default-placement",
    "realm_id": "5b41b1b2-0f92-463d-b582-07552f83e66c"
}




As far as I can see, the master_zone is now set to default.

Now I start the radosgw service:


root@arh-ibstorage1-ib:~# service [email protected] start
root@arh-ibstorage1-ib:~# 
root@arh-ibstorage1-ib:~# 
root@arh-ibstorage1-ib:~# 
root@arh-ibstorage1-ib:~# 
root@arh-ibstorage1-ib:~# 
root@arh-ibstorage1-ib:~# radosgw-admin zone get --rgw-zone=default
{
    "id": "default",
    "name": "default",
    "domain_root": ".rgw",
    "control_pool": ".rgw.control",
    "gc_pool": ".rgw.gc",
    "log_pool": ".log",
    "intent_log_pool": ".intent-log",
    "usage_log_pool": ".usage",
    "user_keys_pool": ".users",
    "user_email_pool": ".users.email",
    "user_swift_pool": ".users.swift",
    "user_uid_pool": ".users.uid",
    "system_key": {
        "access_key": "",
        "secret_key": ""
    },
    "placement_pools": [
        {
            "key": "default-placement",
            "val": {
                "index_pool": ".rgw.buckets.index",
                "data_pool": ".rgw.buckets",
                "data_extra_pool": "default.rgw.buckets.non-ec",
                "index_type": 0
            }
        }
    ],
    "metadata_heap": ".rgw.meta",
    "realm_id": "5b41b1b2-0f92-463d-b582-07552f83e66c"


root@arh-ibstorage1-ib:~# radosgw-admin zonegroup get --rgw-zonegroup=default
{
    "id": "default",
    "name": "default",
    "api_name": "",
    "is_master": "true",
    "endpoints": [],
    "hostnames": [],
    "hostnames_s3website": [],
    "master_zone": "",
    "zones": [
        {
            "id": "default",
            "name": "default",
            "endpoints": [],
            "log_meta": "false",
            "log_data": "false",
            "bucket_index_max_shards": 0,
            "read_only": "false"
        }
    ],
    "placement_targets": [
        {
            "name": "default-placement",
            "tags": []
        }
    ],
    "default_placement": "default-placement",
    "realm_id": ""
}






As you can see from above, the realm_id and the master_zone have reverted back 
to being blank! Crazy!!!


Andrei














----- Original Message -----
> From: "Orit Wasserman" <[email protected]>
> To: "Andrei Mikhailovsky" <[email protected]>
> Cc: "Yoann Moulin" <[email protected]>, "ceph-users" 
> <[email protected]>
> Sent: Thursday, 10 November, 2016 13:36:01
> Subject: Re: [ceph-users] radosgw - http status 400 while creating a bucket

> On Thu, Nov 10, 2016 at 2:24 PM, Andrei Mikhailovsky <[email protected]> 
> wrote:
>>
>> Hi Orit,
>>
>> Thanks for the links.
>>
>> I've had a look at the link that you've sent
>> http://lists.ceph.com/pipermail/ceph-users-ceph.com/2016-July/011157.html and
>> followed the instructions. Created the script as depicted in the email. 
>> Changed
>> the realm name to something relevant. The script ran without any errors. I've
>> restarted radosgw services on both servers, but I am still unable to create
>> buckets. I am getting exactly the same error from the client:
>>
> 
> So you have two radosgw running on different servers, on the same zone
> (default) and on the same rados cluster?
> Are both gateways from the same version?
> Did you shutdown both gateway when you run the script?
> 
>>
>> S3ResponseError: 400 Bad Request
>> <?xml version="1.0"
>> encoding="UTF-8"?><Error><Code>InvalidArgument</Code><BucketName>my-new-bucket-31337</BucketName><RequestId>tx000000000000000000003-00582472a4-995ee8c-default</RequestId><HostId>995ee8c-default-default</HostId></Error>
>>
>>
>> I can delete a bucket, but I can't create a new one.
>>
>>
>> What I did notice is that after running the script and getting the zonegroup
>> info, I do see both the master_zone and the realm_id fields are set:
>>
>>
>>
>>     "master_zone": "default",
>>     "realm_id": "5b41b1b2-0f92-463d-b582-07552f83e66c"
>>
>>
>> However, after I restart radosgw service, they go back to being blank:
>>
>>
>>     "master_zone": "",
>>     "realm_id": ""
>>
>>
>> In any case, creation of buckets doesn't work either way.
>>
>> Cheers
>>
>>
>> Andrei
>>
>>
>> ----- Original Message -----
>>> From: "Orit Wasserman" <[email protected]>
>>> To: "Yoann Moulin" <[email protected]>
>>> Cc: "ceph-users" <[email protected]>
>>> Sent: Thursday, 10 November, 2016 10:04:46
>>> Subject: Re: [ceph-users] radosgw - http status 400 while creating a bucket
>>
>>> On Wed, Nov 9, 2016 at 10:20 PM, Yoann Moulin <[email protected]> wrote:
>>>> Hello,
>>>>
>>>>> many thanks for your help. I've tried setting the zone to master, 
>>>>> followed by
>>>>> the period update --commit command. This is what i've had:
>>>>
>>>> maybe it's related to this issue :
>>>>
>>>> http://tracker.ceph.com/issues/16839 (fixe in Jewel 10.2.3)
>>>>
>>>> or this one :
>>>>
>>>> http://tracker.ceph.com/issues/17239
>>>>
>>>> the "id" of the zonegroup shouldn't be "default" but an uuid afaik
>>>>
>>>> Best regards
>>>>
>>>> Yoann Moulin
>>>>
>>>>> root@arh-ibstorage1-ib:~# radosgw-admin zonegroup get 
>>>>> --rgw-zonegroup=default
>>>>> {
>>>>>     "id": "default",
>>>>>     "name": "default",
>>>>>     "api_name": "",
>>>>>     "is_master": "true",
>>>>>     "endpoints": [],
>>>>>     "hostnames": [],
>>>>>     "hostnames_s3website": [],
>>>>>     "master_zone": "default",
>>>>>     "zones": [
>>>>>         {
>>>>>             "id": "default",
>>>>>             "name": "default",
>>>>>             "endpoints": [],
>>>>>             "log_meta": "false",
>>>>>             "log_data": "false",
>>>>>             "bucket_index_max_shards": 0,
>>>>>             "read_only": "false"
>>>>>         }
>>>>>     ],
>>>>>     "placement_targets": [
>>>>>         {
>>>>>             "name": "default-placement",
>>>>>             "tags": []
>>>>>         }
>>>>>     ],
>>>>>     "default_placement": "default-placement",
>>>>>     "realm_id": "5b41b1b2-0f92-463d-b582-07552f83e66c"
>>>>> }
>>>>>
>>>>>
>>>>> root@arh-ibstorage1-ib:~# radosgw-admin period update --commit
>>>>> cannot commit period: period does not have a master zone of a master 
>>>>> zonegroup
>>>>> failed to commit period: (22) Invalid argument
>>>>>
>>>>>
>>>
>>> This is issue http://tracker.ceph.com/issues/17364 (I am working on it
>>> at he moment).
>>>
>>> Do the same procedure as before without the period update --commit command,
>>> it should fix the master zone problem.
>>> see also
>>> http://lists.ceph.com/pipermail/ceph-users-ceph.com/2016-July/011157.html
>>>
>>> This looks like an upgraded system (the id equals the name after an 
>>> upgrade).
>>>
>>> Orit
>>>
>>>>> root@arh-ibstorage1-ib:~# radosgw-admin zonegroup get 
>>>>> --rgw-zonegroup=default
>>>>> {
>>>>>     "id": "default",
>>>>>     "name": "default",
>>>>>     "api_name": "",
>>>>>     "is_master": "true",
>>>>>     "endpoints": [],
>>>>>     "hostnames": [],
>>>>>     "hostnames_s3website": [],
>>>>>     "master_zone": "",
>>>>>     "zones": [
>>>>>         {
>>>>>             "id": "default",
>>>>>             "name": "default",
>>>>>             "endpoints": [],
>>>>>             "log_meta": "false",
>>>>>             "log_data": "false",
>>>>>             "bucket_index_max_shards": 0,
>>>>>             "read_only": "false"
>>>>>         }
>>>>>     ],
>>>>>     "placement_targets": [
>>>>>         {
>>>>>             "name": "default-placement",
>>>>>             "tags": []
>>>>>         }
>>>>>     ],
>>>>>     "default_placement": "default-placement",
>>>>>     "realm_id": ""
>>>>> }
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>> The strange thing as you can see, following the "radosgw-admin period 
>>>>> update
>>>>> --commit" command, the master_zone and the realm_id values reset to 
>>>>> blank. What
>>>>> could be causing this?
>>>>>
>>>>> Here is my ceph infrastructure setup, perhaps it will help with finding 
>>>>> the
>>>>> issue?:
>>>>>
>>>>> ceph osd and mon servers:
>>>>> arh-ibstorage1-ib (also radosgw server)
>>>>> arh-ibstorage2-ib (also radosgw server)
>>>>> arh-ibstorage3-ib
>>>>>
>>>>> ceph mon server:
>>>>> arh-cloud13-ib
>>>>>
>>>>>
>>>>>
>>>>> Thus, overall, i have 4 mon servers, 3 osd servers and 2 radosgw servers
>>>>>
>>>>> Thanks
>>>>>
>>>>>
>>>>>
>>>>> ----- Original Message -----
>>>>>> From: "Yehuda Sadeh-Weinraub" <[email protected]>
>>>>>> To: "Andrei Mikhailovsky" <[email protected]>
>>>>>> Cc: "ceph-users" <[email protected]>
>>>>>> Sent: Wednesday, 9 November, 2016 17:12:30
>>>>>> Subject: Re: [ceph-users] radosgw - http status 400 while creating a 
>>>>>> bucket
>>>>>
>>>>>> On Wed, Nov 9, 2016 at 1:30 AM, Andrei Mikhailovsky <[email protected]> 
>>>>>> wrote:
>>>>>>> Hi Yehuda,
>>>>>>>
>>>>>>> just tried to run the command to set the master_zone to default 
>>>>>>> followed by the
>>>>>>> bucket create without doing the restart and I still have the same error 
>>>>>>> on the
>>>>>>> client:
>>>>>>>
>>>>>>> <?xml version="1.0"
>>>>>>> encoding="UTF-8"?><Error><Code>InvalidArgument</Code><BucketName>my-new-bucket-31337</BucketName><RequestId>tx000000000000000000010-005822ebbd-9951ad8-default</RequestId><HostId>9951ad8-default-default</HostId></Error>
>>>>>>>
>>>>>>
>>>>>> After setting the master zone, try running:
>>>>>>
>>>>>> $ radosgw-admin period update --commit
>>>>>>
>>>>>> Yehuda
>>>>>>
>>>>>>>
>>>>>>> Andrei
>>>>>>>
>>>>>>> ----- Original Message -----
>>>>>>>> From: "Yehuda Sadeh-Weinraub" <[email protected]>
>>>>>>>> To: "Andrei Mikhailovsky" <[email protected]>
>>>>>>>> Cc: "ceph-users" <[email protected]>
>>>>>>>> Sent: Wednesday, 9 November, 2016 01:13:48
>>>>>>>> Subject: Re: [ceph-users] radosgw - http status 400 while creating a 
>>>>>>>> bucket
>>>>>>>
>>>>>>>> On Tue, Nov 8, 2016 at 5:05 PM, Andrei Mikhailovsky 
>>>>>>>> <[email protected]> wrote:
>>>>>>>>> Hi Yehuda,
>>>>>>>>>
>>>>>>>>> I don't have a multizone setup. The radosgw service was configured 
>>>>>>>>> about two
>>>>>>>>> years ago according to the documentation on ceph.com and haven't 
>>>>>>>>> changed with
>>>>>>>>> numerous version updates. All was working okay until i've upgraded to 
>>>>>>>>> version
>>>>>>>>> 10.2.x.
>>>>>>>>>
>>>>>>>>> Could you please point me in the right direction what exactly needs 
>>>>>>>>> to be done?
>>>>>>>>>
>>>>>>>>>
>>>>>>>>> # radosgw-admin zonegroup get --rgw-zonegroup=default
>>>>>>>>> {
>>>>>>>>>     "id": "default",
>>>>>>>>>     "name": "default",
>>>>>>>>>     "api_name": "",
>>>>>>>>>     "is_master": "true",
>>>>>>>>>     "endpoints": [],
>>>>>>>>>     "hostnames": [],
>>>>>>>>>     "hostnames_s3website": [],
>>>>>>>>>     "master_zone": "",
>>>>>>>>>     "zones": [
>>>>>>>>>         {
>>>>>>>>>             "id": "default",
>>>>>>>>>             "name": "default",
>>>>>>>>>             "endpoints": [],
>>>>>>>>>             "log_meta": "false",
>>>>>>>>>             "log_data": "false",
>>>>>>>>>             "bucket_index_max_shards": 0,
>>>>>>>>>             "read_only": "false"
>>>>>>>>>         }
>>>>>>>>>     ],
>>>>>>>>>     "placement_targets": [
>>>>>>>>>         {
>>>>>>>>>             "name": "default-placement",
>>>>>>>>>             "tags": []
>>>>>>>>>         }
>>>>>>>>>     ],
>>>>>>>>>     "default_placement": "default-placement",
>>>>>>>>>     "realm_id": ""
>>>>>>>>> }
>>>>>>>>
>>>>>>>> Try:
>>>>>>>>
>>>>>>>> $ radosgw-admin zonegroup get --rgw-zonegroup=default > zonegroup.json
>>>>>>>>
>>>>>>>> ... modify the master_zone to be "default"
>>>>>>>>
>>>>>>>> $ radosgw-admin zonegroup set --rgw-zonegroup=default < zonegroup.json
>>>>>>>>
>>>>>>>> (restart radosgw)
>>>>>>>>
>>>>>>>> Yehuda
>>>>>>>>
>>>>>>>>>
>>>>>>>>>
>>>>>>>>>
>>>>>>>>> # radosgw-admin zone get --rgw-zone=default
>>>>>>>>> {
>>>>>>>>>     "id": "default",
>>>>>>>>>     "name": "default",
>>>>>>>>>     "domain_root": ".rgw",
>>>>>>>>>     "control_pool": ".rgw.control",
>>>>>>>>>     "gc_pool": ".rgw.gc",
>>>>>>>>>     "log_pool": ".log",
>>>>>>>>>     "intent_log_pool": ".intent-log",
>>>>>>>>>     "usage_log_pool": ".usage",
>>>>>>>>>     "user_keys_pool": ".users",
>>>>>>>>>     "user_email_pool": ".users.email",
>>>>>>>>>     "user_swift_pool": ".users.swift",
>>>>>>>>>     "user_uid_pool": ".users.uid",
>>>>>>>>>     "system_key": {
>>>>>>>>>         "access_key": "",
>>>>>>>>>         "secret_key": ""
>>>>>>>>>     },
>>>>>>>>>     "placement_pools": [
>>>>>>>>>         {
>>>>>>>>>             "key": "default-placement",
>>>>>>>>>             "val": {
>>>>>>>>>                 "index_pool": ".rgw.buckets.index",
>>>>>>>>>                 "data_pool": ".rgw.buckets",
>>>>>>>>>                 "data_extra_pool": "default.rgw.buckets.non-ec",
>>>>>>>>>                 "index_type": 0
>>>>>>>>>             }
>>>>>>>>>         }
>>>>>>>>>     ],
>>>>>>>>>     "metadata_heap": ".rgw.meta",
>>>>>>>>>     "realm_id": "5b41b1b2-0f92-463d-b582-07552f83e66c"
>>>>>>>>> }
>>>>>>>>>
>>>>>>>>>
>>>>>>>>>
>>>>>>>>> Thanks
>>>>>>>>>
>>>>>>>>>
>>>>>>>>>
>>>>>>>>> ----- Original Message -----
>>>>>>>>>> From: "Yehuda Sadeh-Weinraub" <[email protected]>
>>>>>>>>>> To: "Andrei Mikhailovsky" <[email protected]>
>>>>>>>>>> Cc: "ceph-users" <[email protected]>
>>>>>>>>>> Sent: Wednesday, 9 November, 2016 00:48:50
>>>>>>>>>> Subject: Re: [ceph-users] radosgw - http status 400 while creating a 
>>>>>>>>>> bucket
>>>>>>>>>
>>>>>>>>>> On Tue, Nov 8, 2016 at 3:36 PM, Andrei Mikhailovsky 
>>>>>>>>>> <[email protected]> wrote:
>>>>>>>>>>> Hello
>>>>>>>>>>>
>>>>>>>>>>> I am having issues with creating buckets in radosgw. It started 
>>>>>>>>>>> with an
>>>>>>>>>>> upgrade to version 10.2.x
>>>>>>>>>>>
>>>>>>>>>>> When I am creating a bucket I get the following error on the client 
>>>>>>>>>>> side:
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>> boto.exception.S3ResponseError: S3ResponseError: 400 Bad Request
>>>>>>>>>>> <?xml version="1.0"
>>>>>>>>>>> encoding="UTF-8"?><Error><Code>InvalidArgument</Code><BucketName>my-new-bucket-31337</BucketName><RequestId>tx000000000000000000002-0058225bae-994d148-default</RequestId><HostId>994d148-default-default</HostId></Error>
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>> The radosgw logs are (redacted):
>>>>>>>>>>>
>>>>>>>>>>> ###############################
>>>>>>>>>>>
>>>>>>>>>>> 2016-11-08 23:11:42.876862 7f026d953700 20 enqueued request
>>>>>>>>>>> req=0x7f02ba07b0e0
>>>>>>>>>>> 2016-11-08 23:11:42.876892 7f026d953700 20 RGWWQ:
>>>>>>>>>>> 2016-11-08 23:11:42.876897 7f026d953700 20 req: 0x7f02ba07b0e0
>>>>>>>>>>> 2016-11-08 23:11:42.876912 7f026d953700 10 allocated request
>>>>>>>>>>> req=0x7f02ba07b140
>>>>>>>>>>> 2016-11-08 23:11:42.876975 7f026b94f700 20 dequeued request
>>>>>>>>>>> req=0x7f02ba07b0e0
>>>>>>>>>>> 2016-11-08 23:11:42.876987 7f026b94f700 20 RGWWQ: empty
>>>>>>>>>>> 2016-11-08 23:11:42.877050 7f026b94f700 20 CONTENT_LENGTH=0
>>>>>>>>>>> 2016-11-08 23:11:42.877060 7f026b94f700 20 
>>>>>>>>>>> CONTEXT_DOCUMENT_ROOT=/var/www
>>>>>>>>>>> 2016-11-08 23:11:42.877062 7f026b94f700 20 CONTEXT_PREFIX=
>>>>>>>>>>> 2016-11-08 23:11:42.877063 7f026b94f700 20 DOCUMENT_ROOT=/var/www
>>>>>>>>>>> 2016-11-08 23:11:42.877081 7f026b94f700 20 FCGI_ROLE=RESPONDER
>>>>>>>>>>> 2016-11-08 23:11:42.877083 7f026b94f700 20 GATEWAY_INTERFACE=CGI/1.1
>>>>>>>>>>> 2016-11-08 23:11:42.877084 7f026b94f700 20 
>>>>>>>>>>> HTTP_ACCEPT_ENCODING=identity
>>>>>>>>>>> 2016-11-08 23:11:42.877086 7f026b94f700 20 HTTP_AUTHORIZATION=AWS
>>>>>>>>>>> XXXXXXXEDITEDXXXXXX:EDITEDXXXXXXXeWyiacaN26GcME
>>>>>>>>>>> 2016-11-08 23:11:42.877087 7f026b94f700 20 HTTP_DATE=Tue, 08 Nov 
>>>>>>>>>>> 2016
>>>>>>>>>>> 23:11:37 GMT
>>>>>>>>>>> 2016-11-08 23:11:42.877088 7f026b94f700 20
>>>>>>>>>>> HTTP_HOST=s3service.editedname.com
>>>>>>>>>>> 2016-11-08 23:11:42.877089 7f026b94f700 20 
>>>>>>>>>>> HTTP_USER_AGENT=Boto/2.38.0
>>>>>>>>>>> Python/2.7.12 Linux/4.8.4-040804-generic
>>>>>>>>>>> 2016-11-08 23:11:42.877090 7f026b94f700 20 HTTPS=on
>>>>>>>>>>> 2016-11-08 23:11:42.877092 7f026b94f700 20
>>>>>>>>>>> PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin
>>>>>>>>>>> 2016-11-08 23:11:42.877093 7f026b94f700 20 proxy-nokeepalive=1
>>>>>>>>>>> 2016-11-08 23:11:42.877094 7f026b94f700 20 QUERY_STRING=
>>>>>>>>>>> 2016-11-08 23:11:42.877095 7f026b94f700 20 
>>>>>>>>>>> REMOTE_ADDR=192.168.169.91
>>>>>>>>>>> 2016-11-08 23:11:42.877096 7f026b94f700 20 REMOTE_PORT=45404
>>>>>>>>>>> 2016-11-08 23:11:42.877097 7f026b94f700 20 REQUEST_METHOD=PUT
>>>>>>>>>>> 2016-11-08 23:11:42.877098 7f026b94f700 20 REQUEST_SCHEME=https
>>>>>>>>>>> 2016-11-08 23:11:42.877099 7f026b94f700 20 
>>>>>>>>>>> REQUEST_URI=/my-new-bucket-31337/
>>>>>>>>>>> 2016-11-08 23:11:42.877104 7f026b94f700 20
>>>>>>>>>>> SCRIPT_FILENAME=proxy:fcgi://localhost:9000/my-new-bucket-31337/
>>>>>>>>>>> 2016-11-08 23:11:42.877105 7f026b94f700 20 
>>>>>>>>>>> SCRIPT_NAME=/my-new-bucket-31337/
>>>>>>>>>>> 2016-11-08 23:11:42.877107 7f026b94f700 20
>>>>>>>>>>> SCRIPT_URI=https://s3service.editedname.com/my-new-bucket-31337/
>>>>>>>>>>> 2016-11-08 23:11:42.877108 7f026b94f700 20 
>>>>>>>>>>> SCRIPT_URL=/my-new-bucket-31337/
>>>>>>>>>>> 2016-11-08 23:11:42.877109 7f026b94f700 20 
>>>>>>>>>>> SERVER_ADDR=192.168.169.201
>>>>>>>>>>> 2016-11-08 23:11:42.877110 7f026b94f700 20
>>>>>>>>>>> [email protected]
>>>>>>>>>>> 2016-11-08 23:11:42.877111 7f026b94f700 20
>>>>>>>>>>> SERVER_NAME=s3service.editedname.com
>>>>>>>>>>> 2016-11-08 23:11:42.877112 7f026b94f700 20 SERVER_PORT=443
>>>>>>>>>>> 2016-11-08 23:11:42.877113 7f026b94f700 20 SERVER_PROTOCOL=HTTP/1.1
>>>>>>>>>>> 2016-11-08 23:11:42.877114 7f026b94f700 20 SERVER_SIGNATURE=
>>>>>>>>>>> 2016-11-08 23:11:42.877115 7f026b94f700 20 
>>>>>>>>>>> SERVER_SOFTWARE=Apache/2.4.18
>>>>>>>>>>> (Ubuntu)
>>>>>>>>>>> 2016-11-08 23:11:42.877116 7f026b94f700 20
>>>>>>>>>>> SSL_TLS_SNI=s3service.editedname.com
>>>>>>>>>>> 2016-11-08 23:11:42.877119 7f026b94f700  1 ====== starting new 
>>>>>>>>>>> request
>>>>>>>>>>> req=0x7f02ba07b0e0 =====
>>>>>>>>>>> 2016-11-08 23:11:42.877155 7f026b94f700  2 req 2:0.000035::PUT
>>>>>>>>>>> /my-new-bucket-31337/::initializing for trans_id =
>>>>>>>>>>> tx000000000000000000002-0058225bae-994d148-default
>>>>>>>>>>> 2016-11-08 23:11:42.877175 7f026b94f700 10 rgw api priority: s3=5
>>>>>>>>>>> s3website=4
>>>>>>>>>>> 2016-11-08 23:11:42.877179 7f026b94f700 10 
>>>>>>>>>>> host=s3service.editedname.com
>>>>>>>>>>> 2016-11-08 23:11:42.877199 7f026b94f700 20 subdomain=
>>>>>>>>>>> domain=s3service.editedname.com in_hosted_domain=1
>>>>>>>>>>> in_hosted_domain_s3website=0
>>>>>>>>>>> 2016-11-08 23:11:42.877203 7f026b94f700 20 final domain/bucket 
>>>>>>>>>>> subdomain=
>>>>>>>>>>> domain=s3service.editedname.com in_hosted_domain=1
>>>>>>>>>>> in_hosted_domain_s3website=0 s->info.domain=s3service.editedname.com
>>>>>>>>>>> s->info.request_uri=/my-new-bucket-31337/
>>>>>>>>>>> 2016-11-08 23:11:42.877277 7f026b94f700 20 get_handler
>>>>>>>>>>> handler=25RGWHandler_REST_Bucket_S3
>>>>>>>>>>> 2016-11-08 23:11:42.877286 7f026b94f700 10
>>>>>>>>>>> handler=25RGWHandler_REST_Bucket_S3
>>>>>>>>>>> 2016-11-08 23:11:42.877291 7f026b94f700  2 req 2:0.000172:s3:PUT
>>>>>>>>>>> /my-new-bucket-31337/::getting op 1
>>>>>>>>>>> 2016-11-08 23:11:42.877326 7f026b94f700 10 
>>>>>>>>>>> op=27RGWCreateBucket_ObjStore_S3
>>>>>>>>>>> 2016-11-08 23:11:42.877334 7f026b94f700  2 req 2:0.000215:s3:PUT
>>>>>>>>>>> /my-new-bucket-31337/:create_bucket:authorizing
>>>>>>>>>>> 2016-11-08 23:11:42.877386 7f026b94f700 20 get_system_obj_state:
>>>>>>>>>>> rctx=0x7f026b94b7c0 obj=.users:XXXXXXXEDITEDXXXXXX 
>>>>>>>>>>> state=0x7f02b70cdfe8
>>>>>>>>>>> s->prefetch_data=0
>>>>>>>>>>> 2016-11-08 23:11:42.877403 7f026b94f700 10 cache get:
>>>>>>>>>>> name=.users+XXXXXXXEDITEDXXXXXX : miss
>>>>>>>>>>> 2016-11-08 23:11:42.877516 7f026b94f700  1 -- 
>>>>>>>>>>> 192.168.168.201:0/3703211654
>>>>>>>>>>> --> 192.168.168.201:6829/893 -- osd_op(client.160747848.0:1019 
>>>>>>>>>>> 16.e7d6c03f
>>>>>>>>>>> XXXXXXXEDITEDXXXXXX [getxattrs,stat] snapc 0=[] 
>>>>>>>>>>> ack+read+known_if_red
>>>>>>>>>>> irected e97148) v7 -- ?+0 0x7f02b70ab600 con 0x7f02c581a800
>>>>>>>>>>> 2016-11-08 23:11:42.879265 7f02e9405700  1 -- 
>>>>>>>>>>> 192.168.168.201:0/3703211654
>>>>>>>>>>> <== osd.3 192.168.168.201:6829/893 52 ==== osd_op_reply(1019
>>>>>>>>>>> XXXXXXXEDITEDXXXXXX [getxattrs,stat] v0'0 uv7 ondisk = 0) v7 ==== 
>>>>>>>>>>> 182+0+20
>>>>>>>>>>> (2521936
>>>>>>>>>>> 738 0 3070622072) 0x7f02c9897280 con 0x7f02c581a800
>>>>>>>>>>> 2016-11-08 23:11:42.879391 7f026b94f700 10 cache put:
>>>>>>>>>>> name=.users+XXXXXXXEDITEDXXXXXX info.flags=6
>>>>>>>>>>> 2016-11-08 23:11:42.879421 7f026b94f700 10 adding 
>>>>>>>>>>> .users+XXXXXXXEDITEDXXXXXX
>>>>>>>>>>> to cache LRU end
>>>>>>>>>>> 2016-11-08 23:11:42.879442 7f026b94f700 20 get_system_obj_state: 
>>>>>>>>>>> s->obj_tag
>>>>>>>>>>> was set empty
>>>>>>>>>>> 2016-11-08 23:11:42.879452 7f026b94f700 10 cache get:
>>>>>>>>>>> name=.users+XXXXXXXEDITEDXXXXXX : type miss (requested=1, cached=6)
>>>>>>>>>>> 2016-11-08 23:11:42.879461 7f026b94f700 20 get_system_obj_state:
>>>>>>>>>>> rctx=0x7f026b94b7c0 obj=.users:XXXXXXXEDITEDXXXXXX 
>>>>>>>>>>> state=0x7f02b70cdfe8
>>>>>>>>>>> s->prefetch_data=0
>>>>>>>>>>> 2016-11-08 23:11:42.879480 7f026b94f700 20 rados->read ofs=0 
>>>>>>>>>>> len=524288
>>>>>>>>>>> 2016-11-08 23:11:42.879533 7f026b94f700  1 -- 
>>>>>>>>>>> 192.168.168.201:0/3703211654
>>>>>>>>>>> --> 192.168.168.201:6829/893 -- osd_op(client.160747848.0:1020 
>>>>>>>>>>> 16.e7d6c03f
>>>>>>>>>>> XXXXXXXEDITEDXXXXXX [read 0~524288] snapc 0=[] 
>>>>>>>>>>> ack+read+known_if_redi
>>>>>>>>>>> rected e97148) v7 -- ?+0 0x7f02b70ab980 con 0x7f02c581a800
>>>>>>>>>>> 2016-11-08 23:11:42.880343 7f02e9405700  1 -- 
>>>>>>>>>>> 192.168.168.201:0/3703211654
>>>>>>>>>>> <== osd.3 192.168.168.201:6829/893 53 ==== osd_op_reply(1020
>>>>>>>>>>> XXXXXXXEDITEDXXXXXX [read 0~37] v0'0 uv7 ondisk = 0) v7 ==== 
>>>>>>>>>>> 140+0+37
>>>>>>>>>>> (3791206703 0
>>>>>>>>>>>  1897156770) 0x7f02c9897280 con 0x7f02c581a800
>>>>>>>>>>> 2016-11-08 23:11:42.880433 7f026b94f700 20 rados->read r=0 
>>>>>>>>>>> bl.length=37
>>>>>>>>>>> 2016-11-08 23:11:42.880464 7f026b94f700 10 cache put:
>>>>>>>>>>> name=.users+XXXXXXXEDITEDXXXXXX info.flags=1
>>>>>>>>>>> 2016-11-08 23:11:42.880476 7f026b94f700 10 moving 
>>>>>>>>>>> .users+XXXXXXXEDITEDXXXXXX
>>>>>>>>>>> to cache LRU end
>>>>>>>>>>> 2016-11-08 23:11:42.880509 7f026b94f700 20 get_system_obj_state:
>>>>>>>>>>> rctx=0x7f026b94b3f0 obj=.users.uid:EDITED - client name 
>>>>>>>>>>> state=0x7f02b70cede8
>>>>>>>>>>> s->prefetch_data=0
>>>>>>>>>>> 2016-11-08 23:11:42.880519 7f026b94f700 10 cache get: 
>>>>>>>>>>> name=.users.uid+EDITED
>>>>>>>>>>> - client name : miss
>>>>>>>>>>> 2016-11-08 23:11:42.880587 7f026b94f700  1 -- 
>>>>>>>>>>> 192.168.168.201:0/3703211654
>>>>>>>>>>> --> 192.168.168.203:6825/25725 -- osd_op(client.160747848.0:1021 
>>>>>>>>>>> 15.d14b4318
>>>>>>>>>>> EDITED - client name [call version.read,getxattrs,stat]
>>>>>>>>>>>  snapc 0=[] ack+read+known_if_redirected e97148) v7 -- ?+0 
>>>>>>>>>>> 0x7f02b70abd00
>>>>>>>>>>> con 0x7f02b9057180
>>>>>>>>>>> 2016-11-08 23:11:42.882156 7f02bbefe700  1 -- 
>>>>>>>>>>> 192.168.168.201:0/3703211654
>>>>>>>>>>> <== osd.27 192.168.168.203:6825/25725 33 ==== osd_op_reply(1021 
>>>>>>>>>>> EDITED -
>>>>>>>>>>> client name [call,getxattrs,stat] v0'0 uv14016 ondisk = 0)
>>>>>>>>>>> v7 ==== 237+0+139 (1797997075 0 3786209825) 0x7f02bc847600 con
>>>>>>>>>>> 0x7f02b9057180
>>>>>>>>>>> 2016-11-08 23:11:42.882322 7f026b94f700 10 cache put: 
>>>>>>>>>>> name=.users.uid+EDITED
>>>>>>>>>>> - client name info.flags=22
>>>>>>>>>>> 2016-11-08 23:11:42.882353 7f026b94f700 10 adding .users.uid+EDITED 
>>>>>>>>>>> - client
>>>>>>>>>>> name to cache LRU end
>>>>>>>>>>> 2016-11-08 23:11:42.882363 7f026b94f700 20 get_system_obj_state: 
>>>>>>>>>>> s->obj_tag
>>>>>>>>>>> was set empty
>>>>>>>>>>> 2016-11-08 23:11:42.882374 7f026b94f700 10 cache get: 
>>>>>>>>>>> name=.users.uid+EDITED
>>>>>>>>>>> - client name : type miss (requested=17, cached=22)
>>>>>>>>>>> 2016-11-08 23:11:42.882383 7f026b94f700 20 get_system_obj_state:
>>>>>>>>>>> rctx=0x7f026b94b3f0 obj=.users.uid:EDITED - client name 
>>>>>>>>>>> state=0x7f02b70cede8
>>>>>>>>>>> s->prefetch_data=0
>>>>>>>>>>> 2016-11-08 23:11:42.882427 7f026b94f700 20 rados->read ofs=0 
>>>>>>>>>>> len=524288
>>>>>>>>>>> 2016-11-08 23:11:42.882492 7f026b94f700  1 -- 
>>>>>>>>>>> 192.168.168.201:0/3703211654
>>>>>>>>>>> --> 192.168.168.203:6825/25725 -- osd_op(client.160747848.0:1022 
>>>>>>>>>>> 15.d14b4318
>>>>>>>>>>> EDITED - client name [call version.check_conds,call ver
>>>>>>>>>>> sion.read,read 0~524288] snapc 0=[] ack+read+known_if_redirected 
>>>>>>>>>>> e97148) v7
>>>>>>>>>>> -- ?+0 0x7f02b70ac400 con 0x7f02b9057180
>>>>>>>>>>> 2016-11-08 23:11:42.883550 7f02bbefe700  1 -- 
>>>>>>>>>>> 192.168.168.201:0/3703211654
>>>>>>>>>>> <== osd.27 192.168.168.203:6825/25725 34 ==== osd_op_reply(1022 
>>>>>>>>>>> EDITED -
>>>>>>>>>>> client name [call,call,read 0~401] v0'0 uv14016 ondisk = 0)
>>>>>>>>>>>  v7 ==== 237+0+449 (131137144 0 372490823) 0x7f02bc847980 con 
>>>>>>>>>>> 0x7f02b9057180
>>>>>>>>>>> 2016-11-08 23:11:42.883653 7f026b94f700 20 rados->read r=0 
>>>>>>>>>>> bl.length=401
>>>>>>>>>>> 2016-11-08 23:11:42.883695 7f026b94f700 10 cache put: 
>>>>>>>>>>> name=.users.uid+EDITED
>>>>>>>>>>> - client name info.flags=17
>>>>>>>>>>> 2016-11-08 23:11:42.883698 7f026b94f700 10 moving .users.uid+EDITED 
>>>>>>>>>>> - client
>>>>>>>>>>> name to cache LRU end
>>>>>>>>>>> 2016-11-08 23:11:42.883726 7f026b94f700 10 chain_cache_entry:
>>>>>>>>>>> cache_locator=.users.uid+EDITED - client name
>>>>>>>>>>> 2016-11-08 23:11:42.883777 7f026b94f700 10 get_canon_resource():
>>>>>>>>>>> dest=/my-new-bucket-31337/
>>>>>>>>>>> 2016-11-08 23:11:42.883782 7f026b94f700 10 auth_hdr:
>>>>>>>>>>> PUT
>>>>>>>>>>>
>>>>>>>>>>> Tue, 08 Nov 2016 23:11:37 GMT
>>>>>>>>>>> /my-new-bucket-31337/
>>>>>>>>>>> 2016-11-08 23:11:42.883849 7f026b94f700 15 calculated
>>>>>>>>>>> digest=EDITEDXXXXXXXeWyiacaN26GcME
>>>>>>>>>>> 2016-11-08 23:11:42.883852 7f026b94f700 15
>>>>>>>>>>> auth_sign=EDITEDXXXXXXXeWyiacaN26GcME
>>>>>>>>>>> 2016-11-08 23:11:42.883853 7f026b94f700 15 compare=0
>>>>>>>>>>> 2016-11-08 23:11:42.883858 7f026b94f700  2 req 2:0.006739:s3:PUT
>>>>>>>>>>> /my-new-bucket-31337/:create_bucket:normalizing buckets and tenants
>>>>>>>>>>> 2016-11-08 23:11:42.883864 7f026b94f700 10 s->object=<NULL>
>>>>>>>>>>> s->bucket=my-new-bucket-31337
>>>>>>>>>>> 2016-11-08 23:11:42.883869 7f026b94f700  2 req 2:0.006750:s3:PUT
>>>>>>>>>>> /my-new-bucket-31337/:create_bucket:init permissions
>>>>>>>>>>> 2016-11-08 23:11:42.883872 7f026b94f700  2 req 2:0.006753:s3:PUT
>>>>>>>>>>> /my-new-bucket-31337/:create_bucket:recalculating target
>>>>>>>>>>> 2016-11-08 23:11:42.883875 7f026b94f700  2 req 2:0.006756:s3:PUT
>>>>>>>>>>> /my-new-bucket-31337/:create_bucket:reading permissions
>>>>>>>>>>> 2016-11-08 23:11:42.883879 7f026b94f700  2 req 2:0.006760:s3:PUT
>>>>>>>>>>> /my-new-bucket-31337/:create_bucket:init op
>>>>>>>>>>> 2016-11-08 23:11:42.883884 7f026b94f700  2 req 2:0.006765:s3:PUT
>>>>>>>>>>> /my-new-bucket-31337/:create_bucket:verifying op mask
>>>>>>>>>>> 2016-11-08 23:11:42.883886 7f026b94f700 20 required_mask= 2 
>>>>>>>>>>> user.op_mask=7
>>>>>>>>>>> 2016-11-08 23:11:42.883887 7f026b94f700  2 req 2:0.006769:s3:PUT
>>>>>>>>>>> /my-new-bucket-31337/:create_bucket:verifying op permissions
>>>>>>>>>>> 2016-11-08 23:11:42.883939 7f026b94f700  1 -- 
>>>>>>>>>>> 192.168.168.201:0/3703211654
>>>>>>>>>>> --> 192.168.168.203:6840/36134 -- osd_op(client.160747848.0:1023 
>>>>>>>>>>> 15.efcbc969
>>>>>>>>>>> EDITED - client name.buckets [call user.list_buckets] snapc 0=[]
>>>>>>>>>>> ack+read+known_if_redirected e97148) v7 -- ?+0 0x7f02b70acb00 con
>>>>>>>>>>> 0x7f02cd43a900
>>>>>>>>>>> 2016-11-08 23:11:42.885157 7f02bbfff700  1 -- 
>>>>>>>>>>> 192.168.168.201:0/3703211654
>>>>>>>>>>> <== osd.25 192.168.168.203:6840/36134 37 ==== osd_op_reply(1023 
>>>>>>>>>>> EDITED -
>>>>>>>>>>> client name.buckets [call] v0'0 uv0 ack = -2 ((2) No such file or
>>>>>>>>>>> directory)) v7 ==== 161+0+0 (2433927993 0 0) 0x7f02b886d280 con
>>>>>>>>>>> 0x7f02cd43a900
>>>>>>>>>>> 2016-11-08 23:11:42.885307 7f026b94f700  2 req 2:0.008187:s3:PUT
>>>>>>>>>>> /my-new-bucket-31337/:create_bucket:verifying op params
>>>>>>>>>>> 2016-11-08 23:11:42.885338 7f026b94f700  2 req 2:0.008219:s3:PUT
>>>>>>>>>>> /my-new-bucket-31337/:create_bucket:pre-executing
>>>>>>>>>>> 2016-11-08 23:11:42.885342 7f026b94f700  2 req 2:0.008223:s3:PUT
>>>>>>>>>>> /my-new-bucket-31337/:create_bucket:executing
>>>>>>>>>>> 2016-11-08 23:11:42.885377 7f026b94f700 20 get_system_obj_state:
>>>>>>>>>>> rctx=0x7f026b94c6d0 obj=.rgw:my-new-bucket-31337 
>>>>>>>>>>> state=0x7f02b70cdfe8
>>>>>>>>>>> s->prefetch_data=0
>>>>>>>>>>> 2016-11-08 23:11:42.885390 7f026b94f700 10 cache get:
>>>>>>>>>>> name=.rgw+my-new-bucket-31337 : miss
>>>>>>>>>>> 2016-11-08 23:11:42.885444 7f026b94f700  1 -- 
>>>>>>>>>>> 192.168.168.201:0/3703211654
>>>>>>>>>>> --> 192.168.168.202:6821/14483 -- osd_op(client.160747848.0:1024 
>>>>>>>>>>> 13.501be1c3
>>>>>>>>>>> my-new-bucket-31337 [call version.read,getxattrs,stat] snapc 0=[]
>>>>>>>>>>> ack+read+known_if_redirected e97148) v7 -- ?+0 0x7f02b70ace80 con
>>>>>>>>>>> 0x7f02cd43b980
>>>>>>>>>>> 2016-11-08 23:11:42.886940 7f02bfefe700  1 -- 
>>>>>>>>>>> 192.168.168.201:0/3703211654
>>>>>>>>>>> <== osd.11 192.168.168.202:6821/14483 11 ==== osd_op_reply(1024
>>>>>>>>>>> my-new-bucket-31337 [call,getxattrs,stat] v0'0 uv0 ack = -2 ((2) No 
>>>>>>>>>>> such
>>>>>>>>>>> file or directory)) v7 ==== 223+0+0 (4048274385 0 0) 0x7f02bf480280 
>>>>>>>>>>> con
>>>>>>>>>>> 0x7f02cd43b980
>>>>>>>>>>> 2016-11-08 23:11:42.887120 7f026b94f700 10 cache put:
>>>>>>>>>>> name=.rgw+my-new-bucket-31337 info.flags=0
>>>>>>>>>>> 2016-11-08 23:11:42.887135 7f026b94f700 10 adding 
>>>>>>>>>>> .rgw+my-new-bucket-31337
>>>>>>>>>>> to cache LRU end
>>>>>>>>>>> 2016-11-08 23:11:42.887144 7f026b94f700  0 rest connection is 
>>>>>>>>>>> invalid
>>>>>>>>>>
>>>>>>>>>> Did you configure multi-site? Seems to me that it tries to connect to
>>>>>>>>>> a master zone, so if you have a single zonegroup, single zone
>>>>>>>>>> configuration you need to make sure that the zone (and zonegroup) are
>>>>>>>>>> set to be the master.
>>>>>>>>>>
>>>>>>>>>> Yehuda
>>>>>>>>>>
>>>>>>>>>>> 2016-11-08 23:11:42.887149 7f026b94f700  2 req 2:0.010030:s3:PUT
>>>>>>>>>>> /my-new-bucket-31337/:create_bucket:completing
>>>>>>>>>>> 2016-11-08 23:11:42.887217 7f026b94f700  2 req 2:0.010098:s3:PUT
>>>>>>>>>>> /my-new-bucket-31337/:create_bucket:op status=-22
>>>>>>>>>>> 2016-11-08 23:11:42.887221 7f026b94f700  2 req 2:0.010102:s3:PUT
>>>>>>>>>>> /my-new-bucket-31337/:create_bucket:http status=400
>>>>>>>>>>> 2016-11-08 23:11:42.887244 7f026b94f700  1 ====== req done
>>>>>>>>>>> req=0x7f02ba07b0e0 op status=-22 http_status=400 ======
>>>>>>>>>>> 2016-11-08 23:11:42.887254 7f026b94f700 20 process_request() 
>>>>>>>>>>> returned -22
>>>>>>>>>>>
>>>>>>>>>>> ###############################
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>> Apache proxy logs show the following:
>>>>>>>>>>>
>>>>>>>>>>> 192.168.169.91 - - [08/Nov/2016:23:11:42 +0000] "PUT 
>>>>>>>>>>> /my-new-bucket-31337/
>>>>>>>>>>> HTTP/1.1" 400 4379 "-" "Boto/2.38.0 Python/2.7.12
>>>>>>>>>>> Linux/4.8.4-040804-generic"
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>> The existing buckets work perfectly well, i can list and put 
>>>>>>>>>>> objects. It's
>>>>>>>>>>> the creation of new buckets that i am having issues with.
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>> Could someone please help me to figure out what the issue is?
>>>>>>>>>>>
>>>>>>>>>>> Thanks
>>>>>>>>>>>
>>>>>>>>>>> Andrei
>>>>>>>>>>>
>>>>>>>>>>> _______________________________________________
>>>>>>>>>>> ceph-users mailing list
>>>>>>>>>>> [email protected]
>>>>>>>>>>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>>>> _______________________________________________
>>>>> ceph-users mailing list
>>>>> [email protected]
>>>>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>>>>
>>>>
>>>>
>>>> --
>>>> Yoann Moulin
>>>> EPFL IC-IT
>>> _______________________________________________
>>> ceph-users mailing list
>>> [email protected]
> >> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
_______________________________________________
ceph-users mailing list
[email protected]
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to