2016-08-10 9:30 GMT+05:00 Александр Пивушков :
> I want to use Ceph only as user data storage.
> user program writes data to a folder that is mounted on a Ceph.
> Virtual machine images are not stored on the Сeph.
> Fiber channel and 40GbE are used only for the rapid transmission
Hello Vladimir,
On Wed, 10 Aug 2016 09:12:39 +0500 Дробышевский, Владимир wrote:
> Christian,
>
> I have to say that OpenNebula 5 doesn't need any additional hacks (ok,
> just two lines of code to support rescheduling in case of the original node
> failure and even these patch scheduled to
Hello!
Brad,
is that possible from the default logging or verbose one is needed??
I 've managed to get the UUID of the deleted volume from OpenStack but
don't really know how to get the offsets and OSD maps since "rbd info"
doesn't provide any information for that volume.
Is it possible
I want to use Ceph only as user data storage.
user program writes data to a folder that is mounted on a Ceph.
Virtual machine images are not stored on the Сeph.
Fiber channel and 40GbE are used only for the rapid transmission of
information between the cluster Ceph and the virtual machine on
Christian,
I have to say that OpenNebula 5 doesn't need any additional hacks (ok,
just two lines of code to support rescheduling in case of the original node
failure and even these patch scheduled to 5.2 to be added after my question
a couple of weeks ago; but it isn't about 'live') or an
On Tue, Aug 9, 2016 at 7:39 AM, George Mihaiescu wrote:
> Look in the cinder db, the volumes table to find the Uuid of the deleted
> volume.
You could also look through the logs at the time of the delete and I
suspect you should
be able to see how the rbd image was
On Wed, Aug 10, 2016 at 12:26 AM, agung Laksono wrote:
>
> Hi Ceph users,
>
> I am new in ceph. I've been succeed installing ceph in 4 VM using Quick
> installation guide in ceph documentation.
>
> And I've also done to compile
> ceph from source code, build and install in
Hello,
On Tue, 9 Aug 2016 14:15:59 -0400 Jeff Bailey wrote:
>
>
> On 8/9/2016 10:43 AM, Wido den Hollander wrote:
> >
> >> Op 9 augustus 2016 om 16:36 schreef Александр Пивушков :
> >>
> >>
> >> > >> Hello dear community!
> >> I'm new to the Ceph and not long ago took up
On 8/9/2016 10:43 AM, Wido den Hollander wrote:
Op 9 augustus 2016 om 16:36 schreef Александр Пивушков :
> >> Hello dear community!
I'm new to the Ceph and not long ago took up the theme of building clusters.
Therefore it is very important to your opinion.
It is necessary
Hi Wido,
thanks for your advice.
Best,
Martin
On Tue, Aug 9, 2016 at 10:05 AM, Wido den Hollander wrote:
>
>> Op 8 augustus 2016 om 16:45 schreef Martin Palma :
>>
>>
>> Hi all,
>>
>> we are in the process of expanding our cluster and I would like to
>> know if
Вторник, 9 августа 2016, 17:43 +03:00 от Wido den Hollander :
>
>
>> Op 9 augustus 2016 om 16:36 schreef Александр Пивушков < p...@mail.ru >:
>>
>>
>> > >> Hello dear community!
>> >> >> I'm new to the Ceph and not long ago took up the theme of building
>> >> >> clusters.
>>
> Op 9 augustus 2016 om 16:36 schreef Александр Пивушков :
>
>
> > >> Hello dear community!
> >> >> I'm new to the Ceph and not long ago took up the theme of building
> >> >> clusters.
> >> >> Therefore it is very important to your opinion.
> >> >> It is necessary to create a
Hi Ceph users,
I am new in ceph. I've been succeed installing ceph in 4 VM using Quick
installation guide in ceph documentation.
And I've also done to compile
ceph from source code, build and install in single vm.
What I want to do next is that run ceph multiple nodes in a cluster
but only
On Tue, Aug 9, 2016 at 2:00 AM, Kenneth Waegeman
wrote:
> Hi,
>
> I did a diff on the directories of all three the osds, no difference .. So I
> don't know what's wrong.
omap (as implied by the omap_digest complaint) is stored in the OSD
leveldb, not in the data
Hello,
[re-added the list]
Also try to leave a line-break, paragraph between quoted and new text,
your mail looked like it was all written by me...
On Tue, 09 Aug 2016 11:00:27 +0300 Александр Пивушков wrote:
> Thank you for your response!
>
>
> >Вторник, 9 августа 2016, 5:11 +03:00 от
On Mon, Aug 8, 2016 at 9:39 PM, Georgios Dimitrakakis
wrote:
> Dear David (and all),
>
> the data are considered very critical therefore all this attempt to
> recover them.
>
> Although the cluster hasn't been fully stopped all users actions have. I
> mean services are
Gregory,
I've been given a tip by one of the ceph user list members on tuning values and
data migration and cluster IO. I had an issues twice already where my vms would
simply loose IO and crash while the cluster is being optimised for the new
tunables.
The recommendations were to upgrade the
Hi Mark, thanks for following up. I'm now pretty convinced I have issues
with my network, it's not Ceph related. My cursory iperf tests between
pairs of hosts were looking fine but with multiple clients I'm seeing
really high tcp retransmissions.
On Mon, Aug 8, 2016 at 1:07 PM, Mark Nelson
Hi,
I did a diff on the directories of all three the osds, no difference ..
So I don't know what's wrong.
Only thing I see different is a scrub file in the TEMP folder (it is
already another pg than last mail):
-rw-r--r--1 ceph ceph 0 Aug 9 09:51
> Op 8 augustus 2016 om 16:45 schreef Martin Palma :
>
>
> Hi all,
>
> we are in the process of expanding our cluster and I would like to
> know if there are some best practices in doing so.
>
> Our current cluster is composted as follows:
> - 195 OSDs (14 Storage Nodes)
> -
Hi,
I saved a file sizing 5GB in the cluster. OSD disk "Used space" increases 15GB
in total because replication is 3. And radosgw-admin bucket stats --uid=someuid
shows that num-objects is increased by 1.
However, after I removed the object, I observe this:
the OSD disk usage does NOT
21 matches
Mail list logo