On 10/22/15, 11:52 AM, Ilya Dryomov wrote:
On Thu, Oct 22, 2015 at 5:37 PM, Mike Christie wrote:
On 10/22/2015 06:20 AM, Ilya Dryomov wrote:
If we are just talking about if stable pages are not used, and someone
is re-writing data to a page after the page has already
On Thu, 2015-10-22 at 02:12 +, Allen Samuels wrote:
> One of the biggest changes that flash is making in the storage world is that
> the way basic trade-offs in storage management software architecture are
> being affected. In the HDD world CPU time per IOP was relatively
> inconsequential,
On Wed, Oct 21, 2015 at 10:30:28AM -0700, Sage Weil wrote:
> For example: we need to do an overwrite of an existing object that is
> atomic with respect to a larger ceph transaction (we're updating a bunch
> of other metadata at the same time, possibly overwriting or appending to
> multiple
On Thu, Oct 22, 2015 at 7:22 PM, Mike Christie wrote:
> On 10/22/15, 11:52 AM, Ilya Dryomov wrote:
>>
>> On Thu, Oct 22, 2015 at 5:37 PM, Mike Christie
>> wrote:
>>>
>>> On 10/22/2015 06:20 AM, Ilya Dryomov wrote:
>>
>> If we are just
On Thu, Oct 22, 2015 at 6:07 AM, Mike Christie wrote:
> On 10/21/2015 03:57 PM, Ilya Dryomov wrote:
>> On Wed, Oct 21, 2015 at 10:51 PM, Ilya Dryomov wrote:
>>> On Fri, Oct 16, 2015 at 1:09 PM, Ilya Dryomov wrote:
Hmm... On the
My current situation as I upgrade to v9.1.0 is that client.admin keyring seems
to work fine, for instance for ceph status command. But commands that use
client.bootstrap-osd such as
/usr/bin/ceph --cluster ceph --name client.bootstrap-osd --keyring
/var/lib/ceph/bootstrap-osd/ceph.keyring
Hi Sage and other fellow cephers,
I truly share the pains with you all about filesystem while I am working on
objectstore to improve the performance. As mentioned , there is nothing wrong
with filesystem. Just the Ceph as one of use case need more supports but not
provided in near future
tracker.ceph.com will be brought down today for upgrade and move to a
new host. I plan to do this at about 4PM PST (40 minutes from now).
Expect a downtime of about 15-20 minutes. More notification to follow.
--
Dan Mick
Red Hat, Inc.
Ceph docs: http://ceph.com/docs
--
To unsubscribe from this
It's back. New DNS info is propagating its way around. If you
absolutely must get to it, newtracker.ceph.com is the new address, but
please don't bookmark that, as it will be going away after the transition.
Please let me know of any problems you have.
On 10/22/2015 04:09 PM, Dan Mick wrote:
>
Since the changes which moved the pg log and the pg info into the pg
object space, I think it's now the case that any transaction submitted
to the objectstore updates a disjoint range of objects determined by
the sequencer. It might be easier to exploit that parallelism if we
control allocation
Ah, except for the snapmapper. We can split the snapmapper in the
same way, though, as long as we are careful with the name.
-Sam
On Thu, Oct 22, 2015 at 4:42 PM, Samuel Just wrote:
> Since the changes which moved the pg log and the pg info into the pg
> object space, I think
On Thursday 22 October 2015, Ilya Dryomov wrote:
> Well, checksum mismatches are to be expected given what we are doing
> now, but I wouldn't expect any data corruptions. Ronny writes that he
> saw frequent ext4 corruptions on krbd devices before he enabled stable
> pages, which leads me to
On Thursday 22 October 2015, you wrote:
> It could be due to a recent change. Ronny, tell us about the workload
> and I will check iscsi.
I guess the best testcase is a kernel compilation in a make clean; make -j (>
1); loop. The data-corruptions usually happen in the generated .cmd files,
Fixed a configuration problem preventing updating issues, and switched
the mailer to use ipv4; if you updated and failed, or missed an email
notification, that may have been why.
On 10/22/2015 04:51 PM, Dan Mick wrote:
> It's back. New DNS info is propagating its way around. If you
> absolutely
How would this kind of split affect small transactions? Will each split be
separately transactionally consistent or is there some kind of meta-transaction
that synchronizes each of the splits?
Allen Samuels
Software Architect, Fellow, Systems and Software Solutions
2880 Junction Avenue, San
tracker.ceph.com down now
On 10/22/2015 03:20 PM, Dan Mick wrote:
> tracker.ceph.com will be brought down today for upgrade and move to a
> new host. I plan to do this at about 4PM PST (40 minutes from now).
> Expect a downtime of about 15-20 minutes. More notification to follow.
>
--
Dan
I tried to open a new issue and got this error:
Internal error
An error occurred on the page you were trying to access.
If you continue to experience problems please contact your Redmine
administrator for assistance.
If you are the Redmine administrator, check your log files for details
about
Hi all,
When an osd is started, relative IO will be blocked.
According to the test result,the larger iops the clients send , the
longer it will take to elapse.
Adjustment on all the parameters associate with recovery operations was
also found useless.
How to reduce the impact of this process
Found that issue; reverted the database to the non-backlog-plugin state,
created a test bug. Retry?
On 10/22/2015 06:54 PM, Dan Mick wrote:
> I see that too. I suspect this is because of leftover database columns
> from the backlogs plugin, which is removed. Looking into it.
>
> On 10/22/2015
I see that too. I suspect this is because of leftover database columns
from the backlogs plugin, which is removed. Looking into it.
On 10/22/2015 06:43 PM, Kyle Bader wrote:
> I tried to open a new issue and got this error:
>
> Internal error
>
> An error occurred on the page you were trying
I disagree with your point still - your argument was that customers don't like
to update their code so we cannot rely on them moving to better file system
code. Those same customers would be *just* as reluctant to upgrade OSD code.
Been there, done that in pure block storage, pure object
On 10/22/2015 08:50 AM, Sage Weil wrote:
On Wed, 21 Oct 2015, Ric Wheeler wrote:
You will have to trust me on this as the Red Hat person who spoke to pretty
much all of our key customers about local file systems and storage - customers
all have migrated over to using normal file systems under
Milosz Tanski adfin.com> writes:
>
> On Tue, Oct 20, 2015 at 4:00 PM, Sage Weil redhat.com> wrote:
> > On Tue, 20 Oct 2015, John Spray wrote:
> >> On Mon, Oct 19, 2015 at 8:49 PM, Sage Weil redhat.com> wrote:
> >> > - We have to size the kv backend storage (probably still an XFS
> >> >
On Wed, Oct 21, 2015 at 5:33 PM, John Spray wrote:
> On Wed, Oct 21, 2015 at 10:33 PM, John Spray wrote:
>>> John, I know you've got
>>> https://github.com/ceph/ceph-qa-suite/pull/647. I think that's
>>> supposed to be for this, but I'm not sure if you
On Wed, 21 Oct 2015, Ric Wheeler wrote:
> You will have to trust me on this as the Red Hat person who spoke to pretty
> much all of our key customers about local file systems and storage - customers
> all have migrated over to using normal file systems under Oracle/DB2.
> Typically, they use XFS
On Thu, 22 Oct 2015, John Spray wrote:
> On Thu, Oct 22, 2015 at 1:43 PM, Milosz Tanski wrote:
> > On Wed, Oct 21, 2015 at 5:33 PM, John Spray wrote:
> >> On Wed, Oct 21, 2015 at 10:33 PM, John Spray wrote:
> John, I know you've got
>
On Tue, Oct 20, 2015 at 4:00 PM, Sage Weil wrote:
> On Tue, 20 Oct 2015, John Spray wrote:
>> On Mon, Oct 19, 2015 at 8:49 PM, Sage Weil wrote:
>> > - We have to size the kv backend storage (probably still an XFS
>> > partition) vs the block storage. Maybe
On Thu, Oct 22, 2015 at 1:43 PM, Milosz Tanski wrote:
> On Wed, Oct 21, 2015 at 5:33 PM, John Spray wrote:
>> On Wed, Oct 21, 2015 at 10:33 PM, John Spray wrote:
John, I know you've got
NFS has recently been moving things around to cope with the situation where
a struct file may not be available during an unlock. That work has
presented an opportunity to do a minor cleanup on the locks API.
Users of posix_lock_file_wait() (for FL_POSIX style locks) and
flock_lock_file_wait()
All callers use locks_lock_inode_wait() instead.
Signed-off-by: Benjamin Coddington
---
fs/locks.c |5 +
include/linux/fs.h | 24
2 files changed, 1 insertions(+), 28 deletions(-)
diff --git a/fs/locks.c b/fs/locks.c
index
On Thu, Oct 22, 2015 at 8:48 AM, John Spray wrote:
> On Thu, Oct 22, 2015 at 1:43 PM, Milosz Tanski wrote:
>> On Wed, Oct 21, 2015 at 5:33 PM, John Spray wrote:
>>> On Wed, Oct 21, 2015 at 10:33 PM, John Spray wrote:
On 10/22/2015 06:20 AM, Ilya Dryomov wrote:
>
>> >
>> > If we are just talking about if stable pages are not used, and someone
>> > is re-writing data to a page after the page has already been submitted
>> > to the block layer (I mean the page is on some bio which is on a request
>> > which is on
This patch changes the osd_req_op_data() macro to not evaluate
parameters more than once in order to follow the kernel coding style.
Signed-off-by: Ioana Ciornei
Reviewed-by: Alex Elder
---
net/ceph/osd_client.c | 10 ++
1 file changed, 6
Instead of having users check for FL_POSIX or FL_FLOCK to call the correct
locks API function, use the check within locks_lock_inode_wait(). This
allows for some later cleanup.
Signed-off-by: Benjamin Coddington
---
drivers/staging/lustre/lustre/llite/file.c |8
Users of the locks API commonly call either posix_lock_file_wait() or
flock_lock_file_wait() depending upon the lock type. Add a new function
locks_lock_inode_wait() which will check and call the correct function for
the type of lock passed in.
Signed-off-by: Benjamin Coddington
Hi Benjamin,
[auto build test WARNING on jlayton/linux-next -- if it's inappropriate base,
please suggest rules for selecting the more suitable base]
url:
https://github.com/0day-ci/linux/commits/Benjamin-Coddington/locks-introduce-locks_lock_inode_wait/20151022-233848
reproduce
Signed-off-by: Fengguang Wu
---
locks.c |4 ++--
1 file changed, 2 insertions(+), 2 deletions(-)
diff --git a/fs/locks.c b/fs/locks.c
index daf4664..0d2b326 100644
--- a/fs/locks.c
+++ b/fs/locks.c
@@ -1173,7 +1173,7 @@ EXPORT_SYMBOL(posix_lock_file);
*
* Apply
Hi,
I have a question about the capabilities of the erasure coding API in
Ceph. Let's say that I have 10 data disks and 4 parity disks, is it
possible to create an erasure coding plugin which creates 20 data
chunks and 8 parity chunks, and then places two chunks on each osd?
Or said maybe a bit
On Thu, 22 Oct 2015, Benjamin Coddington wrote:
> Users of the locks API commonly call either posix_lock_file_wait() or
> flock_lock_file_wait() depending upon the lock type. Add a new function
> locks_lock_inode_wait() which will check and call the correct function for
> the type of lock passed
Hi Benjamin,
[auto build test ERROR on jlayton/linux-next -- if it's inappropriate base,
please suggest rules for selecting the more suitable base]
url:
https://github.com/0day-ci/linux/commits/Benjamin-Coddington/locks-introduce-locks_lock_inode_wait/20151022-233848
config: x86_64
On Thu, Oct 22, 2015 at 5:37 PM, Mike Christie wrote:
> On 10/22/2015 06:20 AM, Ilya Dryomov wrote:
>>
>>> >
>>> > If we are just talking about if stable pages are not used, and someone
>>> > is re-writing data to a page after the page has already been submitted
>>> > to the
Not on purpose... out of curiosity, why do you want to do that?
-Sam
On Thu, Oct 22, 2015 at 9:44 AM, Kjetil Babington wrote:
> Hi,
>
> I have a question about the capabilities of the erasure coding API in
> Ceph. Let's say that I have 10 data disks and 4 parity disks, is it
Hi,
On 22/10/2015 18:44, Kjetil Babington wrote:
> Hi,
>
> I have a question about the capabilities of the erasure coding API in
> Ceph. Let's say that I have 10 data disks and 4 parity disks, is it
> possible to create an erasure coding plugin which creates 20 data
> chunks and 8 parity chunks,
43 matches
Mail list logo