Re: [PATCH] mark rbd requiring stable pages

2015-10-22 Thread Mike Christie
On 10/22/15, 11:52 AM, Ilya Dryomov wrote: On Thu, Oct 22, 2015 at 5:37 PM, Mike Christie wrote: On 10/22/2015 06:20 AM, Ilya Dryomov wrote: If we are just talking about if stable pages are not used, and someone is re-writing data to a page after the page has already

Re: newstore direction

2015-10-22 Thread Orit Wasserman
On Thu, 2015-10-22 at 02:12 +, Allen Samuels wrote: > One of the biggest changes that flash is making in the storage world is that > the way basic trade-offs in storage management software architecture are > being affected. In the HDD world CPU time per IOP was relatively > inconsequential,

Re: newstore direction

2015-10-22 Thread Christoph Hellwig
On Wed, Oct 21, 2015 at 10:30:28AM -0700, Sage Weil wrote: > For example: we need to do an overwrite of an existing object that is > atomic with respect to a larger ceph transaction (we're updating a bunch > of other metadata at the same time, possibly overwriting or appending to > multiple

Re: [PATCH] mark rbd requiring stable pages

2015-10-22 Thread Ilya Dryomov
On Thu, Oct 22, 2015 at 7:22 PM, Mike Christie wrote: > On 10/22/15, 11:52 AM, Ilya Dryomov wrote: >> >> On Thu, Oct 22, 2015 at 5:37 PM, Mike Christie >> wrote: >>> >>> On 10/22/2015 06:20 AM, Ilya Dryomov wrote: >> >> If we are just

Re: [PATCH] mark rbd requiring stable pages

2015-10-22 Thread Ilya Dryomov
On Thu, Oct 22, 2015 at 6:07 AM, Mike Christie wrote: > On 10/21/2015 03:57 PM, Ilya Dryomov wrote: >> On Wed, Oct 21, 2015 at 10:51 PM, Ilya Dryomov wrote: >>> On Fri, Oct 16, 2015 at 1:09 PM, Ilya Dryomov wrote: Hmm... On the

keyring issues, 9.1.0

2015-10-22 Thread Deneau, Tom
My current situation as I upgrade to v9.1.0 is that client.admin keyring seems to work fine, for instance for ceph status command. But commands that use client.bootstrap-osd such as /usr/bin/ceph --cluster ceph --name client.bootstrap-osd --keyring /var/lib/ceph/bootstrap-osd/ceph.keyring

RE: newstore direction

2015-10-22 Thread James (Fei) Liu-SSI
Hi Sage and other fellow cephers, I truly share the pains with you all about filesystem while I am working on objectstore to improve the performance. As mentioned , there is nothing wrong with filesystem. Just the Ceph as one of use case need more supports but not provided in near future

tracker.ceph.com downtime today

2015-10-22 Thread Dan Mick
tracker.ceph.com will be brought down today for upgrade and move to a new host. I plan to do this at about 4PM PST (40 minutes from now). Expect a downtime of about 15-20 minutes. More notification to follow. -- Dan Mick Red Hat, Inc. Ceph docs: http://ceph.com/docs -- To unsubscribe from this

Re: tracker.ceph.com downtime today

2015-10-22 Thread Dan Mick
It's back. New DNS info is propagating its way around. If you absolutely must get to it, newtracker.ceph.com is the new address, but please don't bookmark that, as it will be going away after the transition. Please let me know of any problems you have. On 10/22/2015 04:09 PM, Dan Mick wrote: >

Re: newstore direction

2015-10-22 Thread Samuel Just
Since the changes which moved the pg log and the pg info into the pg object space, I think it's now the case that any transaction submitted to the objectstore updates a disjoint range of objects determined by the sequencer. It might be easier to exploit that parallelism if we control allocation

Re: newstore direction

2015-10-22 Thread Samuel Just
Ah, except for the snapmapper. We can split the snapmapper in the same way, though, as long as we are careful with the name. -Sam On Thu, Oct 22, 2015 at 4:42 PM, Samuel Just wrote: > Since the changes which moved the pg log and the pg info into the pg > object space, I think

Re: [PATCH] mark rbd requiring stable pages

2015-10-22 Thread Ronny Hegewald
On Thursday 22 October 2015, Ilya Dryomov wrote: > Well, checksum mismatches are to be expected given what we are doing > now, but I wouldn't expect any data corruptions. Ronny writes that he > saw frequent ext4 corruptions on krbd devices before he enabled stable > pages, which leads me to

Re: [PATCH] mark rbd requiring stable pages

2015-10-22 Thread Ronny Hegewald
On Thursday 22 October 2015, you wrote: > It could be due to a recent change. Ronny, tell us about the workload > and I will check iscsi. I guess the best testcase is a kernel compilation in a make clean; make -j (> 1); loop. The data-corruptions usually happen in the generated .cmd files,

Re: tracker.ceph.com downtime today

2015-10-22 Thread Dan Mick
Fixed a configuration problem preventing updating issues, and switched the mailer to use ipv4; if you updated and failed, or missed an email notification, that may have been why. On 10/22/2015 04:51 PM, Dan Mick wrote: > It's back. New DNS info is propagating its way around. If you > absolutely

RE: newstore direction

2015-10-22 Thread Allen Samuels
How would this kind of split affect small transactions? Will each split be separately transactionally consistent or is there some kind of meta-transaction that synchronizes each of the splits? Allen Samuels Software Architect, Fellow, Systems and Software Solutions 2880 Junction Avenue, San

Re: tracker.ceph.com downtime today

2015-10-22 Thread Dan Mick
tracker.ceph.com down now On 10/22/2015 03:20 PM, Dan Mick wrote: > tracker.ceph.com will be brought down today for upgrade and move to a > new host. I plan to do this at about 4PM PST (40 minutes from now). > Expect a downtime of about 15-20 minutes. More notification to follow. > -- Dan

Re: tracker.ceph.com downtime today

2015-10-22 Thread Kyle Bader
I tried to open a new issue and got this error: Internal error An error occurred on the page you were trying to access. If you continue to experience problems please contact your Redmine administrator for assistance. If you are the Redmine administrator, check your log files for details about

when an osd is started up, IO will be blocked

2015-10-22 Thread wangsongbo
Hi all, When an osd is started, relative IO will be blocked. According to the test result,the larger iops the clients send , the longer it will take to elapse. Adjustment on all the parameters associate with recovery operations was also found useless. How to reduce the impact of this process

Re: tracker.ceph.com downtime today

2015-10-22 Thread Dan Mick
Found that issue; reverted the database to the non-backlog-plugin state, created a test bug. Retry? On 10/22/2015 06:54 PM, Dan Mick wrote: > I see that too. I suspect this is because of leftover database columns > from the backlogs plugin, which is removed. Looking into it. > > On 10/22/2015

Re: tracker.ceph.com downtime today

2015-10-22 Thread Dan Mick
I see that too. I suspect this is because of leftover database columns from the backlogs plugin, which is removed. Looking into it. On 10/22/2015 06:43 PM, Kyle Bader wrote: > I tried to open a new issue and got this error: > > Internal error > > An error occurred on the page you were trying

Re: newstore direction

2015-10-22 Thread Ric Wheeler
I disagree with your point still - your argument was that customers don't like to update their code so we cannot rely on them moving to better file system code. Those same customers would be *just* as reluctant to upgrade OSD code. Been there, done that in pure block storage, pure object

Re: newstore direction

2015-10-22 Thread Ric Wheeler
On 10/22/2015 08:50 AM, Sage Weil wrote: On Wed, 21 Oct 2015, Ric Wheeler wrote: You will have to trust me on this as the Red Hat person who spoke to pretty much all of our key customers about local file systems and storage - customers all have migrated over to using normal file systems under

Re: newstore direction

2015-10-22 Thread Howard Chu
Milosz Tanski adfin.com> writes: > > On Tue, Oct 20, 2015 at 4:00 PM, Sage Weil redhat.com> wrote: > > On Tue, 20 Oct 2015, John Spray wrote: > >> On Mon, Oct 19, 2015 at 8:49 PM, Sage Weil redhat.com> wrote: > >> > - We have to size the kv backend storage (probably still an XFS > >> >

Re: MDS stuck in a crash loop

2015-10-22 Thread Milosz Tanski
On Wed, Oct 21, 2015 at 5:33 PM, John Spray wrote: > On Wed, Oct 21, 2015 at 10:33 PM, John Spray wrote: >>> John, I know you've got >>> https://github.com/ceph/ceph-qa-suite/pull/647. I think that's >>> supposed to be for this, but I'm not sure if you

Re: newstore direction

2015-10-22 Thread Sage Weil
On Wed, 21 Oct 2015, Ric Wheeler wrote: > You will have to trust me on this as the Red Hat person who spoke to pretty > much all of our key customers about local file systems and storage - customers > all have migrated over to using normal file systems under Oracle/DB2. > Typically, they use XFS

Re: MDS stuck in a crash loop

2015-10-22 Thread Sage Weil
On Thu, 22 Oct 2015, John Spray wrote: > On Thu, Oct 22, 2015 at 1:43 PM, Milosz Tanski wrote: > > On Wed, Oct 21, 2015 at 5:33 PM, John Spray wrote: > >> On Wed, Oct 21, 2015 at 10:33 PM, John Spray wrote: > John, I know you've got >

Re: newstore direction

2015-10-22 Thread Milosz Tanski
On Tue, Oct 20, 2015 at 4:00 PM, Sage Weil wrote: > On Tue, 20 Oct 2015, John Spray wrote: >> On Mon, Oct 19, 2015 at 8:49 PM, Sage Weil wrote: >> > - We have to size the kv backend storage (probably still an XFS >> > partition) vs the block storage. Maybe

Re: MDS stuck in a crash loop

2015-10-22 Thread John Spray
On Thu, Oct 22, 2015 at 1:43 PM, Milosz Tanski wrote: > On Wed, Oct 21, 2015 at 5:33 PM, John Spray wrote: >> On Wed, Oct 21, 2015 at 10:33 PM, John Spray wrote: John, I know you've got

[PATCH 0/3] Minor cleanup for locks API

2015-10-22 Thread Benjamin Coddington
NFS has recently been moving things around to cope with the situation where a struct file may not be available during an unlock. That work has presented an opportunity to do a minor cleanup on the locks API. Users of posix_lock_file_wait() (for FL_POSIX style locks) and flock_lock_file_wait()

[PATCH 3/3] locks: cleanup posix_lock_inode_wait and flock_lock_inode_wait

2015-10-22 Thread Benjamin Coddington
All callers use locks_lock_inode_wait() instead. Signed-off-by: Benjamin Coddington --- fs/locks.c |5 + include/linux/fs.h | 24 2 files changed, 1 insertions(+), 28 deletions(-) diff --git a/fs/locks.c b/fs/locks.c index

Re: MDS stuck in a crash loop

2015-10-22 Thread Milosz Tanski
On Thu, Oct 22, 2015 at 8:48 AM, John Spray wrote: > On Thu, Oct 22, 2015 at 1:43 PM, Milosz Tanski wrote: >> On Wed, Oct 21, 2015 at 5:33 PM, John Spray wrote: >>> On Wed, Oct 21, 2015 at 10:33 PM, John Spray wrote:

Re: [PATCH] mark rbd requiring stable pages

2015-10-22 Thread Mike Christie
On 10/22/2015 06:20 AM, Ilya Dryomov wrote: > >> > >> > If we are just talking about if stable pages are not used, and someone >> > is re-writing data to a page after the page has already been submitted >> > to the block layer (I mean the page is on some bio which is on a request >> > which is on

[PATCH] net: ceph: osd_client: change osd_req_op_data() macro

2015-10-22 Thread Ioana Ciornei
This patch changes the osd_req_op_data() macro to not evaluate parameters more than once in order to follow the kernel coding style. Signed-off-by: Ioana Ciornei Reviewed-by: Alex Elder --- net/ceph/osd_client.c | 10 ++ 1 file changed, 6

[PATCH 2/3] Move locks API users to locks_lock_inode_wait()

2015-10-22 Thread Benjamin Coddington
Instead of having users check for FL_POSIX or FL_FLOCK to call the correct locks API function, use the check within locks_lock_inode_wait(). This allows for some later cleanup. Signed-off-by: Benjamin Coddington --- drivers/staging/lustre/lustre/llite/file.c |8

[PATCH 1/3] locks: introduce locks_lock_inode_wait()

2015-10-22 Thread Benjamin Coddington
Users of the locks API commonly call either posix_lock_file_wait() or flock_lock_file_wait() depending upon the lock type. Add a new function locks_lock_inode_wait() which will check and call the correct function for the type of lock passed in. Signed-off-by: Benjamin Coddington

Re: [PATCH 3/3] locks: cleanup posix_lock_inode_wait and flock_lock_inode_wait

2015-10-22 Thread kbuild test robot
Hi Benjamin, [auto build test WARNING on jlayton/linux-next -- if it's inappropriate base, please suggest rules for selecting the more suitable base] url: https://github.com/0day-ci/linux/commits/Benjamin-Coddington/locks-introduce-locks_lock_inode_wait/20151022-233848 reproduce

[RFC PATCH] locks: posix_lock_inode_wait() can be static

2015-10-22 Thread kbuild test robot
Signed-off-by: Fengguang Wu --- locks.c |4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/fs/locks.c b/fs/locks.c index daf4664..0d2b326 100644 --- a/fs/locks.c +++ b/fs/locks.c @@ -1173,7 +1173,7 @@ EXPORT_SYMBOL(posix_lock_file); * * Apply

Ceph erasure coding

2015-10-22 Thread Kjetil Babington
Hi, I have a question about the capabilities of the erasure coding API in Ceph. Let's say that I have 10 data disks and 4 parity disks, is it possible to create an erasure coding plugin which creates 20 data chunks and 8 parity chunks, and then places two chunks on each osd? Or said maybe a bit

Re: [PATCH 1/3] locks: introduce locks_lock_inode_wait()

2015-10-22 Thread Benjamin Coddington
On Thu, 22 Oct 2015, Benjamin Coddington wrote: > Users of the locks API commonly call either posix_lock_file_wait() or > flock_lock_file_wait() depending upon the lock type. Add a new function > locks_lock_inode_wait() which will check and call the correct function for > the type of lock passed

Re: [PATCH 1/3] locks: introduce locks_lock_inode_wait()

2015-10-22 Thread kbuild test robot
Hi Benjamin, [auto build test ERROR on jlayton/linux-next -- if it's inappropriate base, please suggest rules for selecting the more suitable base] url: https://github.com/0day-ci/linux/commits/Benjamin-Coddington/locks-introduce-locks_lock_inode_wait/20151022-233848 config: x86_64

Re: [PATCH] mark rbd requiring stable pages

2015-10-22 Thread Ilya Dryomov
On Thu, Oct 22, 2015 at 5:37 PM, Mike Christie wrote: > On 10/22/2015 06:20 AM, Ilya Dryomov wrote: >> >>> > >>> > If we are just talking about if stable pages are not used, and someone >>> > is re-writing data to a page after the page has already been submitted >>> > to the

Re: Ceph erasure coding

2015-10-22 Thread Samuel Just
Not on purpose... out of curiosity, why do you want to do that? -Sam On Thu, Oct 22, 2015 at 9:44 AM, Kjetil Babington wrote: > Hi, > > I have a question about the capabilities of the erasure coding API in > Ceph. Let's say that I have 10 data disks and 4 parity disks, is it

Re: Ceph erasure coding

2015-10-22 Thread Loic Dachary
Hi, On 22/10/2015 18:44, Kjetil Babington wrote: > Hi, > > I have a question about the capabilities of the erasure coding API in > Ceph. Let's say that I have 10 data disks and 4 parity disks, is it > possible to create an erasure coding plugin which creates 20 data > chunks and 8 parity chunks,