[ceph-users] Re: Hardware for new OSD nodes.

2020-10-23 Thread Brian Topping
with volume size. That said, the disks are portable at that point and really don’t have deep performance bottlenecks because mirroring and striping is cheap. Sent from my iPhone > On Oct 23, 2020, at 03:54, Eneko Lacunza wrote: > > Hi Brian, > >> El 22/10/20 a las 18

[ceph-users] Re: Hardware for new OSD nodes.

2020-10-22 Thread Brian Topping
> On Oct 22, 2020, at 10:34 AM, Anthony D'Atri wrote: > >>- You must really be sure your raid card is dependable. (sorry but I have >> seen so much management problems with top-tier RAID cards I avoid them like >> the plague). > > This. I’d definitely avoid a RAID card. If I can do

[ceph-users] Re: Hardware for new OSD nodes.

2020-10-22 Thread Brian Topping
> On Oct 22, 2020, at 9:14 AM, Eneko Lacunza wrote: > > Don't stripe them, if one NVMe fails you'll lose all OSDs. Just use 1 NVMe > drive for 2 SAS drives and provision 300GB for WAL/DB for each OSD (see > related threads on this mailing list about why that exact size). > > This way if a

[ceph-users] Re: Recommended settings for PostgreSQL

2020-10-19 Thread Brian Topping
Another option is to let PosgreSQL do the replication with local storage. There are great reasons for Ceph, but databases optimize for this kind of thing extremely well. With replication in hand, run snapshots to RADOS buckets for long term storage. > On Oct 17, 2020, at 7:28 AM, Gencer W.

[ceph-users] Re: MONs are down, the quorum is unable to resolve.

2020-10-12 Thread Brian Topping
of the cluster as I can’t lost OSDs data. > > G. > > Le mar. 13 oct. 2020 à 00:01, Brian Topping <mailto:brian.topp...@gmail.com>> a écrit : > Hi there! > > This isn’t a difficult problem to fix. For purposes of clarity, the monmap is > just a part of the monit

[ceph-users] Re: MONs are down, the quorum is unable to resolve.

2020-10-12 Thread Brian Topping
Hi there! This isn’t a difficult problem to fix. For purposes of clarity, the monmap is just a part of the monitor database. You generally have all the details correct though. Have you looked at the process in

[ceph-users] Monitor recovery

2020-10-09 Thread Brian Topping
Hello experts, I have accidentally created a situation where the only monitor in a cluster has been moved to a new node without it’s /var/lib/ceph contents. Not realizing what I had done, I decommissioned the original node, but still have the contents of it’s /var/lib/ceph. Can I shut down

[ceph-users] Re: Feedback for proof of concept OSD Node

2020-10-04 Thread Brian Topping
ng between VMs is as fast as you can get. Feeling like you want physical devices? Plug them into USB and map them to the VMs. Then start doing things like breaking the network, removing disks and still recovering your data. That’s far more important to know than how to plug together big devices

[ceph-users] Re: Feedback for proof of concept OSD Node

2020-10-04 Thread Brian Topping
ls. >> >> Also, in the mid-term, I do have plans to deploy a small OpenStack Cluster. >> >> Because of that, I would like to set up the first small Ceph Cluster that >> can scale as my needs grow, the idea is to have 3 OSD nodes with the same >> characteristics an

[ceph-users] Re: Feedback for proof of concept OSD Node

2020-10-01 Thread Brian Topping
Welcome to Ceph! I think better questions to start with are “what are your objectives in your study?” Is it just seeing Ceph run with many disks, or are you trying to see how much performance you can get out of it with distributed disk? What is your budget? Do you want to try different

[ceph-users] Re: how to "undelete" a pool

2020-09-25 Thread Brian Topping
Haha I figured out you were on Rook. I think you need to add an annotation or label to the CRD. Just create an empty one and do a kubectl get cephcluster -oyaml to see what it generates then figure out what the appropriate analog for the restored CRD is. Once the operator sees the correct

[ceph-users] Re: Thank you!

2020-07-20 Thread Brian Topping
If there was a “like” button, I would have just clicked that to keep the list noise down. I have smaller operations and so my cluster goes down a lot more often. I keep dreading my abuse of the cluster and it just keeps coming back for more. Ceph really is amazing, and it’s hard to fully

[ceph-users] Re: Unsubscribe this mail list

2020-06-25 Thread Brian Topping
That kind of information is ALWAYS in the headers of every email. > List-Unsubscribe: > On Jun 25, 2020, at 9:21 AM, adan wrote: > > hello > > i want to unsubscribe this mail list . help me please. > > > 在 2020/6/25 22:42, ceph-users-requ...@ceph.io 写道: >>

[ceph-users] Re: CEPH failure domain - power considerations

2020-05-29 Thread Brian Topping
Phil, this would be an excellent contribution to the blog or the introductory documentation. I’ve been using Ceph for over a year this brought together a lot of concepts that I hadn’t related so succinctly in the past. One of the things that I hadn’t really conceptualized well was “why size of

[ceph-users] Re: Ceph and Windows - experiences or suggestions

2020-05-29 Thread Brian Topping
Doesn’t SMB support require a paid subscription? > On Feb 13, 2020, at 3:12 AM, Martin Verges wrote: > > Hello Lars, > > we have full SMB Support in our Ceph management solution. You can create > simple (user+pass) or complex SMB (AD) high available shares on CTDB > clustered Samba with

[ceph-users] Re: Clarification of documentation

2020-05-19 Thread Brian Topping
I have been running Ceph over a gigabit WAN for a few months now and have been happy with it. Mine is set up with Strongswan tunnels And dynamic routing with BIRD) (although I would have used transport Mode and iBGP in hindsight). I generally have 300-500kbps flow with 5ms latency. What I

[ceph-users] Re: Dear Abby: Why Is Architecting CEPH So Hard?

2020-04-22 Thread Brian Topping
Great set of suggestions, thanks! One to consider: > On Apr 22, 2020, at 4:14 PM, Jack wrote: > > I use 32GB flash-based satadom devices for root device > They are basically SSD, and do not take front slots > As they are never burning up, we never replace them > Ergo, the need to "open" the

[ceph-users] Re: Understanding monitor requirements

2020-04-11 Thread Brian Topping
> On Apr 11, 2020, at 5:54 PM, Anthony D'Atri wrote: > > Dumb question, can’t you raise the MTU of the tunnel? I’m good with any question, that got it, thank you! I’m not exactly sure what happened, I believe an MTU setting I tried didn’t actually take or the CNI software was somehow not

[ceph-users] Re: Understanding monitor requirements

2020-04-11 Thread Brian Topping
Hi again, after all, this appears to be an MTU issue: Baseline: 1) Two of the nodes have a straight ethernet with 1500MTU, the third (problem) node is on a WAN tunnel with a restricted MTU. It appears that the MTUs were not set up correctly, so no surprise some software has problems. 2) I

[ceph-users] Re: Understanding monitor requirements

2020-04-10 Thread Brian Topping
> On Apr 9, 2020, at 8:39 PM, Brian Topping wrote: > > Hi experts, question about monitors and latency. I think I jumped to a hasty conclusion in my previous email. It’s not about the latency, but about something to do with a problem when mon.c calls an election after sync

[ceph-users] Understanding monitor requirements

2020-04-09 Thread Brian Topping
Hi experts, question about monitors and latency. I am setting up a new cluster and I’d like to have more than one monitor. Unfortunately, the primary site only has two chassis, so to get the third mon, I’ve been trying to bring it up remotely. So far, it’s not working and I wonder if someone

[ceph-users] Re: Q release name

2020-03-23 Thread Brian Topping
I liked the first one a lot. Until I read the second one. > On Mar 23, 2020, at 11:29 AM, Anthony D'Atri wrote: > > That has potential. Another, albeit suboptimal idea would be simply > > Quid > > as in > > ’S quid > > as in “it’s squid”. cf. https://en.wikipedia.org/wiki/%27S_Wonderful

[ceph-users] Re: Q release name

2020-03-23 Thread Brian Topping
Maybe just call it Quincy and have a backstory? Might be fun... > On Mar 23, 2020, at 11:11 AM, Sage Weil wrote: > > Hi everyone, > > As we wrap up Octopus and kick of development for Pacific, now it seems > like a good idea to sort out what to call the Q release. >

[ceph-users] Re: Single machine / multiple monitors

2020-03-12 Thread Brian Topping
2020, at 11:08 AM, Brian Topping wrote: >> >> If the ceph roadmap is getting rid of named clusters, how will multiple >> clusters be supported? How (for instance) would `/var/lib/ceph/mon/{name}` >> directories be resolved? ___ cep

[ceph-users] Re: Single machine / multiple monitors

2020-03-12 Thread Brian Topping
If the ceph roadmap is getting rid of named clusters, how will multiple clusters be supported? How (for instance) would `/var/lib/ceph/mon/{name}` directories be resolved? > On Mar 11, 2020, at 8:29 PM, Brian Topping wrote: > >> On Mar 11, 2020, at 7:59 PM, Anthony D

[ceph-users] Re: Single machine / multiple monitors

2020-03-11 Thread Brian Topping
> On Mar 11, 2020, at 5:28 PM, Anthony D'Atri wrote: > > Custom cluster names are being incrementally deprecated. With ceph-deploy I > thought they were removed in 1.39. Maybe the man page is not up-to-date. The command does not accept `—cluster` as an option. > You could probably achieve

[ceph-users] Single machine / multiple monitors

2020-03-11 Thread Brian Topping
Hi, I’m getting conflicting reads from the documentation. It seems that by using the “cluster name”[1], multiple clusters can be run in parallel on the same hardware. In trying to set this up with `ceph-deploy`, I see the man page[2] says "if it finds the distro.init to be sysvinit (Fedora,

[ceph-users] Monitor / MDS distribution over WAN

2020-02-14 Thread Brian Topping
I had posted about some of this a year ago in [1] and got some really helpful answers. Fortunately, I know a lot more now and feel a lot more comfortable with the scenario. Because I didn’t understand the architecture very well, I took a pause on distributing monitors and MDS over a WAN. I want