We all get burned by the firewall at one time or another. Hence the name 'fire'wall! :) I'm glad you got it working.
On Thu, Mar 12, 2015 at 2:53 PM, Jesus Chavez (jeschave) <[email protected] > wrote: > This is awkard Robert all this time was the firewall :( I cant believe I > spent 2 days trying to figure out :(. Thank you so much! > > > * Jesus Chavez* > SYSTEMS ENGINEER-C.SALES > > [email protected] > Phone: *+52 55 5267 3146 <+52%2055%205267%203146>* > Mobile: *+51 1 5538883255 <+51%201%205538883255>* > > CCIE - 44433 > > On Mar 12, 2015, at 2:48 PM, Robert LeBlanc <[email protected]> wrote: > > I'm not sure why you are having such a hard time. I added monitors (and > removed them) on CentOS 7 by following what I had. The thing that kept > tripping me up was firewalld. Once I either shut it off or created a > service for Ceph, it worked fine. > > What is in in /var/log/ceph/ceph-mon.tauro.log when it is hunting for a > monitor? > > On Thu, Mar 12, 2015 at 2:31 PM, Jesus Chavez (jeschave) < > [email protected]> wrote: > >> Hi Steffen I already had them in my configuration <emoji_u1f61e.png> I >> am stress now because it seems like none of the methods did help :( this is >> bad I think I am going to get back to rhel6.6 where xfs is a damn add on >> and I have to install from centos repo make ceph like patch :( but at last >> with RHEL6.6 work, Shame on RHEL7 the next time I will sell everything with >> ubuntu lol >> >> >> >> * Jesus Chavez* >> SYSTEMS ENGINEER-C.SALES >> >> [email protected] >> Phone: *+52 55 5267 3146 <+52%2055%205267%203146>* >> Mobile: *+51 1 5538883255 <+51%201%205538883255>* >> >> CCIE - 44433 >> >> On Mar 12, 2015, at 1:56 PM, Steffen W Sørensen <[email protected]> wrote: >> >> >> On 12/03/2015, at 20.00, Jesus Chavez (jeschave) <[email protected]> >> wrote: >> >> Thats what I thought and did actually the monmap and keyring were >> copied to the new monitor and there with 2 elements I did the mkfs thing >> and still have that Messages, do I need osd configured? Because I have non >> and I am not sure if it is requiered ... Also is weird that monmap is not >> taking the new monitor I think I should try to configure the 3 monitors as >> initial monitors an see how it goes >> >> Dunno about your config, but I seem to remember when I decommissioned one >> mon instance and addition of a new on another node that I needed to have >> mon.<id> section in ceph.conf inorder to be able to start the monitor. >> >> ceph.conf snippet: >> >> [osd] >> osd mount options xfs = >> "rw,noatime,nobarrier,logbsize=256k,logbufs=8,allocsize=4M,attr2,delaylog,inode64,noquota" >> keyring = /var/lib/ceph/osd/ceph-$id/keyring >> ; Tuning >> ;# By default, Ceph makes 3 replicas of objects. If you want to >> make four >> ;# copies of an object the default value--a primary copy and three >> replica >> ;# copies--reset the default values as shown in 'osd pool default size'. >> ;# If you want to allow Ceph to write a lesser number of copies in a >> degraded >> ;# state, set 'osd pool default min size' to a number less than the >> ;# 'osd pool default size' value. >> >> osd pool default size = 2 # Write an object 2 times. >> osd pool default min size = 1 # Allow writing one copy in a degraded >> state. >> >> ;# Ensure you have a realistic number of placement groups. We recommend >> ;# approximately 100 per OSD. E.g., total number of OSDs multiplied by >> 100 >> ;# divided by the number of replicas (i.e., osd pool default size). So for >> ;# 10 OSDs and osd pool default size = 3, we'd recommend approximately >> ;# (100 * 10) / 3 = 333. >> >> ;# got 24 OSDs => 1200 pg, but this is not a full production site, so >> let's settle for 1024 to lower cpu load >> osd pool default pg num = 1024 >> osd pool default pgp num = 1024 >> >> client cache size = 131072 >> osd client op priority = 40 >> osd op threads = 8 >> osd client message size cap = 512 >> filestore min sync interval = 10 >> filestore max sync interval = 60 >> ;filestore queue max bytes = 10485760 >> ;filestore queue max ops = 50 >> ;filestore queue committing max ops = 500 >> ;filestore queue committing max bytes = 104857600 >> ;filestore op threads = 2 >> recovery max active = 2 >> recovery op priority = 30 >> osd max backfills = 2 >> ; Journal Tuning >> journal size = 5120 >> ;journal max write bytes = 1073714824 >> ;journal max write entries = 10000 >> ;journal queue max ops = 50000 >> ;journal queue max bytes = 10485760000 >> >> >> >> >> [mon.0] >> host = node4 >> mon addr = 10.0.3.4:6789 >> >> [mon.1] >> host = node2 >> mon addr = 10.0.3.2:6789 >> >> [mon.2] >> host = node1 >> mon addr = 10.0.3.1:6789 >> >> >> >> _______________________________________________ >> ceph-users mailing list >> [email protected] >> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com >> >> >
_______________________________________________ ceph-users mailing list [email protected] http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
