Lamar, Excellent email. Thank you so much you have been very informative!!!
On 9/1/2011 11:29 AM, Lamar Owen wrote: > On Wednesday, August 31, 2011 09:21:25 PM Jonathan Vomacka wrote: >> I was >> recently told that this is an old style of partitioning and is not used >> in modern day Linux distributions. > > The only thing old-style I saw in your list was the separate /usr partition. > I like having separate /var, /tmp, and /home. /var because lots of data > resides there that can fill partitions quickly; logs, for instance. I have > built machines where /var/log was separate, specifically for that reason. > >> So more accurately, here are my >> questions to the list: >> >> 1) What is a good partition map/schema for a server OS where it's >> primary purpose is for a LAMP server, DNS (bind), and possibly gameservers > > Splitting out filesystems into partitions makes sense primarily, in my > opinion and experience, in seven basic aspects: > 1.) I/O load balancing across multiple spindles and/or controllers; > 2.) Disk space isolation in case of filesystem 'overflow' (that is, you don't > want your mail spool in /var/spool/mail overflowing to be able to corrupt an > online database in, say, /var/lib/pgsql/data/base) (and while quotas can help > with this when two trees are not fully isolated, filesystems in different > partitons/logical volumes have hard overflow isolation); > 3.) In the case of really large data stores with dynamic data, isolating the > impact of filesystem corruption; > 4.) The ability to stagger fsck's between boots (the fsck time doesn't seem > to increase linearly with filesystem size); > 5.) I/O 'tiering' (like EMC's FAST) where you can allocate your fastest > storage to the most rapidly changing data, and slower storage to data that > doesn't change frequently; > 6.) Putting things into separate filesystem forces the admin to really have > to think through and design the system taking into account all the > requirements, instead of just throwing it all together and then wondering why > performance is suboptimal; > 7.) Filesystems can be mounted with options specific to their use cases, and > using filesystem technology appropriate to the use case (noexec, for > instance, on filesystems that have no business having executables on them; > enabling/disabling journalling and other options as appropriate, and using > XFS, ext4, etc as appropriate, just to mentiona a few things). > >> 2) CentOS docs recommend using 10GB SWAP for 8GB of RAM. 1X the amount >> of physical memory + 2GB added. > > If you put swap on LVM and give yourself room to grow you can increase swap > space size at will if you should find you need to do so. Larger RAM (and > virtual RAM embodied by swap) does not always make things faster. I have a > private e-mail from an admin of a large website showing severe MySQL > performance issues that were reduced by making the RAM size smaller (turned > out to be a caching mismanagement thing with poorly written queries that > caused it). > > Consider swap to be a safety buffer; the Linux kernel is by default > configured to overcommit memory, and swap exists to prevet the oom-killer > from reaping critical processes in this situation. Tuning swap size and the > 'swappiness' of the kernel along with the overcommit policy should be done > together; the default settings produce the default recommendation of 'memory > size plus 2GB' that was for CentOS 5. Not too long ago, the recommendation > was for swap to be twice the memory size. > > Multiple swap partitions can improve performance if those partitions are on > different spindles; however, this reduces reliability, too. I don't have any > experience with benchmarking the performance of multiple 2GB swap spaces; I'd > find results of such benchmarks to be useful information. > >> 3) Is EXT4 better or worse to use then XFS for what I am planning to use >> the system for? > > That depends; consult some file system comparisons (the wikipedia file system > comparison article is a good starting place). I've used both; and I still > use both. XFS as a filesystem is older and presumably more mature than ext4, > but age is not the only indicator of something that will work for you. One > thing to remember is that XFS filesystems cannot currently be reduced in > size, only increased. Ext4 can go either way if you realize you made too > large of a filesystem. > > XFS is very fast to create, but repairing requires absolutely the most RAM of > any recovery process I've ever seen. XFS has seen a lot of use in the field, > particularly with large SGI boxes (Altix series, primarily) running Linux > (with the requisite 'lots of RAM' required for repair/recovery.... > > XFS currently is the only one where I have successfully made a large than > 16TB filesystem. Don't try that on a 32-bit system (in fact, if you care > about data integrity, don't use XFS on a 32-bit system at all, unless you > have rebuilt the kernel with 8k stacks). The mkfs.xfs on a greater than 16TB > partition/logical volume will execute successfully on a 32-bit system (the > last time I tried it), but as soon as you go over 16TB with your data you > will no longer be able to mount the filesystem. The wisdom of making a > greater than 16TB filesystem of any type is left as an exercise for the > reader.... > _______________________________________________ > CentOS mailing list > [email protected] > http://lists.centos.org/mailman/listinfo/centos _______________________________________________ CentOS mailing list [email protected] http://lists.centos.org/mailman/listinfo/centos

