[My previous message was somewhat garbled when reflected back at me. It looks
better in the archives here:
https://marc.info/?l=openbsd-misc&m=161902769301731&w=2. I’m resending as
plain-text to see if the problem is on my end.]
I’m running OpenBSD on top of bHyve using virtual disks allocated out of ZFS
pools. While not the same setup, some concepts carry over...
I have two types of pools:
1) an “expensive" pool for fast random IO:
- this pool is made up stripes of SSD-based vdevs.
- ZFS is configured to use a 16K recordsize for this pool.
- good for small files (guest OS, DBs, web/mail/dns files, etc.)
- When ZFS is told to use the SSD, it starts the partition
on sector 256 (not the default sector 34) to ensure good
SSD NAND alignment.
2) a less-expensive pool for large sequential IO:
- this pool is a single RAIDZ2-based vdev using spinning rust.
- ZFS is configured to use a 1M recordsize for this pool.
- good for large files (movies, high-res images, backups, etc.)
Virtual disks are exposed to the OpenBSD guests from both pools. The guest’s
root-disk is always allocated from pool #1. Typically, a second
application-specific disk is also allocated from pool #1 (e.g., /var/www/sites
on a web server, /home on a mail server, etc.). Only in special circumstances
(e.g., a media server) is a disk allocated from pool #2.
This arrangement steps around needing to read/write 1M blocks for each small
file access, and also the possibility that a guest accessing a given block will
span more than a single physical block.
Can VMWare virtual disks be configured similarly?
K.