[root@sgt-pepper ~]# mount
sysfs on /sys type sysfs (rw,nosuid,nodev,noexec,relatime,seclabel)
proc on /proc type proc (rw,nosuid,nodev,noexec,relatime)
devtmpfs on /dev type devtmpfs
(rw,nosuid,seclabel,size=3991388k,nr_inodes=997847,mode=755)
securityfs on /sys/kernel/security type securityfs
(rw,nosuid,nodev,noexec,relatime)
tmpfs on /dev/shm type tmpfs (rw,nosuid,nodev,seclabel)
devpts on /dev/pts type devpts
(rw,nosuid,noexec,relatime,seclabel,gid=5,mode=620,ptmxmode=000)
tmpfs on /run type tmpfs (rw,nosuid,nodev,seclabel,mode=755)
tmpfs on /sys/fs/cgroup type tmpfs
(ro,nosuid,nodev,noexec,seclabel,mode=755)
cgroup on /sys/fs/cgroup/systemd type cgroup
(rw,nosuid,nodev,noexec,relatime,seclabel,xattr,release_agent=/usr/lib/systemd/systemd-cgroups-agent,name=systemd)
pstore on /sys/fs/pstore type pstore (rw,nosuid,nodev,noexec,relatime)
cgroup on /sys/fs/cgroup/cpuset type cgroup
(rw,nosuid,nodev,noexec,relatime,seclabel,cpuset)
cgroup on /sys/fs/cgroup/hugetlb type cgroup
(rw,nosuid,nodev,noexec,relatime,seclabel,hugetlb)
cgroup on /sys/fs/cgroup/memory type cgroup
(rw,nosuid,nodev,noexec,relatime,seclabel,memory)
cgroup on /sys/fs/cgroup/blkio type cgroup
(rw,nosuid,nodev,noexec,relatime,seclabel,blkio)
cgroup on /sys/fs/cgroup/cpu,cpuacct type cgroup
(rw,nosuid,nodev,noexec,relatime,seclabel,cpuacct,cpu)
cgroup on /sys/fs/cgroup/pids type cgroup
(rw,nosuid,nodev,noexec,relatime,seclabel,pids)
cgroup on /sys/fs/cgroup/freezer type cgroup
(rw,nosuid,nodev,noexec,relatime,seclabel,freezer)
cgroup on /sys/fs/cgroup/net_cls,net_prio type cgroup
(rw,nosuid,nodev,noexec,relatime,seclabel,net_prio,net_cls)
cgroup on /sys/fs/cgroup/devices type cgroup
(rw,nosuid,nodev,noexec,relatime,seclabel,devices)
cgroup on /sys/fs/cgroup/perf_event type cgroup
(rw,nosuid,nodev,noexec,relatime,seclabel,perf_event)
configfs on /sys/kernel/config type configfs (rw,relatime)
/dev/xvda2 on / type ext3 (rw,noatime,seclabel,data=ordered)
selinuxfs on /sys/fs/selinux type selinuxfs (rw,relatime)
mqueue on /dev/mqueue type mqueue (rw,relatime,seclabel)
debugfs on /sys/kernel/debug type debugfs (rw,relatime)
hugetlbfs on /dev/hugepages type hugetlbfs (rw,relatime,seclabel)
/dev/xvda1 on /boot type ext3 (rw,noatime,seclabel,data=ordered)
none on /proc/xen type xenfs (rw,relatime)
systemd-1 on /proc/sys/fs/binfmt_misc type autofs
(rw,relatime,fd=28,pgrp=1,timeout=0,minproto=5,maxproto=5,direct,pipe_ino=28876)
overlay on
/var/lib/docker/overlay2/6dadbda2c0217397cdc88f043163a31abdef1e618db43e1187caf3f88e4be9c5/merged
type overlay
(rw,relatime,seclabel,lowerdir=/var/lib/docker/overlay2/l/CVJZV2GQR7FFUFKDATWUC5XOHD:/var/lib/docker/overlay2/l/3TADZQEQKTFXPQHJ7HLQJH3YBP:/var/lib/docker/overlay2/l/AAIXB6R7K6R7WP76YQY4MNX6DB:/var/lib/docker/overlay2/l/HBG67PHTIV7E4TOGCV5565CJVU:/var/lib/docker/overlay2/l/S3A7TWP3B6UWEWAYG4E33DZC2O:/var/lib/docker/overlay2/l/D5W2VRAGMJU7ED26WDAMKE54MY:/var/lib/docker/overlay2/l/KUZMRHL4IAYCEBT74BBKVDC6IQ:/var/lib/docker/overlay2/l/HYGNGZ6MV3C5YVL2TCXT7E55PS:/var/lib/docker/overlay2/l/ZESSLGAWK2ZYDV3IU2A2RB736Q:/var/lib/docker/overlay2/l/K43FJ3RNBAUET3G45SI47KHUQK:/var/lib/docker/overlay2/l/Q26VC3ZXOGXBYHUANTP3WDI5X6:/var/lib/docker/overlay2/l/K6LQYQTUY75DFZH6NZ34S65GAQ:/var/lib/docker/overlay2/l/E3DMUWQXEBRL5LCV7Z5UK4N6GK,upperdir=/var/lib/docker/overlay2/6dadbda2c0217397cdc88f043163a31abdef1e618db43e1187caf3f88e4be9c5/diff,workdir=/var/lib/docker/overlay2/6dadbda2c0217397cdc88f043163a31abdef1e618db43e1187caf3f88e4be9c5/work)
proc on /run/docker/netns/default type proc
(rw,nosuid,nodev,noexec,relatime)
shm on
/var/lib/docker/containers/4f3bff2fbb9fa38110d3ae70a86a2a2f0a999299597c8d0c7f072eca30d8b186/mounts/shm
type tmpfs (rw,nosuid,nodev,noexec,relatime,seclabel,size=65536k)
overlay on
/var/lib/docker/overlay2/0974ad067d2681ce88bc2a5fe0dd4c27a06921b4db976bfb7fa7e95c9eca1800/merged
type overlay
(rw,relatime,seclabel,lowerdir=/var/lib/docker/overlay2/l/ABKWJRSETYWLWIFWINXQALMKHA:/var/lib/docker/overlay2/l/3TADZQEQKTFXPQHJ7HLQJH3YBP:/var/lib/docker/overlay2/l/AAIXB6R7K6R7WP76YQY4MNX6DB:/var/lib/docker/overlay2/l/HBG67PHTIV7E4TOGCV5565CJVU:/var/lib/docker/overlay2/l/S3A7TWP3B6UWEWAYG4E33DZC2O:/var/lib/docker/overlay2/l/D5W2VRAGMJU7ED26WDAMKE54MY:/var/lib/docker/overlay2/l/KUZMRHL4IAYCEBT74BBKVDC6IQ:/var/lib/docker/overlay2/l/HYGNGZ6MV3C5YVL2TCXT7E55PS:/var/lib/docker/overlay2/l/ZESSLGAWK2ZYDV3IU2A2RB736Q:/var/lib/docker/overlay2/l/K43FJ3RNBAUET3G45SI47KHUQK:/var/lib/docker/overlay2/l/Q26VC3ZXOGXBYHUANTP3WDI5X6:/var/lib/docker/overlay2/l/K6LQYQTUY75DFZH6NZ34S65GAQ:/var/lib/docker/overlay2/l/E3DMUWQXEBRL5LCV7Z5UK4N6GK,upperdir=/var/lib/docker/overlay2/0974ad067d2681ce88bc2a5fe0dd4c27a06921b4db976bfb7fa7e95c9eca1800/diff,workdir=/var/lib/docker/overlay2/0974ad067d2681ce88bc2a5fe0dd4c27a06921b4db976bfb7fa7e95c9eca1800/work)
shm on
/var/lib/docker/containers/6ec82c01ef8fcf8bae69fd55b7fe82ee7e685e28a9331a4f1aeb7425e670a433/mounts/shm
type tmpfs (rw,nosuid,nodev,noexec,relatime,seclabel,size=65536k)
tmpfs on /run/user/0 type tmpfs
(rw,nosuid,nodev,relatime,seclabel,size=800344k,mode=700)
[root@sgt-pepper ~]# cat /proc/meminfo
MemTotal:        8003408 kB
MemFree:          778948 kB
MemAvailable:    7202972 kB
Buffers:          235960 kB
Cached:          6170780 kB
SwapCached:            0 kB
Active:          2572556 kB
Inactive:        4205592 kB
Active(anon):     372284 kB
Inactive(anon):     8072 kB
Active(file):    2200272 kB
Inactive(file):  4197520 kB
Unevictable:           0 kB
Mlocked:               0 kB
SwapTotal:       2096444 kB
SwapFree:        2096444 kB
Dirty:                20 kB
Writeback:             0 kB
AnonPages:        371448 kB
Mapped:          2192120 kB
Shmem:              8956 kB
Slab:             366532 kB
SReclaimable:     329032 kB
SUnreclaim:        37500 kB
KernelStack:        3456 kB
PageTables:         9948 kB
NFS_Unstable:          0 kB
Bounce:                0 kB
WritebackTmp:          0 kB
CommitLimit:     6098148 kB
Committed_AS:    1331684 kB
VmallocTotal:   34359738367 kB
VmallocUsed:       20280 kB
VmallocChunk:   34359715812 kB
HardwareCorrupted:     0 kB
AnonHugePages:    239616 kB
CmaTotal:              0 kB
CmaFree:               0 kB
HugePages_Total:       0
HugePages_Free:        0
HugePages_Rsvd:        0
HugePages_Surp:        0
Hugepagesize:       2048 kB
DirectMap4k:      145408 kB
DirectMap2M:     6141952 kB
DirectMap1G:     2097152 kB

By increased buffer size, you mean in both Crail conf and Java client?

Lou.

On Thu, Mar 5, 2020 at 9:08 AM Adrian Schuepbach <
adrian.schuepb...@gribex.net> wrote:

> Hi Lou
>
> Thanks. Can you show the output of `mount` and `cat /proc/meminfo`?
>
> Can you measure again with an increased buffer size?
>
> Thanks
> Adrian
>
> On 3/5/20 14:48, Lou DeGenaro wrote:
> > Hi Adrian,
> >
> > re: iperf
> >
> > iperf -s
> > iperf -c <hostname>
> >
> > re: bandwidth
> >
> > bandwidth is calculated offline, using the seconds printed and the 1 GB
> > known file size.  The 1 GB file is generated using:
> >
> > base64 /dev/urandom | head -c 1073741824 > G1.txt
> >
> > and is stored into Crail using:
> >
> > bin/crail fs -copyFromLocal /root/G1.txt /G1.txt
> >
> >
> > re: Crail configuration
> >
> > crail.namenode.address            crail://sgt-pepper:9060
> > crail.cachepath                   /tmp/hugepages/cache
> > crail.cachelimit                  1073741824
> > crail.storage.tcp.interface       eth0
> > crail.storage.tcp.datapath        /tmp/hugepages/data
> > crail.storage.tcp.storagelimit    1073741824
> >
> > I did try using iobench sometime ago without success.  See my newsgroup
> > post 2020/02/07
> >
> > Thanks for your help and interest.
> >
> > Lou.
> >
> >
> > On Thu, Mar 5, 2020 at 8:26 AM Adrian Schuepbach <
> > adrian.schuepb...@gribex.net> wrote:
> >
> >> Hi Lou
> >>
> >> It is hard to say without knowing more details.
> >>
> >> Can you post the exact iperf command you used?
> >>
> >> Also the Crail code you show seems not to be the one you
> >> use for the measurements, at least I don't see where the code
> >> computes the bandwidth. Can you post the actual code you are using?
> >>
> >> Have you configured Crail to store data on hugepages? Or will
> >> it access disks?
> >>
> >> There is also iobench, a performance measurement tool that comes
> >> with Crail. Have you tried measuring with this one?
> >>
> >>
> >> Best
> >> Adrian
> >>
> >> On 3/5/20 14:01, Lou DeGenaro wrote:
> >>> Hello,
> >>>
> >>> I'm comparing Crail to iperf.  VM-A is used to run the server(s).  VM-B
> >> is
> >>> used to run the client.  Both VMs are CentOS 7 with 8 GB memory + 2
> CPUs.
> >>> Tests runs are non-overlapping.
> >>>
> >>> For the Crail case , a 1 GB file is posted to server and a simple Java
> >>> client is employed (see below).
> >>>
> >>> Results:
> >>>
> >>> iperf: 4.05 Gb/seb
> >>> Crail: 2.52 Gb/sec
> >>>
> >>> Why is iperf so much better??
> >>>
> >>> Lou.
> >>>
> >>> -----
> >>>
> >>> Crail Java Client:
> >>>
> >>> public static void main(String[] args) {
> >>> try {
> >>> //initialize
> >>> String filename = "/G1.txt";
> >>> int filesize = 1024*1024*1024;
> >>> int bufsize = 1024*1024;
> >>> CrailBuffer buf =
> OffHeapBuffer.wrap(ByteBuffer.allocateDirect(bufsize));
> >>> CrailConfiguration cconf =
> >> CrailConfiguration.createConfigurationFromFile();
> >>> CrailStore cstore = CrailStore.newInstance(cconf);
> >>> CrailFile file = cstore.lookup(filename).get().asFile();
> >>> CrailInputStream directStream =
> >>> file.getDirectInputStream(file.getCapacity());
> >>> long sumbytes = 0;
> >>> //run test
> >>> long start = System.currentTimeMillis();
> >>> while (sumbytes < filesize) {
> >>> buf.clear();
> >>> CrailResult cr = directStream.read(buf).get();
> >>> long ret = cr.getLen();
> >>> sumbytes = sumbytes + ret;
> >>> }
> >>> long end = System.currentTimeMillis();
> >>> //print result and clean-up
> >>> double executionTime = ((double) (end - start)) / 1000.0;
> >>> System.out.println("time: "+executionTime);
> >>> cstore.close();
> >>> directStream.close();
> >>> }
> >>> catch(Exception e) {
> >>> e.printStackTrace();
> >>> }
> >>> }
> >>>
> >> --
> >> Adrian Schüpbach, Dr. sc. ETH Zürich
> >>
> >>
> --
> Adrian Schüpbach, Dr. sc. ETH Zürich
>
>

Reply via email to