Hi,
I'm having some problems, and i'm on a tight schedule if some could help i
apreciate :)
- I've configured lustre to access a storage resource. We have 2 machines
connected to it. The config generation script is:
#!/bin/sh
# config.sh
DEV_MDS='/dev/sds'
DEV_HOME1='/dev/sdr'
DEV_HOME2='/dev/sdq'
DEV_SOFT1='/dev/sdo'
DEV_SOFT2='/dev/sdp'
DEV_VOL1='/dev/sdm'
DEV_VOL2='/dev/sdn'
DEV_DPM1='/dev/sdk'
DEV_DPM2='/dev/sdl'
# Create nodes
rm -f config.xml
lmc -m config.xml --add net --node zephyr.up.pt --nid zephyr.up.pt --nettype tcp
lmc -m config.xml --add net --node hades.up.pt --nid hades.up.pt --nettype tcp
# New client profiles
lmc -m config.xml --add net --node SE-vpool --nid '*' --nettype tcp
lmc -m config.xml --add net --node SE-ppool --nid '*' --nettype tcp
lmc -m config.xml --add net --node VO-soft --nid '*' --nettype tcp
lmc -m config.xml --add net --node homes --nid '*' --nettype tcp
lmc -m config.xml --add net --node shared --nid '*' --nettype tcp
# Cofigure MDS
lmc -m config.xml --add mds --node hades.up.pt --mds mds-polo3 --fstype ext3
--dev $DEV_MDS
# Configure LOVS
# Homes
lmc -m config.xml --add lov --lov lov-home --mds mds-polo3 --stripe_sz 1048576
--stripe_cnt 0 --stripe_pattern 0
# VOs software space
lmc -m config.xml --add lov --lov lov-soft --mds mds-polo3 --stripe_sz 1048576
--stripe_cnt 0 --stripe_pattern 0
# Volatil space for DPM
lmc -m config.xml --add lov --lov lov-tmp --mds mds-polo3 --stripe_sz 1048576
--stripe_cnt 0 --stripe_pattern 0
# DPM space
lmc -m config.xml --add lov --lov lov-dpm --mds mds-polo3 --stripe_sz 1048576
--stripe_cnt 0 --stripe_pattern 0
# Configure OSTs
# Homes
lmc -m config.xml --add ost --node hades.up.pt --lov lov-home --ost home-hades
--fstype ldiskfs --dev $DEV_HOME1
lmc -m config.xml --add ost --node zephyr.up.pt --lov lov-home --ost
home-zephyr --fstype ldiskfs --dev $DEV_HOME2
# Soft
lmc -m config.xml --add ost --node hades.up.pt --lov lov-soft --ost soft-hades
--fstype ldiskfs --dev $DEV_SOFT1
lmc -m config.xml --add ost --node zephyr.up.pt --lov lov-soft --ost
soft-zephyr --fstype ldiskfs --dev $DEV_SOFT2
# Volatil
lmc -m config.xml --add ost --node hades.up.pt --lov lov-tmp --ost tmp-hades
--fstype ldiskfs --dev $DEV_VOL1
lmc -m config.xml --add ost --node zephyr.up.pt --lov lov-tmp --ost tmp-zephyr
--fstype ldiskfs --dev $DEV_VOL2
# DPM
lmc -m config.xml --add ost --node hades.up.pt --lov lov-dpm --ost dpm-hades
--fstype ldiskfs --dev $DEV_DPM1
lmc -m config.xml --add ost --node zephyr.up.pt --lov lov-dpm --ost dpm-zephyr
--fstype ldiskfs --dev $DEV_DPM2
# homes client profile
lmc -m config.xml --add mtpt --node homes --path /local/home --mds mds-polo3
--lov lov-home
# Vo software share client profile
lmc -m config.xml --add mtpt --node VO-soft --path /vosoft --mds mds-polo3
--lov lov-soft
# Volatil pool client profile
lmc -m config.xml --add mtpt --node SE-vpool --path /mnt/vpool --mds mds-polo3
--lov lov-tmp
# Permanent pool client profile
lmc -m config.xml --add mtpt --node SE-ppool --path /mnt/ppool --mds mds-polo3
--lov lov-dpm
I've add it to fstab the mount points for the clients:
# Lustre mount points
hades.up.pt:/mds-polo3/SE-vpool /mnt/vpool lustre rw 0 0
hades.up.pt:/mds-polo3/SE-ppool /mnt/ppool lustre rw 0 0
hades.up.pt:/mds-polo3/VO-soft /vosoft lustre rw 0 0
hades.up.pt:/mds-polo3/homes /local/home lustre rw 0 0
And doing a 'df -h' i can see the devices are mounted:
hades.up.pt:/mds-polo3/SE-vpool
50G 153M 47G 1% /mnt/vpool
hades.up.pt:/mds-polo3/SE-ppool
1.2T 209M 1.1T 1% /mnt/ppool
hades.up.pt:/mds-polo3/VO-soft
306G 185M 290G 1% /vosoft
hades.up.pt:/mds-polo3/homes
591G 193M 561G 1% /local/home
Now. if i go to one of the partition lets say '/local/home' and create some
file, this file is created also in the other devices.!
doing "lfs df -h"
UUID bytes Used Available Use% Mounted on
mds-polo3_UUID 39.4G 2.1G 37.3G 5% /mnt/vpool[MDT:0]
tmp-hades_UUID 24.6G 1.3G 23.3G 5% /mnt/vpool[OST:0]
tmp-zephyr_UUID 24.6G 1.3G 23.3G 5% /mnt/vpool[OST:1]
filesystem summary: 49.2G 2.6G 46.6G 5% /mnt/vpool
UUID bytes Used Available Use% Mounted on
mds-polo3_UUID 39.4G 2.1G 37.3G 5% /mnt/ppool[MDT:0]
dpm-hades_UUID 564.5G 28.8G 535.7G 5% /mnt/ppool[OST:0]
dpm-zephyr_UUID 564.5G 28.8G 535.7G 5% /mnt/ppool[OST:1]
filesystem summary: 1.1T 57.6G 1.0T 5% /mnt/ppool
UUID bytes Used Available Use% Mounted on
mds-polo3_UUID 39.4G 2.1G 37.3G 5% /vosoft[MDT:0]
soft-hades_UUID 152.6G 7.8G 144.7G 5% /vosoft[OST:0]
soft-zephyr_UUID 152.6G 7.8G 144.7G 5% /vosoft[OST:1]
filesystem summary: 305.1G 15.7G 289.5G 5% /vosoft
UUID bytes Used Available Use% Mounted on
mds-polo3_UUID 39.4G 2.1G 37.3G 5% /local/home[MDT:0]
home-hades_UUID 295.3G 15.1G 280.2G 5% /local/home[OST:0]
home-zephyr_UUID 295.3G 15.1G 280.2G 5% /local/home[OST:1]
filesystem summary: 590.6G 30.2G 560.4G 5% /local/home
The only think i can think of is, that i use the same mds for each lov. Do i
have to create a MDS for each lov ?
Thanks in advance
--
Rui Ramos
==============================================
Universidade do Porto - IRICUP
Praça Gomes Teixeira, 4099-002 Porto, Portugal
email: rramos[at]reit.up.pt
phone: +351 22 040 8164
==============================================
_______________________________________________
Lustre-discuss mailing list
[email protected]
https://mail.clusterfs.com/mailman/listinfo/lustre-discuss