Hi,

  I'm having some problems, and i'm on a tight schedule if some could help i 
apreciate :)

 - I've configured lustre to access a storage resource. We have 2 machines 
connected to it. The config generation script is:

#!/bin/sh

# config.sh 

DEV_MDS='/dev/sds'
DEV_HOME1='/dev/sdr'
DEV_HOME2='/dev/sdq'
DEV_SOFT1='/dev/sdo'
DEV_SOFT2='/dev/sdp'
DEV_VOL1='/dev/sdm'
DEV_VOL2='/dev/sdn'
DEV_DPM1='/dev/sdk'
DEV_DPM2='/dev/sdl'

# Create nodes
rm -f config.xml
lmc -m config.xml --add net --node zephyr.up.pt --nid zephyr.up.pt --nettype tcp
lmc -m config.xml --add net --node hades.up.pt --nid hades.up.pt --nettype tcp

# New client profiles
lmc -m config.xml --add net --node SE-vpool --nid '*' --nettype tcp
lmc -m config.xml --add net --node SE-ppool --nid '*' --nettype tcp
lmc -m config.xml --add net --node VO-soft --nid '*' --nettype tcp
lmc -m config.xml --add net --node homes --nid '*' --nettype tcp
lmc -m config.xml --add net --node shared --nid '*' --nettype tcp

# Cofigure MDS
lmc -m config.xml --add mds --node hades.up.pt --mds mds-polo3 --fstype ext3 
--dev $DEV_MDS

# Configure LOVS
# Homes
lmc -m config.xml --add lov --lov lov-home --mds mds-polo3 --stripe_sz 1048576 
--stripe_cnt 0 --stripe_pattern 0
# VOs software space
lmc -m config.xml --add lov --lov lov-soft --mds mds-polo3 --stripe_sz 1048576 
--stripe_cnt 0 --stripe_pattern 0
# Volatil space for DPM
lmc -m config.xml --add lov --lov lov-tmp --mds mds-polo3 --stripe_sz 1048576 
--stripe_cnt 0 --stripe_pattern 0
# DPM space
lmc -m config.xml --add lov --lov lov-dpm --mds mds-polo3 --stripe_sz 1048576 
--stripe_cnt 0 --stripe_pattern 0

# Configure OSTs
# Homes
lmc -m config.xml --add ost --node hades.up.pt --lov lov-home --ost home-hades 
--fstype ldiskfs --dev $DEV_HOME1
lmc -m config.xml --add ost --node zephyr.up.pt --lov lov-home --ost 
home-zephyr --fstype ldiskfs --dev $DEV_HOME2

# Soft
lmc -m config.xml --add ost --node hades.up.pt --lov lov-soft --ost soft-hades 
--fstype ldiskfs --dev $DEV_SOFT1
lmc -m config.xml --add ost --node zephyr.up.pt --lov lov-soft --ost 
soft-zephyr --fstype ldiskfs --dev $DEV_SOFT2

# Volatil
lmc -m config.xml --add ost --node hades.up.pt --lov lov-tmp --ost tmp-hades 
--fstype ldiskfs --dev $DEV_VOL1
lmc -m config.xml --add ost --node zephyr.up.pt --lov lov-tmp --ost tmp-zephyr 
--fstype ldiskfs --dev $DEV_VOL2

# DPM
lmc -m config.xml --add ost --node hades.up.pt --lov lov-dpm --ost dpm-hades 
--fstype ldiskfs --dev $DEV_DPM1
lmc -m config.xml --add ost --node zephyr.up.pt --lov lov-dpm --ost dpm-zephyr 
--fstype ldiskfs --dev $DEV_DPM2

# homes client profile
lmc -m config.xml --add mtpt --node homes --path /local/home --mds mds-polo3 
--lov lov-home
# Vo software share client profile
lmc -m config.xml --add mtpt --node VO-soft --path /vosoft --mds mds-polo3 
--lov lov-soft
# Volatil pool client profile
lmc -m config.xml --add mtpt --node SE-vpool --path /mnt/vpool --mds mds-polo3 
--lov lov-tmp
# Permanent pool client profile
lmc -m config.xml --add mtpt --node SE-ppool --path /mnt/ppool --mds mds-polo3 
--lov lov-dpm

I've add it to fstab the mount points for the clients:
# Lustre mount points
hades.up.pt:/mds-polo3/SE-vpool         /mnt/vpool      lustre  rw      0 0
hades.up.pt:/mds-polo3/SE-ppool         /mnt/ppool      lustre  rw      0 0
hades.up.pt:/mds-polo3/VO-soft          /vosoft         lustre  rw      0 0
hades.up.pt:/mds-polo3/homes            /local/home     lustre  rw      0 0

And doing a 'df -h' i can see the devices are mounted:

hades.up.pt:/mds-polo3/SE-vpool
                       50G  153M   47G   1% /mnt/vpool
hades.up.pt:/mds-polo3/SE-ppool
                      1.2T  209M  1.1T   1% /mnt/ppool
hades.up.pt:/mds-polo3/VO-soft
                      306G  185M  290G   1% /vosoft
hades.up.pt:/mds-polo3/homes
                      591G  193M  561G   1% /local/home


Now. if i go to one of the partition lets say '/local/home' and create some 
file, this file is created also in the other devices.!

doing "lfs df -h"

UUID                     bytes      Used Available  Use% Mounted on
mds-polo3_UUID           39.4G      2.1G     37.3G    5% /mnt/vpool[MDT:0]
tmp-hades_UUID           24.6G      1.3G     23.3G    5% /mnt/vpool[OST:0]
tmp-zephyr_UUID          24.6G      1.3G     23.3G    5% /mnt/vpool[OST:1]

filesystem summary:      49.2G      2.6G     46.6G    5% /mnt/vpool

UUID                     bytes      Used Available  Use% Mounted on
mds-polo3_UUID           39.4G      2.1G     37.3G    5% /mnt/ppool[MDT:0]
dpm-hades_UUID          564.5G     28.8G    535.7G    5% /mnt/ppool[OST:0]
dpm-zephyr_UUID         564.5G     28.8G    535.7G    5% /mnt/ppool[OST:1]

filesystem summary:       1.1T     57.6G      1.0T    5% /mnt/ppool

UUID                     bytes      Used Available  Use% Mounted on
mds-polo3_UUID           39.4G      2.1G     37.3G    5% /vosoft[MDT:0]
soft-hades_UUID         152.6G      7.8G    144.7G    5% /vosoft[OST:0]
soft-zephyr_UUID        152.6G      7.8G    144.7G    5% /vosoft[OST:1]

filesystem summary:     305.1G     15.7G    289.5G    5% /vosoft

UUID                     bytes      Used Available  Use% Mounted on
mds-polo3_UUID           39.4G      2.1G     37.3G    5% /local/home[MDT:0]
home-hades_UUID         295.3G     15.1G    280.2G    5% /local/home[OST:0]
home-zephyr_UUID        295.3G     15.1G    280.2G    5% /local/home[OST:1]

filesystem summary:     590.6G     30.2G    560.4G    5% /local/home

The only think i can think of is, that i use the same mds for each lov. Do i 
have to create a MDS for each lov ?

  Thanks in advance

-- 
Rui Ramos
==============================================
Universidade do Porto - IRICUP
Praça Gomes Teixeira, 4099-002 Porto, Portugal
email: rramos[at]reit.up.pt
phone: +351 22 040 8164
==============================================


_______________________________________________
Lustre-discuss mailing list
[email protected]
https://mail.clusterfs.com/mailman/listinfo/lustre-discuss

Reply via email to