https://cds.sun.com/is-bin/INTERSHOP.enfinity/WFS/CDS-CDS_SMI-Site/en_US/-/USD/viewproductdetail-start?productref=rdacdvr60002500-09.03.0bc02.0013-lx-...@cds-cds_smi
rdac


On 8/13/2010 12:05 PM, LaoTsao wrote:
U mean stk 2540?
Iirc one can download drivers from oracle sun site

------- Original message -------
From: David Noriega <[email protected]>
To: [email protected]
Sent: 13.8.'10,  11:51

We have three Sun StorageTek 2150, one connected to the metadata
server and two crossconnected to the two data storage nodes. They are
connected via fiber using the qla2xxx driver that comes with CentOS
5.5.  The multipath daemon has the following config:

defaults {
        udev_dir                /dev
        polling_interval        10
        selector                "round-robin 0"
        path_grouping_policy    multibus
        getuid_callout          "/sbin/scsi_id -g -u -s /block/%n"
        prio_callout "/sbin/mpath_prio_rdac /dev/%n"
        path_checker            rdac
        rr_min_io               100
        max_fds                 8192
        rr_weight               priorities
        failback                immediate
        no_path_retry           fail
        user_friendly_names     yes
}

Comment out from multipath.conf file:

blacklist {
        devnode "*"
}


On Fri, Aug 13, 2010 at 4:31 AM, Wojciech Turek <[email protected]> wrote:
Hi David,

I have seen simmilar errors given out by some storage arrays. There were caused by arrays exporting volumes via more then a single path without multi
path driver installed or configured properly. Some times the array
controllers requires a special driver to be installed on Linux host (for example RDAC mpp driver) to properly present and handle configured volumes
in the OS. What sort of disk raid array are you using?

Best gerads,

Wojciech

On 12 August 2010 17:58, David Noriega <[email protected]> wrote:

We just setup a lustre system, and all looks good, but there is this
nagging error thats floating about. When I reboot any of the nodes, be
it a OSS or MDS, I will get this:

[r...@meta1 ~]# dmesg | grep sdc
sdc : very big device. try to use READ CAPACITY(16).
SCSI device sdc: 4878622720 512-byte hdwr sectors (2497855 MB)
sdc: Write Protect is off
sdc: Mode Sense: 77 00 10 08
SCSI device sdc: drive cache: write back w/ FUA
sdc : very big device. try to use READ CAPACITY(16).
SCSI device sdc: 4878622720 512-byte hdwr sectors (2497855 MB)
sdc: Write Protect is off
sdc: Mode Sense: 77 00 10 08
SCSI device sdc: drive cache: write back w/ FUA
 sdc:end_request: I/O error, dev sdc, sector 0
Buffer I/O error on device sdc, logical block 0
end_request: I/O error, dev sdc, sector 0

This doesn't seem to affect anything. fdisk -l doesn't even report the
device. The same(thought of course different block device sdd, sde, on
the OSSs), happens on all the nodes.

If I run pvdisplay or lvdisplay, I'll get this:
/dev/sdc: read failed after 0 of 4096 at 0: Input/output error

Any ideas?
David
--
Personally, I liked the university. They gave us money and facilities,
we didn't have to produce anything! You've never been out of college!
You don't know what it's like out there! I've worked in the private
sector. They expect results. -Ray Ghostbusters
_______________________________________________
Lustre-discuss mailing list
[email protected]
http://lists.lustre.org/mailman/listinfo/lustre-discuss



--
Wojciech Turek

Senior System Architect

High Performance Computing Service
University of Cambridge
Email: [email protected]
Tel: (+)44 1223 763517




--
Personally, I liked the university. They gave us money and facilities,
we didn't have to produce anything! You've never been out of college!
You don't know what it's like out there! I've worked in the private
sector. They expect results. -Ray Ghostbusters
_______________________________________________
Lustre-discuss mailing list
[email protected]
http://lists.lustre.org/mailman/listinfo/lustre-discuss

<<attachment: laotsao.vcf>>

_______________________________________________
Lustre-discuss mailing list
[email protected]
http://lists.lustre.org/mailman/listinfo/lustre-discuss

Reply via email to