[Lustre-discuss] Future of LusterFS?
Hi, Today we have a storage system based on NFS, but we are really concerned about redundancy and are at the brink to take the step to a cluster file system as glusterfs, but we have got suggestions on that lusterfs would have been the best option for us, but at the same time those who recommended lusterfs has said that Oracle has pulled the plug and put the resources into OCFS2. If using lusterfs in a production environment, it would be good to know that it won't be discontinued. Will there be a long term future for lusterfs? Or should we be looking for something else for a long term solution? Thanks in advance for your reply for my a bit cloudy question. -- Janne Aho (Developer) | City Network Hosting AB - www.citynetwork.se Phone: +46 455 690022 | Cell: +46 733 312775 EMail/MSN: ja...@citynetwork.se ICQ: 567311547 | Skype: janne_mz | AIM: janne4cn | Gadu: 16275665 ___ Lustre-discuss mailing list Lustre-discuss@lists.lustre.org http://lists.lustre.org/mailman/listinfo/lustre-discuss
Re: [Lustre-discuss] Future of LusterFS?
Am Donnerstag, 22. April 2010 08:33:14 schrieb Janne Aho: Hi, Today we have a storage system based on NFS, but we are really concerned about redundancy and are at the brink to take the step to a cluster file system as glusterfs, but we have got suggestions on that lusterfs would have been the best option for us, but at the same time those who recommended lusterfs has said that Oracle has pulled the plug and put the resources into OCFS2. If using lusterfs in a production environment, it would be good to know that it won't be discontinued. Will there be a long term future for lusterfs? Or should we be looking for something else for a long term solution? Thanks in advance for your reply for my a bit cloudy question. Hi, for me Lustre is a very good option. But you also could consider a system composed from - corosync for the cluster communication - pacemaker as a cluster resource manager - DRBD for the replication of data between nodes in a cluster and - NFS or - OCFS2 or GFS or ... especially the NFS option provides you with a high available NFS server on real cluster stack all managed by pacemaker. Greetings, -- Dr. Michael Schwartzkopff MultiNET Services GmbH Addresse: Bretonischer Ring 7; 85630 Grasbrunn; Germany Tel: +49 - 89 - 45 69 11 0 Fax: +49 - 89 - 45 69 11 21 mob: +49 - 174 - 343 28 75 mail: mi...@multinet.de web: www.multinet.de Sitz der Gesellschaft: 85630 Grasbrunn Registergericht: Amtsgericht München HRB 114375 Geschäftsführer: Günter Jurgeneit, Hubert Martens --- PGP Fingerprint: F919 3919 FF12 ED5A 2801 DEA6 AA77 57A4 EDD8 979B Skype: misch42 ___ Lustre-discuss mailing list Lustre-discuss@lists.lustre.org http://lists.lustre.org/mailman/listinfo/lustre-discuss
Re: [Lustre-discuss] Future of LusterFS?
On 22/04/10 08:56, Michael Schwartzkopff wrote: On 22/04/10 08:53, Andreas Dilger wrote: Am Donnerstag, 22. April 2010 08:33:14 schrieb Janne Aho: Hi, Today we have a storage system based on NFS, but we are really concerned about redundancy and are at the brink to take the step to a cluster file system as glusterfs, but we have got suggestions on that lusterfs would have been the best option for us, but at the same time those who recommended lusterfs has said that Oracle has pulled the plug and put the resources into OCFS2. If using lusterfs in a production environment, it would be good to know that it won't be discontinued. On 22/04/10 08:53, Andreas Dilger wrote: Whoever told you that is mis-informed. We still have our full team developing Lustre at Oracle, and are planning development into the future. Andreas, thanks for the reply. :) Will there be a long term future for lusterfs? Or should we be looking for something else for a long term solution? Thanks in advance for your reply for my a bit cloudy question. On 22/04/10 08:56, Michael Schwartzkopff wrote: for me Lustre is a very good option. But you also could consider a system composed from - corosync for the cluster communication - pacemaker as a cluster resource manager - DRBD for the replication of data between nodes in a cluster Michael, thanks for the advice, we will take a big dive into this. and - NFS or - OCFS2 or GFS or ... especially the NFS option provides you with a high available NFS server on real cluster stack all managed by pacemaker. We have had quite a deal of issues with our setup using NFS, one of the problems has been caused by poor file system selection thanks to the poor selection of file systems in CentOS/RedHat EL. -- Janne Aho (Developer) | City Network Hosting AB - www.citynetwork.se Phone: +46 455 690022 | Cell: +46 733 312775 EMail/MSN: ja...@citynetwork.se ICQ: 567311547 | Skype: janne_mz | AIM: janne4cn | Gadu: 16275665 ___ Lustre-discuss mailing list Lustre-discuss@lists.lustre.org http://lists.lustre.org/mailman/listinfo/lustre-discuss
[Lustre-discuss] Lustre loadgen error
Hi all, I'm trying to test Lustre using loadgen, but got sergmentation fault error: * *I have succesfully added obdecho.ko on both OSS previously [lustre]# loadgen loadgen dev lustre-OST-osc 192.168.11...@o2ib Added uuid OSS_UUID: 192.168.11...@o2ib Target OST name is 'lustre-OST-osc' loadgen st 3 start 0 to 3 loadgen: running thread #1 Segmentation fault Meet same error on both OSS-es and client using any number of clients. What's wrong? _ Thanks, Katya ___ Lustre-discuss mailing list Lustre-discuss@lists.lustre.org http://lists.lustre.org/mailman/listinfo/lustre-discuss
[Lustre-discuss] Newbie Q : 16TB OST with 1.8.2
Hi, I am looking at using Lustre 1.8.2 with Centos 5.3 or 5.4 64bit. I believe this should support 16TB ext4/lsfdisk backed OSTs. Is this correct ? TIA!! B _ {Beto|Norberto|Numard} Meijome With sufficient thrust, pigs fly just fine. However, this is not necessarily a good idea. It is hard to be sure where they are going to land, and it could be dangerous sitting under them as they fly overhead. [RFC1925 - section 2, subsection 3] ___ Lustre-discuss mailing list Lustre-discuss@lists.lustre.org http://lists.lustre.org/mailman/listinfo/lustre-discuss
Re: [Lustre-discuss] Newbie Q : 16TB OST with 1.8.2
On Thu, Apr 22, 2010 at 4:01 PM, Norberto Meijome numard...@gmail.com wrote: I am looking at using Lustre 1.8.2 with Centos 5.3 or 5.4 64bit. I believe this should support 16TB ext4/lsfdisk backed OSTs. Is this correct ? Hi Norberto, yes, it does. I'm using it on Scientific Linux 5.4: # lfs df -h /lustre/fs4/ UUID bytes Used Available Use% Mounted on fs4-MDT_UUID895.9G471.9M844.3G0% /lustre/fs4[MDT:0] fs4-OST_UUID 11.5T 27.1M 10.9T0% /lustre/fs4[OST:0] fs4-OST0001_UUID 11.5T 27.1M 10.9T0% /lustre/fs4[OST:1] fs4-OST0002_UUID 11.5T 27.1M 10.9T0% /lustre/fs4[OST:2] fs4-OST0003_UUID 11.5T 27.1M 10.9T0% /lustre/fs4[OST:3] fs4-OST0004_UUID 11.5T 27.1M 10.9T0% /lustre/fs4[OST:4] fs4-OST0005_UUID 11.5T 27.1M 10.9T0% /lustre/fs4[OST:5] fs4-OST0006_UUID 11.5T 27.1M 10.9T0% /lustre/fs4[OST:6] fs4-OST0007_UUID 11.5T 27.1M 10.9T0% /lustre/fs4[OST:7] filesystem summary: 92.2T216.5M 87.5T0% /lustre/fs4 Regards, Götz -- AL I:40: Do what thou wilt shall be the whole of the Law. ___ Lustre-discuss mailing list Lustre-discuss@lists.lustre.org http://lists.lustre.org/mailman/listinfo/lustre-discuss
[Lustre-discuss] Lustre client bug (?)
Hi, I'm not sure where I should report it but I couldn't find the error text in Google so I guess it's not in bug tracker yet. This appeared on CentOS 64-bit client under light traffic. Lustre 1.8.2 patchless client from Sun, Linux 2.6.28.10 #4 SMP, both without custom patches. I'm not sure what more details I could supply. mx1 kernel: LustreError: 20716:0:(statahead.c:149:ll_sai_entry_cleanup()) ASSERTION(list_empty(entry-se_list)) failed Message from syslogd@ at Thu Apr 22 04:31:50 2010 ... mx1 kernel: LustreError: 20716:0:(statahead.c:149:ll_sai_entry_cleanup()) LBUG -- Andrzej Godziuk http://CloudAccess.net/ ___ Lustre-discuss mailing list Lustre-discuss@lists.lustre.org http://lists.lustre.org/mailman/listinfo/lustre-discuss
Re: [Lustre-discuss] Newbie Q : 16TB OST with 1.8.2
Norberto Meijome wrote: Hi, I am looking at using Lustre 1.8.2 with Centos 5.3 or 5.4 64bit. I believe this should support 16TB ext4/lsfdisk backed OSTs. Is this correct ? Yes, but not by default. Kevin ___ Lustre-discuss mailing list Lustre-discuss@lists.lustre.org http://lists.lustre.org/mailman/listinfo/lustre-discuss
Re: [Lustre-discuss] Newbie Q : 16TB OST with 1.8.2
On 23 April 2010 00:21, Kevin Van Maren kevin.van.ma...@oracle.com wrote: Norberto Meijome wrote: Hi, I am looking at using Lustre 1.8.2 with Centos 5.3 or 5.4 64bit. I believe this should support 16TB ext4/lsfdisk backed OSTs. Is this correct ? Yes, but not by default. Kevin thanks, Götz Kevin. _ {Beto|Norberto|Numard} Meijome With sufficient thrust, pigs fly just fine. However, this is not necessarily a good idea. It is hard to be sure where they are going to land, and it could be dangerous sitting under them as they fly overhead. [RFC1925 - section 2, subsection 3] ___ Lustre-discuss mailing list Lustre-discuss@lists.lustre.org http://lists.lustre.org/mailman/listinfo/lustre-discuss
Re: [Lustre-discuss] Lustre client bug (?)
On Thu, 2010-04-22 at 16:36 +0200, Andrew Godziuk wrote: Hi, I'm not sure where I should report it but I couldn't find the error text in Google so I guess it's not in bug tracker yet. Hrm. I'd not be too sure that Google has indexed the entire of our Bugzilla. Maybe it has, but searching it directly is probably more definitive. This appeared on CentOS 64-bit client under light traffic. Lustre 1.8.2 patchless client from Sun, Linux 2.6.28.10 #4 SMP, both without custom patches. I'm not sure what more details I could supply. mx1 kernel: LustreError: 20716:0:(statahead.c:149:ll_sai_entry_cleanup()) ASSERTION(list_empty(entry-se_list)) failed Message from syslogd@ at Thu Apr 22 04:31:50 2010 ... mx1 kernel: LustreError: 20716:0:(statahead.c:149:ll_sai_entry_cleanup()) LBUG I can't find an existing bug in our bugzilla regarding this ASSERTION/LBUG. ASSERTION/LBUGs are logic conditions that were unexpected. They are also fatal errors that need the node to rebooted to resolve. Can you please a file a bug in our bugzilla about this one. Please attach the syslog from the node that hit the LBUG. Include a few hours of syslog prior to the LBUG if you can. b. signature.asc Description: This is a digitally signed message part ___ Lustre-discuss mailing list Lustre-discuss@lists.lustre.org http://lists.lustre.org/mailman/listinfo/lustre-discuss
Re: [Lustre-discuss] Future of LusterFS?
Make sure you read and understand the Lustre 2.0 release notes before you buy. There seemed to be some specifics in there about using Oracle hardware. In all fairness ... that only matters if you pay Oracle for support. If you aren't paying Oracle for support (or have no plans to), then it doesn't matter. --Ken ___ Lustre-discuss mailing list Lustre-discuss@lists.lustre.org http://lists.lustre.org/mailman/listinfo/lustre-discuss
Re: [Lustre-discuss] Future of LusterFS?
If you read the presentation, at http://wiki.lustre.org/images/6/6f/LUG_Keynote_Presentation-Bojanic-100415.pdf it does say that Lustre 2 will be supported on qualified configurations -- from Oracle and others. Kevin Lundgren, Andrew wrote: Make sure you read and understand the Lustre 2.0 release notes before you buy. There seemed to be some specifics in there about using Oracle hardware. -Original Message- From: lustre-discuss-boun...@lists.lustre.org [mailto:lustre-discuss-boun...@lists.lustre.org] On Behalf Of Andreas Dilger Sent: Thursday, April 22, 2010 12:53 AM To: Janne Aho Cc: lusterfs Subject: Re: [Lustre-discuss] Future of LusterFS? On 2010-04-22, at 00:33, Janne Aho wrote: Today we have a storage system based on NFS, but we are really concerned about redundancy and are at the brink to take the step to a cluster file system as glusterfs, but we have got suggestions on that lusterfs would have been the best option for us, but at the same time those who recommended lusterfs has said that Oracle has pulled the plug and put the resources into OCFS2. Whoever told you that is mis-informed. We still have our full team developing Lustre at Oracle, and are planning development into the future. If using lusterfs in a production environment, it would be good to know that it won't be discontinued. Will there be a long term future for lusterfs? Yes. Cheers, Andreas -- Andreas Dilger Principal Engineer, Lustre Group Oracle Corporation Canada Inc. ___ Lustre-discuss mailing list Lustre-discuss@lists.lustre.org http://lists.lustre.org/mailman/listinfo/lustre-discuss ___ Lustre-discuss mailing list Lustre-discuss@lists.lustre.org http://lists.lustre.org/mailman/listinfo/lustre-discuss ___ Lustre-discuss mailing list Lustre-discuss@lists.lustre.org http://lists.lustre.org/mailman/listinfo/lustre-discuss
Re: [Lustre-discuss] fseeks on lustre
We were able to find where to tune the stat_blksize after loading the patch mentioned below and the fseek function is working correctly with this patch installed. Thanks Rocky After doing some more digging it looks as though a bug was reported on this in 2007. https://bugzilla.lustre.org/show_bug.cgi?id=12739 We have loaded the patch for lustre attached to this bug, however when running the set_param command I am getting the following error. lctl set_param llite*.*.stat_blksize=4096 error: set_param: /proc/{fs,sys}/{lnet,lustre}/llite/lustre*/stat_blksize: No such process Is this patch still valid for 2.6.9-78.0.22.EL_lustre.1.6.7.2smp Thanks again Rocky ___ Lustre-discuss mailing list Lustre-discuss@lists.lustre.org http://lists.lustre.org/mailman/listinfo/lustre-discuss
Re: [Lustre-discuss] Lustre loadgen error
On 2010-04-22, at 05:23, Katya Tutlyaeva wrote: I'm trying to test Lustre using loadgen, but got sergmentation fault error: * *I have succesfully added obdecho.ko on both OSS previously [lustre]# loadgen loadgen dev lustre-OST-osc 192.168.11...@o2ib Added uuid OSS_UUID: 192.168.11...@o2ib Target OST name is 'lustre-OST-osc' loadgen st 3 start 0 to 3 loadgen: running thread #1 Segmentation fault Meet same error on both OSS-es and client using any number of clients. I believe there is a fix for loadgen in bugzilla. Cheers, Andreas -- Andreas Dilger Lustre Technical Lead Oracle Corporation Canada Inc. ___ Lustre-discuss mailing list Lustre-discuss@lists.lustre.org http://lists.lustre.org/mailman/listinfo/lustre-discuss