Shailendra Mudgal wrote: > Hi , > We have upgraded our code to nutch-0.9 with hadoop-0.12.2-core.jar. After > running say 50 nutch jobs(which includes inject/generate/fetch/parse etc.) > we start getting "Too many open files" error on our cluster. We are using > Linux box with kernel 2.6.9 and the open files number is 1024 on these > machine which is default. I read several mails from nutch-user, or > hadoop-user mailing lists. And i found only way was to increase the number > of open files using ulimit. Is there any other solution for this problem at > code level. BTW the value for io.sort.factor is 8 in our hadoop-site.xml. > > Is anybody having any idea in this regard? Any help will be appreciated.
Apparently datanodes that perform intensive IO operations need a higher limit. Try increasing this number to 16k or so. -- Best regards, Andrzej Bialecki <>< ___. ___ ___ ___ _ _ __________________________________ [__ || __|__/|__||\/| Information Retrieval, Semantic Web ___|||__|| \| || | Embedded Unix, System Integration http://www.sigram.com Contact: info at sigram dot com ------------------------------------------------------------------------- This SF.net email is sponsored by DB2 Express Download DB2 Express C - the FREE version of DB2 express and take control of your XML. No limits. Just data. Click to get it now. http://sourceforge.net/powerbar/db2/ _______________________________________________ Nutch-general mailing list [email protected] https://lists.sourceforge.net/lists/listinfo/nutch-general
