Shawn Gervais wrote:
Greetings list,

This is my DFS report:

Total raw bytes: 709344133120 (660.62 Gb)
Used raw bytes: 302794461922 (281.99 Gb)
% used: 42.68%

Total effective bytes: 11826067632 (11.01 Gb)
Effective replication multiplier: 25.6039853097637

These numbers seem to me to be completely insane -- a 25 times replication of blocks. I have my replication factor set to 3.

"Used raw bytes" goes up when I run jobs, and if I delete files those jobs produce within DFS (e.g. a segment for a failed fetch), it doesn't appear that hadoop immediately reclaims the space used by the deleted files' blocks.

Am I right? Is this a bug?

What does 'hadoop fsck /' say?

--
Best regards,
Andrzej Bialecki     <><
___. ___ ___ ___ _ _   __________________________________
[__ || __|__/|__||\/|  Information Retrieval, Semantic Web
___|||__||  \|  ||  |  Embedded Unix, System Integration
http://www.sigram.com  Contact: info at sigram dot com




-------------------------------------------------------
This SF.Net email is sponsored by xPML, a groundbreaking scripting language
that extends applications into web and mobile media. Attend the live webcast
and join the prime developer group breaking into this new coding territory!
http://sel.as-us.falkag.net/sel?cmd=lnk&kid=110944&bid=241720&dat=121642
_______________________________________________
Nutch-general mailing list
[email protected]
https://lists.sourceforge.net/lists/listinfo/nutch-general

Reply via email to