Bryan, could you describe in more details how your files "rot"?
Do you loose entire files or blocks? How do you detect missing files?
How regular is "regular basis"?
Should we have a jira issue for that?

I think Jagadeesh' q#3 was about running a spare name node on the same cluster,
rather than running overlapping clusters.

On the logging issue. I think we should change the default logging level,
which is INFO at the moment.

Thanks,
Konstantin


Bryan A. Pendleton wrote:

I'm not sure I agree with Konstantin's assurances on reliability, just yet. I've been running a persistent cluster of hadoop since early this year, and have repeatedly had data "rot". Things have gotten better lately, but it's not clear that all of the necessary tests and ruggedness testing have been
done to stick this into a live production environment, especially without
additional backups in other storage systems. Still, a small portion of my
files "rot" on a regular basis.

To 1: Yes, with a replication of 3, any 2 nodes can fail with no loss of
data. If the failures happen slowly enough, more failures are tolerated,
supposing that there's time and space for the re-replications to take place.
I once had 5 of my ~30 nodes fail across a weekend, with no data becoming
unavailable for ongoing processes. There are currently no active tests that
blocks which are reported present can actually be read from disk, though,
unless attempts to read the data are actually made.

To 2: Not at this time. However, if you temporarily turn replication *up*
for a large subset of your data, wait for things to become quiescent, then turn replication back - you will effectively redistribute your free space. Be aware that things might be "wonky" during heavy up-replications. I've had map jobs fail because the replication load made normal block accesses take
long enough to start losing tasks.

To 3: Using the current state of the code, you could run overlapping DFS
clusters. For each namenode you run, you'd want to use different storage
directories on the cluster nodes.

I don't know if you're likely to run into namenode performance problems just yet, though. It is worth noting that, for a very busy cluster, you'll want
to periodically restart the namenode, to incorporate the change long into
the disk-stored image. Otherwise, the edits log might grow forever. Probably
take a lot longer than running out of space from your over-full logs,
though.

On 8/10/06, Jagadeesh <[EMAIL PROTECTED]> wrote:


Dear Konstantin Shvachko,

Thanks for your reply.

I have anyway decided to try Hadoop for our application and I successfully
connected 22 nodes and it's working fine so far. I had one issue though
where the master node generated 5 log files of size 125GB each in 5 days and
resulted in crashing the server. Anyway I have changed the properties in
log4j and fixed it.

I have a few more queries and really appreciate if you can give me an
answer to those.

1. If I set the replication level to 3 or 4, will I be able to access the
files even if one or two slave nodes go down.

2. If one of the slave nodes run out of disk space, will hadoop perform
any defragmentation process by moving some data blocks onto other nodes?

3. Can I run multiple master nodes with the same set of slaves and then by
modifying the code have the master nodes communicating to each other
informing the data chunks stored in slave nodes? By this way we can run
multiple master nodes and provide options for load balancing and clustering. You would be the right person to suggest an approach and I can work on it.

Please let me know your thougts...

Thanks and Best
Jugs

-----Original Message-----
    From: "Konstantin Shvachko"<[EMAIL PROTECTED]>
    Sent: 8/10/06 11:27:13 PM
    To: "[email protected]"<[email protected]>
    Subject: Re: Some queries about stability and reliability

    Hi Jagadeesh,

    >I am very much new to Hadoop and would like to know some details
about the
    >reliability and stability. I am developing flickr kind of an
application for
>storing and sharing movies and would like to use Hadoop as my storage >backend. I am planning to put in atleast 100 nodes and would like to
know
>more about the product. I will appreciate if you could answer some of
my
    >queries.
    >
    >
    This is a very interesting application for Hadoop.
    Did you have any progress with the system?

    >1. Is the product matured enough for using in an application like
this?
    >
    >
    Yes.

    >2. Has somebody tested it using atleast 100 nodes?
    >
    >
    Yes, there are even larger installations.

    >3. Can I have multiple master nodes in Hadoop to do load balancing
and
    >fail-overs?
    >
    >
    Not yet.

>4. What is the maximum number of simultaneous connections possible in
    >Hadoop?
    >
    >
    Hadoop is designed to support and actually supports high volume of
    simultaneous connections.
    E.g., on a 100 node cluster an extensive map-reduce job can generate
400
    concurrent connections.

    Creation time and date is not implemented for DFS files.
    Do you have a good application for ctime?

    Thank you,

    --Konstantin





Reply via email to