Where is the default replication factor on job files set? Is it different then the dfs.replication setting in hdfs-site.xml?
Sent from my iPad On May 18, 2011, at 9:10 PM, Joey Echeverria <j...@cloudera.com> wrote: > Did you run a map reduce job? > > I think the default replication factor on job files is 10, which > obviously doesn't work well on a psuedo-distributed cluster. > > -Joey > > On Wed, May 18, 2011 at 5:07 PM, Steve Cohen <mail4st...@gmail.com> wrote: >> Thanks for the answer. Earlier, I asked about why I get occasional not >> replicated yet errors. Now, I had dfs.replication set to one. What >> replication could it have been doing? Did the error messages actually mean >> that the file couldn't get created in the cluster? >> >> Thanks, >> Steve Cohen >> >> >> >> On May 18, 2011, at 6:39 PM, Todd Lipcon <t...@cloudera.com> wrote: >> >>> Tried to send this, but apparently SpamAssassin finds emails about >>> "replicas" to be spammy. This time with less rich text :) >>> >>> On Wed, May 18, 2011 at 3:35 PM, Todd Lipcon <t...@cloudera.com> wrote: >>>> >>>> Hi Steve, >>>> Running setrep will indeed change those files. Changing "dfs.replication" >>>> just changes the default replication value for files created in the >>>> future. Replication level is a file-specific property. >>>> Thanks >>>> -Todd >>>> >>>> On Wed, May 18, 2011 at 3:32 PM, Steve Cohen <mail4st...@gmail.com> wrote: >>>>> >>>>> Say I add a datanode to a pseudo cluster and I want to change the >>>>> replication factor to 2. I see that I can either run hadoop fs -setrep >>>>> or change the hdfs-site.xml value for dfs.replication. But do either >>>>> of these cause the existing blocks to replicate? >>>>> >>>>> Thanks, >>>>> Steve Cohen >>>> >>>> >>>> >>>> -- >>>> Todd Lipcon >>>> Software Engineer, Cloudera >>> >>> >>> >>> -- >>> Todd Lipcon >>> Software Engineer, Cloudera >> > > > > -- > Joseph Echeverria > Cloudera, Inc. > 443.305.9434