On Fri, Apr 11, 2008 at 9:32 AM, Rong-en Fan <[EMAIL PROTECTED]> wrote:
> On Fri, Apr 11, 2008 at 1:14 AM, stack <[EMAIL PROTECTED]> wrote:
>  > Rong-en Fan wrote:
>  >
>  > > I did so. I even rm -rf on dfs's dir and do namenode -format
>  > > before starting my dfs. hadoop fsck reports the default replication
>  > > is 1, avg. block replication is 2.9x after I wrote some data into
>  > > hbase. The underlying dfs is used by hbase. No other apps on
>  > > it.
>  > >
>  > >
>  >
>  >  What if you add a file using './bin/hadoop fs ....' -- i.e. don't have
>  > hbase in the mix at all -- does the file show as replicated?
>
>  It's 1 replication.
>
>
>  >  If you copy your hadoop-conf.xml to $HBASE_HOME/conf, does it then do the
>  > right thing?  Maybe whats happening is that hbase writing files, we're 
> using
>  > hadoop defaults.
>
>  Yes, I can verify by doing so, HBase respects my customized config. Shall I 
> file
>  a JIRA against HBase or Hadoop itself?

When HBase was in hadoop/contrib, the hbase script set both HADOOP_CONF_DIR
and HBASE_CONF_DIR to CLASSPATH, so that dfs's configuration can be loaded
correctly. However, when moved out hadoop/contrib, it only sets HBASE_CONF_DIR.

I can think of several possible solutions:

1) set HADOOP_CONF_DIR in hbase-env.sh, then add HADOOP_CONF_DIR to
    CLASSPATH as before
2) Instruct user to create links for hadoop-*.xml if they want to
customize some dfs settings.
3) If only a small set of dfs confs are related to dfs's client, maybe
they can be set via
    hbase-site.xml, then hbase sets these for us when create a FileSystem obj.

Regards,
Rong-En Fan

>
>
>  > > Hmm... as far as I understand the hadoop FileSystem, you can
>  > > specify # of replication when creating a file. But I did not find hbase
>  > > use it, correct?
>  > >
>  > >
>  >
>  >  We don't do it explicitly, but as I suggest above, we're probably using
>  > defaults instead of your custom config.
>  >
>  >  St.Ack
>
>  Thanks,
>  Rong-En Fan
>

Reply via email to