*bin/hbase classpath:* */opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../conf:/usr/java/default/lib/tools.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/..:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../hbase-0.94.15-cdh4.7.0-security.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../hbase-0.94.15-cdh4.7.0-security-tests.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../hbase.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../lib/activation-1.1.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../lib/aopalliance-1.0.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../lib/asm-3.2.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../lib/avro-1.7.4.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../lib/cloudera-jets3t-2.0.0-cdh4.7.0.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../lib/commons-beanutils-1.7.0.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../lib/commons-beanutils-core-1.8.0.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../lib/commons-cli-1.2.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../lib/commons-codec-1.4.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../lib/commons-collections-3.2.1.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../lib/commons-compress-1.4.1.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../lib/commons-configuration-1.6.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../lib/commons-daemon-1.0.3.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../lib/commons-digester-1.8.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../lib/commons-el-1.0.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../lib/commons-httpclient-3.1.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../lib/commons-io-2.1.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../lib/commons-lang-2.5.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../lib/commons-logging-1.1.1.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../lib/commons-net-3.1.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../lib/core-3.1.1.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../lib/gmbal-api-only-3.0.0-b023.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../lib/grizzly-framework-2.1.1.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../lib/grizzly-framework-2.1.1-tests.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../lib/grizzly-http-2.1.1.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../lib/grizzly-http-server-2.1.1.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../lib/grizzly-http-servlet-2.1.1.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../lib/grizzly-rcm-2.1.1.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../lib/guava-11.0.2.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../lib/guice-3.0.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../lib/guice-servlet-3.0.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../lib/hamcrest-core-1.3.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../lib/* *high-scale-lib-1.1.1.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../lib/httpclient-4.2.5.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../lib/httpcore-4.2.5.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../lib/jackson-core-asl-1.8.8.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../lib/jackson-jaxrs-1.8.8.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../lib/jackson-mapper-asl-1.8.8.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../lib/jackson-xc-1.8.8.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../lib/jamon-runtime-2.3.1.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../lib/jasper-compiler-5.5.23.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../lib/jasper-runtime-5.5.23.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../lib/javax.inject-1.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../lib/javax.servlet-3.0.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../lib/jaxb-api-2.1.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../lib/jaxb-impl-2.2.3-1.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../lib/jersey-client-1.8.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../lib/jersey-core-1.8.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../lib/jersey-grizzly2-1.8.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../lib/jersey-guice-1.8.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../lib/jersey-json-1.8.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../lib/jersey-server-1.8.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../lib/jersey-test-framework-core-1.8.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../lib/jersey-test-framework-grizzly2-1.8.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../lib/jets3t-0.6.1.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../lib/jettison-1.1.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../lib/jetty-6.1.26.cloudera.2.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../lib/jetty-util-6.1.26.cloudera.2.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../lib/jruby-complete-1.6.5.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../lib/jsch-0.1.42.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../lib/jsp-2.1-6.1.14.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../lib/jsp-api-2.1-6.1.14.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../lib/jsp-api-2.1.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../lib/jsr305-1.3.9.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../lib/junit-4.11.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../lib/kfs-0.3.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../lib/libthrift-0.9.0.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../lib/log4j-1.2.17.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../lib/management-api-3.0.0-b012.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../lib/metrics-core-2.1.2.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../lib/netty-3.2.4.Final.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../lib/netty-3.6.6.Final.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../lib/paranamer-2.3.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../lib/phoenix-core-3.0.0-incubating.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../lib/protobuf-java-2.4.0a.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../lib/servlet-api-2.5-6.1.14.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../lib/servlet-api-2.5.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../lib/slf4j-api-1.6.1.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../lib/snappy-java-1.0.4.1.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../lib/stax-api-1.0.1.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../lib/xmlenc-0.52.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../lib/xz-1.0.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../lib/zookeeper.jar:/etc/hadoop/conf/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/bin/../lib/hadoop/*:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/bin/../lib/hadoop/lib/*:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/bin/../lib/zookeeper/*:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/bin/../lib/zookeeper/lib/*::/etc/hadoop/conf:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hadoop/libexec/../../hadoop/lib/*:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hadoop/libexec/../../hadoop/.//*:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hadoop/libexec/../../hadoop-hdfs/./:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hadoop/libexec/../../hadoop-hdfs/lib/*:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hadoop/libexec/../../hadoop-hdfs/.//*:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hadoop/libexec/../../hadoop-yarn/lib/*:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hadoop/libexec/../../hadoop-yarn/.//*:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hadoop/libexec/../../hadoop-0.20-mapreduce/./:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hadoop/libexec/../../hadoop-0.20-mapreduce/lib/*:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hadoop/libexec/../../hadoop-0.20-mapreduce/.//*:/etc/hadoop/conf/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/bin/../lib/hadoop/*:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/bin/../lib/hadoop/lib/*:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/bin/../lib/zookeeper/*:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/bin/../lib/zookeeper/lib/*:*
this is the result i got for Hbase classpath command...and this is the "/opt/cloudera/parcels/CDH/lib/hbase/lib/" path i executed the code... On Mon, Aug 11, 2014 at 9:29 PM, Jesse Yates <jesse.k.ya...@gmail.com> wrote: > The classpath you are using above doesn't include the HBase config files, > so the code executed will correctly tell you that the class exists, but is > not configured. > > Have you tried running > "bin/hbase classpath" > to see what you're classpath is at RS startup? If its the same as the -cp > argument, its missing the config files. > On Aug 11, 2014 6:10 AM, "Saravanan A" <asarava...@alphaworkz.com> wrote: > >> *This is the command i run in hbase classpath (test1.jar is my jar file)*: >> hbase -cp >> .:hadoop-common-2.0.0-cdh4.7.0.jar:commons-logging-1.1.1.jar:hbase-0.94.15-cdh4.7.0-security.jar:com.google.collections.jar:commons-collections-3.2.1.jar:phoenix-core-3.0.0-incubating.jar:com.google.guava_1.6.0.jar:test1.jar >> FixConfigFile >> >> *The Output:* >> Found >> Not Found >> >> *This is my full code:* >> >> import org.apache.hadoop.conf.Configuration; >> >> public class FixConfigFile { >> >> public static final String INDEX_WAL_EDIT_CODEC_CLASS_NAME = >> "org.apache.hadoop.hbase.regionserver.wal.IndexedWALEditCodec"; >> public static final String WAL_EDIT_CODEC_CLASS_KEY = >> "org.apache.hadoop.hbase.regionserver.wal.codec"; >> public static void main(String[] args) { >> Configuration config=new Configuration(); >> isWALEditCodecSet(config); >> >> } >> public static boolean isWALEditCodecSet(Configuration conf) { >> // check to see if the WALEditCodec is installed >> try { >> // Use reflection to load the IndexedWALEditCodec, since it >> may not load with an older version >> // of HBase >> Class.forName(INDEX_WAL_EDIT_CODEC_CLASS_NAME); >> System.out.println("Found"); >> } catch (Throwable t) { >> System.out.println("Error"); >> return false; >> } >> if >> (INDEX_WAL_EDIT_CODEC_CLASS_NAME.equals(conf.get(WAL_EDIT_CODEC_CLASS_KEY, >> null))) { >> // its installed, and it can handle compression and >> non-compression cases >> System.out.println("True"); >> return true; >> } >> System.out.println("Not Found"); >> return false; >> } >> >> } >> ************ >> >> am not sure this is how you want me to execute the code...If am wrong >> please guide me... >> >> >> >> On Sat, Aug 9, 2014 at 8:32 PM, Jesse Yates <jesse.k.ya...@gmail.com> >> wrote: >> >>> When you run >>> $ bin/hbase classpath >>> What do you get? Should help illuminate if everything is setup right. >>> >>> If the phoenix jar is there, then check the contents of the jar ( >>> http://docs.oracle.com/javase/tutorial/deployment/jar/view.html) and >>> make sure the classes are present. >>> On Aug 9, 2014 1:03 AM, "Saravanan A" <asarava...@alphaworkz.com> >>> wrote: >>> >>>> Hi Jesse, >>>> >>>> I ran the following code to test the existence of the classes you asked >>>> me to check. I initialized the two constants to the following values. >>>> >>>> ======= >>>> public static final String INDEX_WAL_EDIT_CODEC_CLASS_NAME = >>>> "org.apache.hadoop.hbase.regionserver.wal.IndexedWALEditCodec"; >>>> >>>> public static final String WAL_EDIT_CODEC_CLASS_KEY = >>>> "hbase.regionserver.wal.codec"; >>>> ====== >>>> >>>> Then I ran the following code and got the error "Not found" in the >>>> equality test. >>>> >>>> ==== >>>> if >>>> (INDEX_WAL_EDIT_CODEC_CLASS_NAME.equals(conf.get(WAL_EDIT_CODEC_CLASS_KEY, >>>> null))) { >>>> // its installed, and it can handle compression and >>>> non-compression cases >>>> System.out.println("True"); >>>> return true; >>>> } >>>> System.out.println("Not Found"); >>>> ==== >>>> >>>> I am not sure, if I initialized the values for the constants correctly. >>>> If I did, then I think some jars are missing or I have incorrect version. >>>> We use CDH 4.7 which has the Hbase version of 0.94.15 and Phoenix >>>> version of 3.0 >>>> >>>> Can you tell me how to make this work? Your assistance is greatly >>>> appreciated. >>>> >>>> Regards, >>>> Saravanan.A >>>> >>>> Full code >>>> ========== >>>> public static void main(String[] args) { >>>> Configuration config=new Configuration(); >>>> isWALEditCodecSet(config); >>>> >>>> } >>>> public static boolean isWALEditCodecSet(Configuration conf) { >>>> // check to see if the WALEditCodec is installed >>>> try { >>>> // Use reflection to load the IndexedWALEditCodec, since it >>>> may not load with an older version >>>> // of HBase >>>> Class.forName(INDEX_WAL_EDIT_CODEC_CLASS_NAME); >>>> System.out.println("Found"); >>>> } catch (Throwable t) { >>>> System.out.println("Error"); >>>> return false; >>>> } >>>> if >>>> (INDEX_WAL_EDIT_CODEC_CLASS_NAME.equals(conf.get(WAL_EDIT_CODEC_CLASS_KEY, >>>> null))) { >>>> // its installed, and it can handle compression and >>>> non-compression cases >>>> System.out.println("True"); >>>> return true; >>>> } >>>> System.out.println("Not Found"); >>>> return false; >>>> } >>>> >>>> >>>> >>>> On Sat, Aug 9, 2014 at 12:02 AM, Jesse Yates <jesse.k.ya...@gmail.com> >>>> wrote: >>>> >>>>> This error is thrown when on the server-side, the following code >>>>> returns false (IndexManagementUtil#isWALEditCodecSet): >>>>> >>>>> public static boolean isWALEditCodecSet(Configuration conf) { >>>>>> // check to see if the WALEditCodec is installed >>>>>> try { >>>>>> // Use reflection to load the IndexedWALEditCodec, since >>>>>> it may not load with an older version >>>>>> // of HBase >>>>>> Class.forName(INDEX_WAL_EDIT_CODEC_CLASS_NAME); >>>>>> } catch (Throwable t) { >>>>>> return false; >>>>>> } >>>>>> if >>>>>> (INDEX_WAL_EDIT_CODEC_CLASS_NAME.equals(conf.get(WAL_EDIT_CODEC_CLASS_KEY, >>>>>> null))) { >>>>>> // its installed, and it can handle compression and >>>>>> non-compression cases >>>>>> return true; >>>>>> } >>>>>> return false; >>>>>> } >>>>>> >>>>> >>>>> You could just put this into a main method in a java class, put that >>>>> in the classpath of your HBase install on one of the machines on your >>>>> cluster and run it from the HBase command line to make sure that it >>>>> passes. >>>>> Otherwise, you might not have the actual right configs (copy-paste error?) >>>>> or missing the right jars. >>>>> >>>>> >>>>> Also, FWIW, this property: >>>>> >>>>> <property> >>>>>> <name>hbase.region.server.rpc.scheduler.factory.class</name> >>>>>> >>>>>> >>>>>> <value>org.apache.phoenix.hbase.index.ipc.PhoenixIndexRpcSchedulerFactory</value> >>>>>> <description>Factory to create the Phoenix RPC Scheduler that >>>>>> knows to put index updates into index queues</description> >>>>>> >>>>> </property> >>>>>> >>>>> >>>>> is only valid in HBase 0.98.4+ (as pointed out in the section >>>>> "Advanced Setup - Removing Index Deadlocks (0.98.4+)"). However, it should >>>>> still be fine to have in older versions. >>>>> >>>>> >>>>> >>>>> >>>>> ------------------- >>>>> Jesse Yates >>>>> @jesse_yates >>>>> jyates.github.com >>>>> >>>>> >>>>> On Fri, Aug 8, 2014 at 2:18 AM, Saravanan A <asarava...@alphaworkz.com >>>>> > wrote: >>>>> >>>>>> This is my Hbase-site.xml file... >>>>>> >>>>>> >>>>>> <?xml version="1.0" encoding="UTF-8"?> >>>>>> <!--Autogenerated by Cloudera CM on 2014-06-16T11:10:16.319Z--> >>>>>> <configuration> >>>>>> >>>>>> <property> >>>>>> <name>hbase.regionserver.wal.codec</name> >>>>>> >>>>>> >>>>>> <value>org.apache.hadoop.hbase.regionserver.wal.IndexedWALEditCodec</value> >>>>>> </property> >>>>>> <property> >>>>>> <name>hbase.region.server.rpc.scheduler.factory.class</name> >>>>>> >>>>>> >>>>>> <value>org.apache.phoenix.hbase.index.ipc.PhoenixIndexRpcSchedulerFactory</value> >>>>>> <description>Factory to create the Phoenix RPC Scheduler that >>>>>> knows to put index updates into index queues</description> >>>>>> </property> >>>>>> >>>>>> <property> >>>>>> <name>hbase.rootdir</name> >>>>>> <value>hdfs://alpmas.alp.com:8020/hbase</value> >>>>>> </property> >>>>>> <property> >>>>>> <name>hbase.client.write.buffer</name> >>>>>> <value>2097152</value> >>>>>> </property> >>>>>> <property> >>>>>> <name>hbase.client.pause</name> >>>>>> <value>1000</value> >>>>>> </property> >>>>>> <property> >>>>>> <name>hbase.client.retries.number</name> >>>>>> <value>10</value> >>>>>> </property> >>>>>> <property> >>>>>> <name>hbase.client.scanner.caching</name> >>>>>> <value>1000</value> >>>>>> </property> >>>>>> <property> >>>>>> <name>hbase.client.keyvalue.maxsize</name> >>>>>> <value>20971520</value> >>>>>> </property> >>>>>> <property> >>>>>> <name>hbase.rpc.timeout</name> >>>>>> <value>1200000</value> >>>>>> </property> >>>>>> <property> >>>>>> <name>hbase.security.authentication</name> >>>>>> <value>simple</value> >>>>>> </property> >>>>>> <property> >>>>>> <name>zookeeper.session.timeout</name> >>>>>> <value>240000</value> >>>>>> </property> >>>>>> <property> >>>>>> <name>zookeeper.retries</name> >>>>>> <value>5</value> >>>>>> </property> >>>>>> <property> >>>>>> <name>zookeeper.pause</name> >>>>>> <value>5000</value> >>>>>> </property> >>>>>> <property> >>>>>> <name>zookeeper.znode.parent</name> >>>>>> <value>/hbase</value> >>>>>> </property> >>>>>> <property> >>>>>> <name>zookeeper.znode.rootserver</name> >>>>>> <value>root-region-server</value> >>>>>> </property> >>>>>> <property> >>>>>> <name>hbase.zookeeper.quorum</name> >>>>>> <value>zk3.alp.com,zk2.alp.com,zk1.alp.com</value> >>>>>> </property> >>>>>> <property> >>>>>> <name>hbase.zookeeper.property.clientPort</name> >>>>>> <value>2181</value> >>>>>> </property> >>>>>> </configuration> >>>>>> >>>>>> >>>>>> >>>>>> On Fri, Aug 8, 2014 at 2:46 PM, Saravanan A < >>>>>> asarava...@alphaworkz.com> wrote: >>>>>> >>>>>>> I already included this property in hbase-site.xml in all region >>>>>>> servers..but still am getting that error...If i define my view as >>>>>>> IMMUTABLE_ROWS = true, then i can able to create view..but i want to >>>>>>> create >>>>>>> index for mutable.. >>>>>>> >>>>>>> >>>>>>> On Fri, Aug 8, 2014 at 2:10 PM, Abhilash L L < >>>>>>> abhil...@capillarytech.com> wrote: >>>>>>> >>>>>>>> Really sorry, shared the wrong config >>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> <property> >>>>>>>> <name>hbase.regionserver.wal.codec</name> >>>>>>>> >>>>>>>> <value>org.apache.hadoop.hbase.regionserver.wal.IndexedWALEditCodec</value> >>>>>>>> </property> >>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> Regards, >>>>>>>> Abhilash L L >>>>>>>> Capillary Technologies >>>>>>>> M:919886208262 >>>>>>>> abhil...@capillarytech.com | www.capillarytech.com >>>>>>>> >>>>>>>> Email from people at capillarytech.com may not represent official >>>>>>>> policy of Capillary Technologies unless explicitly stated. Please see >>>>>>>> our >>>>>>>> Corporate-Email-Policy >>>>>>>> <http://support.capillary.co.in/policy-public/Corporate-Email-Policy.pdf> >>>>>>>> for details. Contents of this email are confidential. Please contact >>>>>>>> the >>>>>>>> Sender if you have received this email in error. >>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> On Fri, Aug 8, 2014 at 1:07 PM, Saravanan A < >>>>>>>> asarava...@alphaworkz.com> wrote: >>>>>>>> >>>>>>>>> Hi Abhilash, >>>>>>>>> >>>>>>>>> Thanks for the replay...i included above property and restarted >>>>>>>>> the region servers but still am getting the same error... >>>>>>>>> >>>>>>>>> >>>>>>>>> >>>>>>>>> On Fri, Aug 8, 2014 at 12:39 PM, Abhilash L L < >>>>>>>>> abhil...@capillarytech.com> wrote: >>>>>>>>> >>>>>>>>>> Hi Saravanan, >>>>>>>>>> >>>>>>>>>> Please check the Setup section here >>>>>>>>>> >>>>>>>>>> http://phoenix.apache.org/secondary_indexing.html >>>>>>>>>> >>>>>>>>>> You will need to add this config to all Region Servers in >>>>>>>>>> hbase-site. xml, as the error says as well (You will need to restart >>>>>>>>>> the >>>>>>>>>> servers after the change) >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> <property> >>>>>>>>>> <name>hbase.region.server.rpc.scheduler.factory.class</name> >>>>>>>>>> >>>>>>>>>> <value>org.apache.phoenix.hbase.index.ipc.PhoenixIndexRpcSchedulerFactory</value> >>>>>>>>>> <description>Factory to create the Phoenix RPC Scheduler that >>>>>>>>>> knows to put index updates into index queues</description> >>>>>>>>>> </property> >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> Regards, >>>>>>>>>> Abhilash L L >>>>>>>>>> Capillary Technologies >>>>>>>>>> M:919886208262 >>>>>>>>>> abhil...@capillarytech.com | www.capillarytech.com >>>>>>>>>> >>>>>>>>>> Email from people at capillarytech.com may not represent >>>>>>>>>> official policy of Capillary Technologies unless explicitly stated. >>>>>>>>>> Please >>>>>>>>>> see our Corporate-Email-Policy >>>>>>>>>> <http://support.capillary.co.in/policy-public/Corporate-Email-Policy.pdf> >>>>>>>>>> for details. Contents of this email are confidential. Please contact >>>>>>>>>> the >>>>>>>>>> Sender if you have received this email in error. >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> On Fri, Aug 8, 2014 at 12:22 PM, Saravanan A < >>>>>>>>>> asarava...@alphaworkz.com> wrote: >>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> Hi, >>>>>>>>>>> >>>>>>>>>>> I have a table in hbase and created view in phoenix and try >>>>>>>>>>> to create index on a column on the view..but i got following error.. >>>>>>>>>>> >>>>>>>>>>> Error: ERROR 1029 (42Y88): Mutable secondary indexes must have >>>>>>>>>>> the hbase.regionserver.wal.codec property set to >>>>>>>>>>> org.apache.hadoop.hbase.regionserver.wal.IndexedWALEditCodec in the >>>>>>>>>>> hbase-sites.xml of every region server tableName=tab2_col4 >>>>>>>>>>> (state=42Y88,code=1029) >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> but i have added the hbase.regionserver.wal.codec property in >>>>>>>>>>> all my region server...i can able to create IMMUTABLE index for >>>>>>>>>>> that... >>>>>>>>>>> >>>>>>>>>>> Am using Hbase ---0.94.15-cdh4.7.0 >>>>>>>>>>> Phoenix---3.0 >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> am i missing something??? >>>>>>>>>>> thanks in advance... >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> Regards, >>>>>>>>>>> Saravanan >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>> >>>>>>>>>> Email from people at capillarytech.com may not represent >>>>>>>>>> official policy of Capillary Technologies unless explicitly stated. >>>>>>>>>> Please >>>>>>>>>> see our Corporate-Email-Policy for details.Contents of this email are >>>>>>>>>> confidential. Please contact the Sender if you have received this >>>>>>>>>> email in >>>>>>>>>> error. >>>>>>>>>> >>>>>>>>> >>>>>>>>> >>>>>>>> >>>>>>>> Email from people at capillarytech.com may not represent official >>>>>>>> policy of Capillary Technologies unless explicitly stated. Please see >>>>>>>> our >>>>>>>> Corporate-Email-Policy for details.Contents of this email are >>>>>>>> confidential. >>>>>>>> Please contact the Sender if you have received this email in error. >>>>>>>> >>>>>>> >>>>>>> >>>>>> >>>>> >>>> >>