phoenix-user mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From Jesse Yates <jesse.k.ya...@gmail.com>
Subject Re: Error when try to create mutable secondary index...
Date Mon, 11 Aug 2014 17:08:44 GMT
Well now, that is strange. Maybe its something to do with CDH? Have you
talked to those fellas? Or maybe someone from Cloudera has an insight?

Seems like it should work
On Aug 11, 2014 9:55 AM, "Saravanan A" <asaravanan@alphaworkz.com> wrote:

> *bin/hbase classpath:*
>
>
> */opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../conf:/usr/java/default/lib/tools.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/..:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../hbase-0.94.15-cdh4.7.0-security.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../hbase-0.94.15-cdh4.7.0-security-tests.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../hbase.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../lib/activation-1.1.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../lib/aopalliance-1.0.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../lib/asm-3.2.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../lib/avro-1.7.4.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../lib/cloudera-jets3t-2.0.0-cdh4.7.0.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../lib/commons-beanutils-1.7.0.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../lib/commons-beanutils-core-1.8.0.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../lib/commons-cli-1.2.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../lib/commons-codec-1.4.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../lib/commons-collections-3.2.1.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../lib/commons-compress-1.4.1.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../lib/commons-configuration-1.6.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../lib/commons-daemon-1.0.3.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../lib/commons-digester-1.8.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../lib/commons-el-1.0.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../lib/commons-httpclient-3.1.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../lib/commons-io-2.1.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../lib/commons-lang-2.5.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../lib/commons-logging-1.1.1.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../lib/commons-net-3.1.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../lib/core-3.1.1.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../lib/gmbal-api-only-3.0.0-b023.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../lib/grizzly-framework-2.1.1.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../lib/grizzly-framework-2.1.1-tests.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../lib/grizzly-http-2.1.1.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../lib/grizzly-http-server-2.1.1.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../lib/grizzly-http-servlet-2.1.1.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../lib/grizzly-rcm-2.1.1.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../lib/guava-11.0.2.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../lib/guice-3.0.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../lib/guice-servlet-3.0.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../lib/hamcrest-core-1.3.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../lib/*
> *high-scale-lib-1.1.1.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../lib/httpclient-4.2.5.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../lib/httpcore-4.2.5.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../lib/jackson-core-asl-1.8.8.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../lib/jackson-jaxrs-1.8.8.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../lib/jackson-mapper-asl-1.8.8.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../lib/jackson-xc-1.8.8.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../lib/jamon-runtime-2.3.1.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../lib/jasper-compiler-5.5.23.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../lib/jasper-runtime-5.5.23.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../lib/javax.inject-1.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../lib/javax.servlet-3.0.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../lib/jaxb-api-2.1.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../lib/jaxb-impl-2.2.3-1.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../lib/jersey-client-1.8.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../lib/jersey-core-1.8.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../lib/jersey-grizzly2-1.8.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../lib/jersey-guice-1.8.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../lib/jersey-json-1.8.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../lib/jersey-server-1.8.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../lib/jersey-test-framework-core-1.8.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../lib/jersey-test-framework-grizzly2-1.8.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../lib/jets3t-0.6.1.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../lib/jettison-1.1.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../lib/jetty-6.1.26.cloudera.2.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../lib/jetty-util-6.1.26.cloudera.2.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../lib/jruby-complete-1.6.5.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../lib/jsch-0.1.42.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../lib/jsp-2.1-6.1.14.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../lib/jsp-api-2.1-6.1.14.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../lib/jsp-api-2.1.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../lib/jsr305-1.3.9.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../lib/junit-4.11.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../lib/kfs-0.3.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../lib/libthrift-0.9.0.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../lib/log4j-1.2.17.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../lib/management-api-3.0.0-b012.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../lib/metrics-core-2.1.2.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../lib/netty-3.2.4.Final.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../lib/netty-3.6.6.Final.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../lib/paranamer-2.3.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../lib/phoenix-core-3.0.0-incubating.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../lib/protobuf-java-2.4.0a.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../lib/servlet-api-2.5-6.1.14.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../lib/servlet-api-2.5.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../lib/slf4j-api-1.6.1.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../lib/snappy-java-1.0.4.1.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../lib/stax-api-1.0.1.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../lib/xmlenc-0.52.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../lib/xz-1.0.jar:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hbase/bin/../lib/zookeeper.jar:/etc/hadoop/conf/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/bin/../lib/hadoop/*:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/bin/../lib/hadoop/lib/*:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/bin/../lib/zookeeper/*:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/bin/../lib/zookeeper/lib/*::/etc/hadoop/conf:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hadoop/libexec/../../hadoop/lib/*:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hadoop/libexec/../../hadoop/.//*:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hadoop/libexec/../../hadoop-hdfs/./:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hadoop/libexec/../../hadoop-hdfs/lib/*:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hadoop/libexec/../../hadoop-hdfs/.//*:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hadoop/libexec/../../hadoop-yarn/lib/*:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hadoop/libexec/../../hadoop-yarn/.//*:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hadoop/libexec/../../hadoop-0.20-mapreduce/./:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hadoop/libexec/../../hadoop-0.20-mapreduce/lib/*:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/lib/hadoop/libexec/../../hadoop-0.20-mapreduce/.//*:/etc/hadoop/conf/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/bin/../lib/hadoop/*:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/bin/../lib/hadoop/lib/*:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/bin/../lib/zookeeper/*:/opt/cloudera/parcels/CDH-4.7.0-1.cdh4.7.0.p0.40/bin/../lib/zookeeper/lib/*:
> *
>
> this is the result i got for Hbase classpath command...and this is the
> "/opt/cloudera/parcels/CDH/lib/hbase/lib/" path i executed the code...
>
>
> On Mon, Aug 11, 2014 at 9:29 PM, Jesse Yates <jesse.k.yates@gmail.com>
> wrote:
>
>> The classpath you are using above doesn't include the HBase config files,
>> so the code executed will correctly tell you that the class exists, but is
>> not configured.
>>
>> Have you tried running
>> "bin/hbase classpath"
>> to see what you're classpath is at RS startup? If its the same as the -cp
>> argument, its missing the config files.
>> On Aug 11, 2014 6:10 AM, "Saravanan A" <asaravanan@alphaworkz.com> wrote:
>>
>>> *This is the command i run in hbase classpath (test1.jar is my jar file)*:
>>> hbase -cp
>>> .:hadoop-common-2.0.0-cdh4.7.0.jar:commons-logging-1.1.1.jar:hbase-0.94.15-cdh4.7.0-security.jar:com.google.collections.jar:commons-collections-3.2.1.jar:phoenix-core-3.0.0-incubating.jar:com.google.guava_1.6.0.jar:test1.jar
>>> FixConfigFile
>>>
>>> *The Output:*
>>> Found
>>> Not Found
>>>
>>> *This is my full code:*
>>>
>>> import org.apache.hadoop.conf.Configuration;
>>>
>>> public class FixConfigFile {
>>>
>>> public static final String INDEX_WAL_EDIT_CODEC_CLASS_NAME =
>>> "org.apache.hadoop.hbase.regionserver.wal.IndexedWALEditCodec";
>>>  public static final String WAL_EDIT_CODEC_CLASS_KEY =
>>> "org.apache.hadoop.hbase.regionserver.wal.codec";
>>>  public static void main(String[] args) {
>>> Configuration config=new Configuration();
>>> isWALEditCodecSet(config);
>>>
>>> }
>>> public static boolean isWALEditCodecSet(Configuration conf) {
>>>         // check to see if the WALEditCodec is installed
>>>         try {
>>>             // Use reflection to load the IndexedWALEditCodec, since it
>>> may not load with an older version
>>>             // of HBase
>>>             Class.forName(INDEX_WAL_EDIT_CODEC_CLASS_NAME);
>>>             System.out.println("Found");
>>>         } catch (Throwable t) {
>>>         System.out.println("Error");
>>>             return false;
>>>         }
>>>         if
>>> (INDEX_WAL_EDIT_CODEC_CLASS_NAME.equals(conf.get(WAL_EDIT_CODEC_CLASS_KEY,
>>> null))) {
>>>             // its installed, and it can handle compression and
>>> non-compression cases
>>>         System.out.println("True");
>>>             return true;
>>>         }
>>>         System.out.println("Not Found");
>>>         return false;
>>>     }
>>>
>>> }
>>> ************
>>>
>>> am not sure this is how you want me to execute the code...If am wrong
>>> please guide me...
>>>
>>>
>>>
>>> On Sat, Aug 9, 2014 at 8:32 PM, Jesse Yates <jesse.k.yates@gmail.com>
>>> wrote:
>>>
>>>> When you run
>>>>    $ bin/hbase classpath
>>>> What do you get? Should help illuminate if everything is setup right.
>>>>
>>>> If the phoenix jar is there, then check the contents of the jar (
>>>> http://docs.oracle.com/javase/tutorial/deployment/jar/view.html) and
>>>> make sure the classes are present.
>>>>  On Aug 9, 2014 1:03 AM, "Saravanan A" <asaravanan@alphaworkz.com>
>>>> wrote:
>>>>
>>>>> Hi Jesse,
>>>>>
>>>>> I ran the following code to test the existence of the classes you
>>>>> asked me to check. I initialized the two constants to the following values.
>>>>>
>>>>> =======
>>>>> public static final String INDEX_WAL_EDIT_CODEC_CLASS_NAME =
>>>>> "org.apache.hadoop.hbase.regionserver.wal.IndexedWALEditCodec";
>>>>>
>>>>> public static final String WAL_EDIT_CODEC_CLASS_KEY =
>>>>> "hbase.regionserver.wal.codec";
>>>>> ======
>>>>>
>>>>> Then I ran the following code and got the error "Not found" in the
>>>>> equality test.
>>>>>
>>>>> ====
>>>>>         if
>>>>> (INDEX_WAL_EDIT_CODEC_CLASS_NAME.equals(conf.get(WAL_EDIT_CODEC_CLASS_KEY,
>>>>> null))) {
>>>>>             // its installed, and it can handle compression and
>>>>> non-compression cases
>>>>>             System.out.println("True");
>>>>>             return true;
>>>>>         }
>>>>>         System.out.println("Not Found");
>>>>> ====
>>>>>
>>>>> I am not sure, if I initialized the values for the constants
>>>>> correctly. If I did, then I think some jars are missing or I have incorrect
>>>>> version.
>>>>> We use CDH 4.7 which has the Hbase version of 0.94.15 and Phoenix
>>>>> version of 3.0
>>>>>
>>>>> Can you tell me how to make this work? Your assistance is greatly
>>>>> appreciated.
>>>>>
>>>>> Regards,
>>>>> Saravanan.A
>>>>>
>>>>> Full code
>>>>> ==========
>>>>> public static void main(String[] args) {
>>>>>         Configuration config=new Configuration();
>>>>>         isWALEditCodecSet(config);
>>>>>
>>>>>     }
>>>>>     public static boolean isWALEditCodecSet(Configuration conf) {
>>>>>         // check to see if the WALEditCodec is installed
>>>>>         try {
>>>>>             // Use reflection to load the IndexedWALEditCodec, since
>>>>> it may not load with an older version
>>>>>             // of HBase
>>>>>             Class.forName(INDEX_WAL_EDIT_CODEC_CLASS_NAME);
>>>>>             System.out.println("Found");
>>>>>         } catch (Throwable t) {
>>>>>             System.out.println("Error");
>>>>>             return false;
>>>>>         }
>>>>>         if
>>>>> (INDEX_WAL_EDIT_CODEC_CLASS_NAME.equals(conf.get(WAL_EDIT_CODEC_CLASS_KEY,
>>>>> null))) {
>>>>>             // its installed, and it can handle compression and
>>>>> non-compression cases
>>>>>             System.out.println("True");
>>>>>             return true;
>>>>>         }
>>>>>         System.out.println("Not Found");
>>>>>         return false;
>>>>>     }
>>>>>
>>>>>
>>>>>
>>>>> On Sat, Aug 9, 2014 at 12:02 AM, Jesse Yates <jesse.k.yates@gmail.com>
>>>>> wrote:
>>>>>
>>>>>> This error is thrown when on the server-side, the following code
>>>>>> returns false (IndexManagementUtil#isWALEditCodecSet):
>>>>>>
>>>>>>     public static boolean isWALEditCodecSet(Configuration conf) {
>>>>>>>         // check to see if the WALEditCodec is installed
>>>>>>>         try {
>>>>>>>             // Use reflection to load the IndexedWALEditCodec,
since
>>>>>>> it may not load with an older version
>>>>>>>             // of HBase
>>>>>>>             Class.forName(INDEX_WAL_EDIT_CODEC_CLASS_NAME);
>>>>>>>         } catch (Throwable t) {
>>>>>>>             return false;
>>>>>>>         }
>>>>>>>         if
>>>>>>> (INDEX_WAL_EDIT_CODEC_CLASS_NAME.equals(conf.get(WAL_EDIT_CODEC_CLASS_KEY,
>>>>>>> null))) {
>>>>>>>             // its installed, and it can handle compression and
>>>>>>> non-compression cases
>>>>>>>             return true;
>>>>>>>         }
>>>>>>>         return false;
>>>>>>>     }
>>>>>>>
>>>>>>
>>>>>>  You could just put this into a main method in a java class, put
that
>>>>>> in the classpath of your HBase install on one of the machines on
your
>>>>>> cluster and run it from the HBase command line to make sure that
it passes.
>>>>>> Otherwise, you might not have the actual right configs (copy-paste
error?)
>>>>>> or missing the right jars.
>>>>>>
>>>>>>
>>>>>> Also, FWIW, this property:
>>>>>>
>>>>>>  <property>
>>>>>>>      <name>hbase.region.server.rpc.scheduler.factory.class</name>
>>>>>>>
>>>>>>>  <value>org.apache.phoenix.hbase.index.ipc.PhoenixIndexRpcSchedulerFactory</value>
>>>>>>>      <description>Factory to create the Phoenix RPC Scheduler
that
>>>>>>> knows to put index updates into index queues</description>
>>>>>>>
>>>>>> </property>
>>>>>>>
>>>>>>
>>>>>>  is only valid in HBase 0.98.4+ (as pointed out in the section
>>>>>> "Advanced Setup - Removing Index Deadlocks (0.98.4+)"). However,
it should
>>>>>> still be fine to have in older versions.
>>>>>>
>>>>>>
>>>>>>
>>>>>>
>>>>>> -------------------
>>>>>> Jesse Yates
>>>>>> @jesse_yates
>>>>>> jyates.github.com
>>>>>>
>>>>>>
>>>>>> On Fri, Aug 8, 2014 at 2:18 AM, Saravanan A <
>>>>>> asaravanan@alphaworkz.com> wrote:
>>>>>>
>>>>>>> This is my Hbase-site.xml file...
>>>>>>>
>>>>>>>
>>>>>>> <?xml version="1.0" encoding="UTF-8"?>
>>>>>>> <!--Autogenerated by Cloudera CM on 2014-06-16T11:10:16.319Z-->
>>>>>>> <configuration>
>>>>>>>
>>>>>>>  <property>
>>>>>>>      <name>hbase.regionserver.wal.codec</name>
>>>>>>>
>>>>>>>  <value>org.apache.hadoop.hbase.regionserver.wal.IndexedWALEditCodec</value>
>>>>>>>  </property>
>>>>>>>  <property>
>>>>>>>      <name>hbase.region.server.rpc.scheduler.factory.class</name>
>>>>>>>
>>>>>>>  <value>org.apache.phoenix.hbase.index.ipc.PhoenixIndexRpcSchedulerFactory</value>
>>>>>>>      <description>Factory to create the Phoenix RPC Scheduler
that
>>>>>>> knows to put index updates into index queues</description>
>>>>>>>  </property>
>>>>>>>
>>>>>>>   <property>
>>>>>>>     <name>hbase.rootdir</name>
>>>>>>>     <value>hdfs://alpmas.alp.com:8020/hbase</value>
>>>>>>>   </property>
>>>>>>>   <property>
>>>>>>>     <name>hbase.client.write.buffer</name>
>>>>>>>     <value>2097152</value>
>>>>>>>   </property>
>>>>>>>   <property>
>>>>>>>     <name>hbase.client.pause</name>
>>>>>>>     <value>1000</value>
>>>>>>>   </property>
>>>>>>>   <property>
>>>>>>>     <name>hbase.client.retries.number</name>
>>>>>>>     <value>10</value>
>>>>>>>   </property>
>>>>>>>   <property>
>>>>>>>     <name>hbase.client.scanner.caching</name>
>>>>>>>     <value>1000</value>
>>>>>>>   </property>
>>>>>>>   <property>
>>>>>>>     <name>hbase.client.keyvalue.maxsize</name>
>>>>>>>     <value>20971520</value>
>>>>>>>   </property>
>>>>>>>   <property>
>>>>>>>     <name>hbase.rpc.timeout</name>
>>>>>>>     <value>1200000</value>
>>>>>>>   </property>
>>>>>>>   <property>
>>>>>>>     <name>hbase.security.authentication</name>
>>>>>>>     <value>simple</value>
>>>>>>>   </property>
>>>>>>>   <property>
>>>>>>>     <name>zookeeper.session.timeout</name>
>>>>>>>     <value>240000</value>
>>>>>>>   </property>
>>>>>>>   <property>
>>>>>>>     <name>zookeeper.retries</name>
>>>>>>>     <value>5</value>
>>>>>>>   </property>
>>>>>>>   <property>
>>>>>>>     <name>zookeeper.pause</name>
>>>>>>>     <value>5000</value>
>>>>>>>   </property>
>>>>>>>   <property>
>>>>>>>     <name>zookeeper.znode.parent</name>
>>>>>>>     <value>/hbase</value>
>>>>>>>   </property>
>>>>>>>   <property>
>>>>>>>     <name>zookeeper.znode.rootserver</name>
>>>>>>>     <value>root-region-server</value>
>>>>>>>   </property>
>>>>>>>   <property>
>>>>>>>     <name>hbase.zookeeper.quorum</name>
>>>>>>>     <value>zk3.alp.com,zk2.alp.com,zk1.alp.com</value>
>>>>>>>   </property>
>>>>>>>   <property>
>>>>>>>     <name>hbase.zookeeper.property.clientPort</name>
>>>>>>>     <value>2181</value>
>>>>>>>   </property>
>>>>>>> </configuration>
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>> On Fri, Aug 8, 2014 at 2:46 PM, Saravanan A <
>>>>>>> asaravanan@alphaworkz.com> wrote:
>>>>>>>
>>>>>>>> I already included this property in hbase-site.xml in all
region
>>>>>>>> servers..but still am getting that error...If i define my
view as
>>>>>>>> IMMUTABLE_ROWS = true, then i can able to create view..but
i want to create
>>>>>>>> index for mutable..
>>>>>>>>
>>>>>>>>
>>>>>>>> On Fri, Aug 8, 2014 at 2:10 PM, Abhilash L L <
>>>>>>>> abhilash@capillarytech.com> wrote:
>>>>>>>>
>>>>>>>>> Really sorry, shared the wrong config
>>>>>>>>>
>>>>>>>>>
>>>>>>>>>
>>>>>>>>>
>>>>>>>>>
>>>>>>>>>
>>>>>>>>> <property>
>>>>>>>>>   <name>hbase.regionserver.wal.codec</name>
>>>>>>>>>   <value>org.apache.hadoop.hbase.regionserver.wal.IndexedWALEditCodec</value>
>>>>>>>>> </property>
>>>>>>>>>
>>>>>>>>>
>>>>>>>>>
>>>>>>>>> Regards,
>>>>>>>>> Abhilash L L
>>>>>>>>> Capillary Technologies
>>>>>>>>> M:919886208262
>>>>>>>>> abhilash@capillarytech.com | www.capillarytech.com
>>>>>>>>>
>>>>>>>>> Email from people at capillarytech.com may not represent
official
>>>>>>>>> policy of  Capillary Technologies unless explicitly stated.
Please see our
>>>>>>>>> Corporate-Email-Policy
>>>>>>>>> <http://support.capillary.co.in/policy-public/Corporate-Email-Policy.pdf>
>>>>>>>>> for details. Contents of this email are confidential.
Please contact the
>>>>>>>>> Sender if you have received this email in error.
>>>>>>>>>
>>>>>>>>>
>>>>>>>>>
>>>>>>>>> On Fri, Aug 8, 2014 at 1:07 PM, Saravanan A <
>>>>>>>>> asaravanan@alphaworkz.com> wrote:
>>>>>>>>>
>>>>>>>>>> Hi Abhilash,
>>>>>>>>>>
>>>>>>>>>> Thanks for the replay...i included above property
and restarted
>>>>>>>>>> the region servers but still am getting the same
error...
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> On Fri, Aug 8, 2014 at 12:39 PM, Abhilash L L <
>>>>>>>>>> abhilash@capillarytech.com> wrote:
>>>>>>>>>>
>>>>>>>>>>> Hi Saravanan,
>>>>>>>>>>>
>>>>>>>>>>>     Please check the Setup section here
>>>>>>>>>>>
>>>>>>>>>>> http://phoenix.apache.org/secondary_indexing.html
>>>>>>>>>>>
>>>>>>>>>>>    You will need to add this config to all Region
Servers in
>>>>>>>>>>> hbase-site. xml, as the error says as well (You
will need to restart the
>>>>>>>>>>> servers after the change)
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>> <property>
>>>>>>>>>>>   <name>hbase.region.server.rpc.scheduler.factory.class</name>
>>>>>>>>>>>   <value>org.apache.phoenix.hbase.index.ipc.PhoenixIndexRpcSchedulerFactory</value>
>>>>>>>>>>>   <description>Factory to create the Phoenix
RPC Scheduler that knows to put index updates into index queues</description>
>>>>>>>>>>> </property>
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>> Regards,
>>>>>>>>>>> Abhilash L L
>>>>>>>>>>> Capillary Technologies
>>>>>>>>>>> M:919886208262
>>>>>>>>>>> abhilash@capillarytech.com | www.capillarytech.com
>>>>>>>>>>>
>>>>>>>>>>> Email from people at capillarytech.com may not
represent
>>>>>>>>>>> official policy of  Capillary Technologies unless
explicitly stated. Please
>>>>>>>>>>> see our Corporate-Email-Policy
>>>>>>>>>>> <http://support.capillary.co.in/policy-public/Corporate-Email-Policy.pdf>
>>>>>>>>>>> for details. Contents of this email are confidential.
Please contact the
>>>>>>>>>>> Sender if you have received this email in error.
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>> On Fri, Aug 8, 2014 at 12:22 PM, Saravanan A
<
>>>>>>>>>>> asaravanan@alphaworkz.com> wrote:
>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>> Hi,
>>>>>>>>>>>>
>>>>>>>>>>>>     I have a table in hbase and created view
in phoenix and try
>>>>>>>>>>>> to create index on a column on the view..but
i got following error..
>>>>>>>>>>>>
>>>>>>>>>>>> Error: ERROR 1029 (42Y88): Mutable secondary
indexes must have
>>>>>>>>>>>> the hbase.regionserver.wal.codec property
set to
>>>>>>>>>>>> org.apache.hadoop.hbase.regionserver.wal.IndexedWALEditCodec
in the
>>>>>>>>>>>> hbase-sites.xml of every region server tableName=tab2_col4
>>>>>>>>>>>> (state=42Y88,code=1029)
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>> but i have added the hbase.regionserver.wal.codec
property in
>>>>>>>>>>>> all my region server...i can able to create
IMMUTABLE index for that...
>>>>>>>>>>>>
>>>>>>>>>>>> Am using Hbase ---0.94.15-cdh4.7.0
>>>>>>>>>>>>             Phoenix---3.0
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>> am i missing something???
>>>>>>>>>>>> thanks in advance...
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>> Regards,
>>>>>>>>>>>> Saravanan
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>> Email from people at capillarytech.com may not
represent
>>>>>>>>>>> official policy of Capillary Technologies unless
explicitly stated. Please
>>>>>>>>>>> see our Corporate-Email-Policy for details.Contents
of this email are
>>>>>>>>>>> confidential. Please contact the Sender if you
have received this email in
>>>>>>>>>>> error.
>>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>
>>>>>>>>> Email from people at capillarytech.com may not represent
official
>>>>>>>>> policy of Capillary Technologies unless explicitly stated.
Please see our
>>>>>>>>> Corporate-Email-Policy for details.Contents of this email
are confidential.
>>>>>>>>> Please contact the Sender if you have received this email
in error.
>>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>
>>>>>>
>>>>>
>>>
>

Mime
View raw message