phoenix-user mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From Francis Chuang <francischu...@apache.org>
Subject Re: org.apache.phoenix.shaded.org.apache.thrift.TException: Unable to discover transaction service. -> TException: Unable to discover transaction service.
Date Fri, 28 Sep 2018 05:39:28 GMT
I updated the docker image file to remove the hadoop jars bundled with 
HBase and replace them with the ones from hadoop 2.8.5 from maven. Maven 
does not host any of the test jars which were present in the HBase 
distribution, but I did not notice any adverse effects.

The image works correctly on all my machines after this update, however, 
I am still baffled as to why it the one using Hadoop 2.7.4 jars worked 
correctly on some machines but failed on others.

On 28/09/2018 8:24 AM, Francis Chuang wrote:
> I tried updating my hbase-phoenix-all-in-one image to use HBase built 
> with Hadoop 3. Unfortunately, it didn't work. I think this might be 
> because Hadoop 3.0.0 is too new for tephra (which uses 2.2.0):
>
> starting Query Server, logging to 
> /tmp/phoenix/phoenix-root-queryserver.log
> Thu Sep 27 02:50:56 UTC 2018 Starting tephra service on 
> m401b01-phoenix.m401b01
> Running class org.apache.tephra.TransactionServiceMain
> SLF4J: Class path contains multiple SLF4J bindings.
> SLF4J: Found binding in 
> [jar:file:/opt/hbase/lib/phoenix-5.0.0-HBase-2.0-client.jar!/org/slf4j/impl/StaticLoggerBinder.class]
> SLF4J: Found binding in 
> [jar:file:/opt/hbase/lib/slf4j-log4j12-1.7.25.jar!/org/slf4j/impl/StaticLoggerBinder.class]
> SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an 
> explanation.
> Exception in thread "main" java.lang.NoSuchMethodError: 
> com.ctc.wstx.stax.WstxInputFactory.createSR(Lcom/ctc/wstx/api/ReaderConfig;Lcom/ctc/wstx/io/SystemId;Lcom/ctc/wstx/io/InputBootstrapper;ZZ)Lorg/codehaus/stax2/XMLStreamReader2;
>     at 
> org.apache.hadoop.conf.Configuration.parse(Configuration.java:2803)
>     at 
> org.apache.hadoop.conf.Configuration.parse(Configuration.java:2787)
>     at 
> org.apache.hadoop.conf.Configuration.loadResource(Configuration.java:2838)
>     at 
> org.apache.hadoop.conf.Configuration.loadResources(Configuration.java:2812)
>     at 
> org.apache.hadoop.conf.Configuration.getProps(Configuration.java:2689)
>     at org.apache.hadoop.conf.Configuration.get(Configuration.java:1160)
>     at 
> org.apache.hadoop.conf.Configuration.getTrimmed(Configuration.java:1214)
>     at 
> org.apache.hadoop.conf.Configuration.getBoolean(Configuration.java:1620)
>     at 
> org.apache.hadoop.hbase.HBaseConfiguration.checkDefaultsVersion(HBaseConfiguration.java:66)
>     at 
> org.apache.hadoop.hbase.HBaseConfiguration.addHbaseResources(HBaseConfiguration.java:80)
>     at 
> org.apache.hadoop.hbase.HBaseConfiguration.create(HBaseConfiguration.java:94)
>     at 
> org.apache.hadoop.hbase.util.HBaseConfTool.main(HBaseConfTool.java:39)
> SLF4J: Class path contains multiple SLF4J bindings.
> SLF4J: Found binding in 
> [jar:file:/opt/hbase/lib/phoenix-5.0.0-HBase-2.0-client.jar!/org/slf4j/impl/StaticLoggerBinder.class]
> SLF4J: Found binding in 
> [jar:file:/opt/hbase/lib/slf4j-log4j12-1.7.25.jar!/org/slf4j/impl/StaticLoggerBinder.class]
> SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an 
> explanation.
> Exception in thread "main" java.lang.NoSuchMethodError: 
> com.ctc.wstx.stax.WstxInputFactory.createSR(Lcom/ctc/wstx/api/ReaderConfig;Lcom/ctc/wstx/io/SystemId;Lcom/ctc/wstx/io/InputBootstrapper;ZZ)Lorg/codehaus/stax2/XMLStreamReader2;
>     at 
> org.apache.hadoop.conf.Configuration.parse(Configuration.java:2803)
>     at 
> org.apache.hadoop.conf.Configuration.parse(Configuration.java:2787)
>     at 
> org.apache.hadoop.conf.Configuration.loadResource(Configuration.java:2838)
>     at 
> org.apache.hadoop.conf.Configuration.loadResources(Configuration.java:2812)
>     at 
> org.apache.hadoop.conf.Configuration.getProps(Configuration.java:2689)
>     at org.apache.hadoop.conf.Configuration.get(Configuration.java:1160)
>     at 
> org.apache.hadoop.conf.Configuration.getTrimmed(Configuration.java:1214)
>     at 
> org.apache.hadoop.conf.Configuration.getBoolean(Configuration.java:1620)
>     at 
> org.apache.hadoop.hbase.HBaseConfiguration.checkDefaultsVersion(HBaseConfiguration.java:66)
>     at 
> org.apache.hadoop.hbase.HBaseConfiguration.addHbaseResources(HBaseConfiguration.java:80)
>     at 
> org.apache.hadoop.hbase.HBaseConfiguration.create(HBaseConfiguration.java:94)
>     at 
> org.apache.hadoop.hbase.zookeeper.ZKServerTool.main(ZKServerTool.java:63)
> running master, logging to 
> /opt/hbase/logs/hbase--master-m401b01-phoenix.m401b01.out
> SLF4J: Class path contains multiple SLF4J bindings.
> SLF4J: Found binding in 
> [jar:file:/opt/hbase/lib/phoenix-5.0.0-HBase-2.0-client.jar!/org/slf4j/impl/StaticLoggerBinder.class]
> SLF4J: Found binding in 
> [jar:file:/opt/hbase/lib/slf4j-log4j12-1.7.25.jar!/org/slf4j/impl/StaticLoggerBinder.class]
> SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an 
> explanation.
> Exception in thread "main" java.lang.NoSuchMethodError: 
> com.ctc.wstx.stax.WstxInputFactory.createSR(Lcom/ctc/wstx/api/ReaderConfig;Lcom/ctc/wstx/io/SystemId;Lcom/ctc/wstx/io/InputBootstrapper;ZZ)Lorg/codehaus/stax2/XMLStreamReader2;
>     at 
> org.apache.hadoop.conf.Configuration.parse(Configuration.java:2803)
>     at 
> org.apache.hadoop.conf.Configuration.parse(Configuration.java:2787)
>     at 
> org.apache.hadoop.conf.Configuration.loadResource(Configuration.java:2838)
>     at 
> org.apache.hadoop.conf.Configuration.loadResources(Configuration.java:2812)
>     at 
> org.apache.hadoop.conf.Configuration.getProps(Configuration.java:2689)
> : running regionserver, logging to 
> /opt/hbase/logs/hbase--regionserver-m401b01-phoenix.m401b01.out
> : SLF4J: Class path contains multiple SLF4J bindings.
> : SLF4J: Found binding in 
> [jar:file:/opt/hbase/lib/phoenix-5.0.0-HBase-2.0-client.jar!/org/slf4j/impl/StaticLoggerBinder.class]
> : SLF4J: Found binding in 
> [jar:file:/opt/hbase/lib/slf4j-log4j12-1.7.25.jar!/org/slf4j/impl/StaticLoggerBinder.class]
> : SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an 
> explanation.
> : Exception in thread "main" java.lang.NoSuchMethodError: 
> com.ctc.wstx.stax.WstxInputFactory.createSR(Lcom/ctc/wstx/api/ReaderConfig;Lcom/ctc/wstx/io/SystemId;Lcom/ctc/wstx/io/InputBootstrapper;ZZ)Lorg/codehaus/stax2/XMLStreamReader2;
> :     at 
> org.apache.hadoop.conf.Configuration.parse(Configuration.java:2803)
> :     at 
> org.apache.hadoop.conf.Configuration.parse(Configuration.java:2787)
> :     at 
> org.apache.hadoop.conf.Configuration.loadResource(Configuration.java:2838)
> :     at 
> org.apache.hadoop.conf.Configuration.loadResources(Configuration.java:2812)
> :     at 
> org.apache.hadoop.conf.Configuration.getProps(Configuration.java:2689)
>
> I built HBase with 2.8.5 and was able to resolve the hasKerberosKeyTab 
> methodNotFound error. The only problem is that it took 4 hours to run 
> an automated build on Docker cloud to build HBase and the build failed 
> eventually. I think I am going to download the Hadoop jars
> from maven, rather than build HBase.
>
> On 27/09/2018 12:56 AM, Josh Elser wrote:
>> If you're using HBase with Hadoop3, HBase should have Hadoop3 jars.
>>
>> Re-build HBase using the -Dhadoop.profile=3.0 (I think it is) CLI 
>> option.
>>
>> On 9/26/18 7:21 AM, Francis Chuang wrote:
>>> Upon further investigation, it appears that this is because 
>>> org.apache.hadoop.security.authentication.util.KerberosUtil.hasKerberosKeyTab

>>> is only available in Hadoop 2.8+. HBase ships with Hadoop 2.7.4 jars.
>>>
>>> I noticed that Hadoop was bumped from 2.7.4 to 3.0.0 a few months 
>>> ago to fix PQS/Avatica issues: 
>>> https://github.com/apache/phoenix/blame/master/pom.xml#L70
>>>
>>> I think this causes Phoenix to expect some things that are available 
>>> in Hadoop 3.0.0, but are not present in HBase's Hadoop 2.7.4 jars.
>>>
>>> I think I can try and replace the hadoop-*.jar files in hbase/lib 
>>> with the equivalent 2.8.5 versions, however I am not familiar with 
>>> Java and the hadoop project, so I am not sure if this is going to 
>>> introduce issues.
>>>
>>> On 26/09/2018 4:44 PM, Francis Chuang wrote:
>>>> I wonder if this is because:
>>>> - HBase's binary distribution ships with Hadoop 2.7.4 jars.
>>>> - Phoenix 5.0.0 has Hadoop 3.0.0 declared in its pom.xml: 
>>>> https://github.com/apache/phoenix/blob/8a819c6c3b4befce190c6ac759f744df511de61d/pom.xml#L70

>>>>
>>>> - Tephra has Hadoop 2.2.0 declared in its pom.xml: 
>>>> https://github.com/apache/incubator-tephra/blob/master/pom.xml#L211
>>>>
>>>> On 26/09/2018 4:03 PM, Francis Chuang wrote:
>>>>> Hi all,
>>>>>
>>>>> I am using Phoenix 5.0.0 with HBase 2.0.0. I am seeing errors 
>>>>> while trying to create transactional tables using Phoenix.
>>>>>
>>>>> I am using my Phoenix + HBase all in one docker image available 
>>>>> here: https://github.com/Boostport/hbase-phoenix-all-in-one
>>>>>
>>>>> This is the error: 
>>>>> org.apache.phoenix.shaded.org.apache.thrift.TException: Unable to 
>>>>> discover transaction service. -> TException: Unable to discover 
>>>>> transaction service.
>>>>>
>>>>> I checked the tephra logs and got the following:
>>>>>
>>>>> Exception in thread "HDFSTransactionStateStorage STARTING" 
>>>>> Exception in thread "ThriftRPCServer" 
>>>>> com.google.common.util.concurrent.ExecutionError: 
>>>>> java.lang.NoSuchMethodError: 
>>>>> org.apache.hadoop.security.authentication.util.KerberosUtil.hasKerberosKeyTab(Ljavax/security/auth/Subject;)Z

>>>>>
>>>>>     at 
>>>>> com.google.common.util.concurrent.Futures.wrapAndThrowUnchecked(Futures.java:1008)

>>>>>
>>>>>     at 
>>>>> com.google.common.util.concurrent.Futures.getUnchecked(Futures.java:1001)

>>>>>
>>>>>     at 
>>>>> com.google.common.util.concurrent.AbstractService.startAndWait(AbstractService.java:220)

>>>>>
>>>>>     at 
>>>>> com.google.common.util.concurrent.AbstractIdleService.startAndWait(AbstractIdleService.java:106)

>>>>>
>>>>>     at 
>>>>> org.apache.tephra.TransactionManager.doStart(TransactionManager.java:245)

>>>>>
>>>>>     at 
>>>>> com.google.common.util.concurrent.AbstractService.start(AbstractService.java:170)

>>>>>
>>>>>     at 
>>>>> com.google.common.util.concurrent.AbstractService.startAndWait(AbstractService.java:220)

>>>>>
>>>>>     at 
>>>>> org.apache.tephra.distributed.TransactionServiceThriftHandler.init(TransactionServiceThriftHandler.java:249)

>>>>>
>>>>>     at 
>>>>> org.apache.tephra.rpc.ThriftRPCServer.startUp(ThriftRPCServer.java:177)

>>>>>
>>>>>     at 
>>>>> com.google.common.util.concurrent.AbstractExecutionThreadService$1$1.run(AbstractExecutionThreadService.java:47)

>>>>>
>>>>>     at java.lang.Thread.run(Thread.java:748)
>>>>> Caused by: java.lang.NoSuchMethodError: 
>>>>> org.apache.hadoop.security.authentication.util.KerberosUtil.hasKerberosKeyTab(Ljavax/security/auth/Subject;)Z

>>>>>
>>>>>     at 
>>>>> org.apache.hadoop.security.UserGroupInformation.<init>(UserGroupInformation.java:715)

>>>>>
>>>>>     at 
>>>>> org.apache.hadoop.security.UserGroupInformation.loginUserFromSubject(UserGroupInformation.java:925)

>>>>>
>>>>>     at 
>>>>> org.apache.hadoop.security.UserGroupInformation.getLoginUser(UserGroupInformation.java:873)

>>>>>
>>>>>     at 
>>>>> org.apache.hadoop.security.UserGroupInformation.getCurrentUser(UserGroupInformation.java:740)

>>>>>
>>>>>     at 
>>>>> org.apache.hadoop.fs.FileSystem$Cache$Key.<init>(FileSystem.java:3472)
>>>>>     at 
>>>>> org.apache.hadoop.fs.FileSystem$Cache.getUnique(FileSystem.java:3310)
>>>>>     at 
>>>>> org.apache.hadoop.fs.FileSystem.newInstance(FileSystem.java:529)
>>>>>     at 
>>>>> org.apache.tephra.persist.HDFSTransactionStateStorage.startUp(HDFSTransactionStateStorage.java:104)

>>>>>
>>>>>     at 
>>>>> com.google.common.util.concurrent.AbstractIdleService$1$1.run(AbstractIdleService.java:43)

>>>>>
>>>>>     ... 1 more
>>>>> 2018-09-26 04:31:11,290 INFO [leader-election-tx.service-leader] 
>>>>> distributed.TransactionService 
>>>>> (TransactionService.java:leader(115)) - Transaction Thrift Service 
>>>>> didn't start on /0.0.0.0:15165
>>>>> java.lang.NoSuchMethodError: 
>>>>> org.apache.hadoop.security.authentication.util.KerberosUtil.hasKerberosKeyTab(Ljavax/security/auth/Subject;)Z

>>>>>
>>>>>     at 
>>>>> org.apache.hadoop.security.UserGroupInformation.<init>(UserGroupInformation.java:715)

>>>>>
>>>>>     at 
>>>>> org.apache.hadoop.security.UserGroupInformation.loginUserFromSubject(UserGroupInformation.java:925)

>>>>>
>>>>>     at 
>>>>> org.apache.hadoop.security.UserGroupInformation.getLoginUser(UserGroupInformation.java:873)

>>>>>
>>>>>     at 
>>>>> org.apache.hadoop.security.UserGroupInformation.getCurrentUser(UserGroupInformation.java:740)

>>>>>
>>>>>     at 
>>>>> org.apache.hadoop.fs.FileSystem$Cache$Key.<init>(FileSystem.java:3472)
>>>>>     at 
>>>>> org.apache.hadoop.fs.FileSystem$Cache.getUnique(FileSystem.java:3310)
>>>>>     at 
>>>>> org.apache.hadoop.fs.FileSystem.newInstance(FileSystem.java:529)
>>>>>     at 
>>>>> org.apache.tephra.persist.HDFSTransactionStateStorage.startUp(HDFSTransactionStateStorage.java:104)

>>>>>
>>>>>     at 
>>>>> com.google.common.util.concurrent.AbstractIdleService$1$1.run(AbstractIdleService.java:43)

>>>>>
>>>>>     at java.lang.Thread.run(Thread.java:748)
>>>>>
>>>>> I know that HBase ships with the Hadoop 2.7.4 jars and I was not 
>>>>> able to find "hasKerberosKeyTab" grepping through the source code 
>>>>> for hadoop 2.7.4. However, I checked the Hadoop 2.7.4 source files 
>>>>> from the stack trace above and the line numbers do not match up.
>>>>>
>>>>> Interestingly, I only see this issue on my older machine (Core i7 
>>>>> 920 with 12GB of RAM) and Gitlab's CI environment (a Google Cloud 
>>>>> n1-standard-1 instance with 1vCPU and 3.75GB of RAM). I know 
>>>>> Michael also encountered this problem while running the Phoenix 
>>>>> tests for calcite-avatica-go on an older i5 machine from 2011.
>>>>>
>>>>> It does seem to be pretty weird that we are only seeing this on 
>>>>> machines where the CPU is not very powerful.
>>>>>
>>>>> I also printed the classpath for tephra by doing:
>>>>>
>>>>> $ # export HBASE_CONF_DIR=/opt/hbase/conf
>>>>> $ # export HBASE_CP=/opt/hbase/lib
>>>>> $ # export HBASE_HOME=/opt/hbase
>>>>> $ # /opt/hbase/bin/tephra classpath
>>>>> /opt/hbase/bin/../lib/*:/opt/hbase/bin/../conf/:/opt/hbase/phoenix-client/target/*:/opt/hbase/conf:/usr/lib/jvm/java-1.8-openjdk/jre/lib/tools.jar:/opt/hbase:/opt/hbase/lib/aopalliance-1.0.jar:/opt/hbase/lib/aopalliance-repackaged-2.5.0-b32.jar:/opt/hbase/lib/apacheds-i18n-2.0.0-M15.jar:/opt/hbase/lib/apacheds-kerberos-codec-2.0.0-M15.jar:/opt/hbase/lib/api-asn1-api-1.0.0-M20.jar:/opt/hbase/lib/api-util-1.0.0-M20.jar:/opt/hbase/lib/asm-3.1.jar:/opt/hbase/lib/audience-annotations-0.5.0.jar:/opt/hbase/lib/avro-1.7.7.jar:/opt/hbase/lib/commons-beanutils-core-1.8.0.jar:/opt/hbase/lib/commons-cli-1.2.jar:/opt/hbase/lib/commons-codec-1.10.jar:/opt/hbase/lib/commons-collections-3.2.2.jar:/opt/hbase/lib/commons-compress-1.4.1.jar:/opt/hbase/lib/commons-configuration-1.6.jar:/opt/hbase/lib/commons-crypto-1.0.0.jar:/opt/hbase/lib/commons-daemon-1.0.13.jar:/opt/hbase/lib/commons-digester-1.8.jar:/opt/hbase/lib/commons-httpclient-3.1.jar:/opt/hbase/lib/commons-io-2.5.jar:/opt/hbase/lib/commons-lang-2.6.jar:/opt/hbase/lib/commons-lang3-3.6.jar:/opt/hbase/lib/commons-logging-1.2.jar:/opt/hbase/lib/commons-math3-3.6.1.jar:/opt/hbase/lib/commons-net-3.1.jar:/opt/hbase/lib/curator-client-4.0.0.jar:/opt/hbase/lib/curator-framework-4.0.0.jar:/opt/hbase/lib/curator-recipes-4.0.0.jar:/opt/hbase/lib/disruptor-3.3.6.jar:/opt/hbase/lib/findbugs-annotations-1.3.9-1.jar:/opt/hbase/lib/gson-2.2.4.jar:/opt/hbase/lib/guice-3.0.jar:/opt/hbase/lib/guice-servlet-3.0.jar:/opt/hbase/lib/hadoop-annotations-2.7.4.jar:/opt/hbase/lib/hadoop-auth-2.7.4.jar:/opt/hbase/lib/hadoop-client-2.7.4.jar:/opt/hbase/lib/hadoop-common-2.7.4-tests.jar:/opt/hbase/lib/hadoop-common-2.7.4.jar:/opt/hbase/lib/hadoop-distcp-2.7.4.jar:/opt/hbase/lib/hadoop-hdfs-2.7.4-tests.jar:/opt/hbase/lib/hadoop-hdfs-2.7.4.jar:/opt/hbase/lib/hadoop-mapreduce-client-app-2.7.4.jar:/opt/hbase/lib/hadoop-mapreduce-client-common-2.7.4.jar:/opt/hbase/lib/hadoop-mapreduce-client-core-2.7.4.jar:/opt/hbase/lib/hadoop-mapreduce-client-hs-2.7.4.jar:/opt/hbase/lib/hadoop-mapreduce-client-jobclient-2.7.4.jar:/opt/hbase/lib/hadoop-mapreduce-client-shuffle-2.7.4.jar:/opt/hbase/lib/hadoop-minicluster-2.7.4.jar:/opt/hbase/lib/hadoop-yarn-api-2.7.4.jar:/opt/hbase/lib/hadoop-yarn-client-2.7.4.jar:/opt/hbase/lib/hadoop-yarn-common-2.7.4.jar:/opt/hbase/lib/hadoop-yarn-server-applicationhistoryservice-2.7.4.jar:/opt/hbase/lib/hadoop-yarn-server-common-2.7.4.jar:/opt/hbase/lib/hadoop-yarn-server-nodemanager-2.7.4.jar:/opt/hbase/lib/hadoop-yarn-server-resourcemanager-2.7.4.jar:/opt/hbase/lib/hadoop-yarn-server-tests-2.7.4-tests.jar:/opt/hbase/lib/hadoop-yarn-server-web-proxy-2.7.4.jar:/opt/hbase/lib/hbase-annotations-2.0.0-tests.jar:/opt/hbase/lib/hbase-annotations-2.0.0.jar:/opt/hbase/lib/hbase-client-2.0.0.jar:/opt/hbase/lib/hbase-common-2.0.0-tests.jar:/opt/hbase/lib/hbase-common-2.0.0.jar:/opt/hbase/lib/hbase-endpoint-2.0.0.jar:/opt/hbase/lib/hbase-examples-2.0.0.jar:/opt/hbase/lib/hbase-external-blockcache-2.0.0.jar:/opt/hbase/lib/hbase-hadoop-compat-2.0.0-tests.jar:/opt/hbase/lib/hbase-hadoop-compat-2.0.0.jar:/opt/hbase/lib/hbase-hadoop2-compat-2.0.0-tests.jar:/opt/hbase/lib/hbase-hadoop2-compat-2.0.0.jar:/opt/hbase/lib/hbase-http-2.0.0.jar:/opt/hbase/lib/hbase-it-2.0.0-tests.jar:/opt/hbase/lib/hbase-it-2.0.0.jar:/opt/hbase/lib/hbase-mapreduce-2.0.0-tests.jar:/opt/hbase/lib/hbase-mapreduce-2.0.0.jar:/opt/hbase/lib/hbase-metrics-2.0.0.jar:/opt/hbase/lib/hbase-metrics-api-2.0.0.jar:/opt/hbase/lib/hbase-procedure-2.0.0.jar:/opt/hbase/lib/hbase-protocol-2.0.0.jar:/opt/hbase/lib/hbase-protocol-shaded-2.0.0.jar:/opt/hbase/lib/hbase-replication-2.0.0.jar:/opt/hbase/lib/hbase-resource-bundle-2.0.0.jar:/opt/hbase/lib/hbase-rest-2.0.0.jar:/opt/hbase/lib/hbase-rsgroup-2.0.0-tests.jar:/opt/hbase/lib/hbase-rsgroup-2.0.0.jar:/opt/hbase/lib/hbase-server-2.0.0-tests.jar:/opt/hbase/lib/hbase-server-2.0.0.jar:/opt/hbase/lib/hbase-shaded-miscellaneous-2.1.0.jar:/opt/hbase/lib/hbase-shaded-netty-2.1.0.jar:/opt/hbase/lib/hbase-shaded-protobuf-2.1.0.jar:/opt/hbase/lib/hbase-shell-2.0.0.jar:/opt/hbase/lib/hbase-testing-util-2.0.0.jar:/opt/hbase/lib/hbase-thrift-2.0.0.jar:/opt/hbase/lib/hbase-zookeeper-2.0.0-tests.jar:/opt/hbase/lib/hbase-zookeeper-2.0.0.jar:/opt/hbase/lib/hk2-api-2.5.0-b32.jar:/opt/hbase/lib/hk2-locator-2.5.0-b32.jar:/opt/hbase/lib/hk2-utils-2.5.0-b32.jar:/opt/hbase/lib/htrace-core-3.2.0-incubating.jar:/opt/hbase/lib/htrace-core4-4.2.0-incubating.jar:/opt/hbase/lib/httpclient-4.5.3.jar:/opt/hbase/lib/httpcore-4.4.6.jar:/opt/hbase/lib/jackson-annotations-2.9.0.jar:/opt/hbase/lib/jackson-core-2.9.2.jar:/opt/hbase/lib/jackson-core-asl-1.9.13.jar:/opt/hbase/lib/jackson-databind-2.9.2.jar:/opt/hbase/lib/jackson-jaxrs-1.8.3.jar:/opt/hbase/lib/jackson-jaxrs-base-2.9.2.jar:/opt/hbase/lib/jackson-jaxrs-json-provider-2.9.2.jar:/opt/hbase/lib/jackson-mapper-asl-1.9.13.jar:/opt/hbase/lib/jackson-module-jaxb-annotations-2.9.2.jar:/opt/hbase/lib/jackson-xc-1.8.3.jar:/opt/hbase/lib/jamon-runtime-2.4.1.jar:/opt/hbase/lib/java-xmlbuilder-0.4.jar:/opt/hbase/lib/javassist-3.20.0-GA.jar:/opt/hbase/lib/javax.annotation-api-1.2.jar:/opt/hbase/lib/javax.el-3.0.1-b08.jar:/opt/hbase/lib/javax.inject-2.5.0-b32.jar:/opt/hbase/lib/javax.servlet-api-3.1.0.jar:/opt/hbase/lib/javax.servlet.jsp-2.3.2.jar:/opt/hbase/lib/javax.servlet.jsp-api-2.3.1.jar:/opt/hbase/lib/javax.servlet.jsp.jstl-1.2.0.v201105211821.jar:/opt/hbase/lib/javax.servlet.jsp.jstl-1.2.2.jar:/opt/hbase/lib/javax.ws.rs-api-2.0.1.jar:/opt/hbase/lib/jaxb-api-2.2.12.jar:/opt/hbase/lib/jaxb-impl-2.2.3-1.jar:/opt/hbase/lib/jcodings-1.0.18.jar:/opt/hbase/lib/jersey-client-2.25.1.jar:/opt/hbase/lib/jersey-common-2.25.1.jar:/opt/hbase/lib/jersey-container-servlet-core-2.25.1.jar:/opt/hbase/lib/jersey-guava-2.25.1.jar:/opt/hbase/lib/jersey-media-jaxb-2.25.1.jar:/opt/hbase/lib/jersey-server-2.25.1.jar:/opt/hbase/lib/jets3t-0.9.0.jar:/opt/hbase/lib/jettison-1.3.8.jar:/opt/hbase/lib/jetty-6.1.26.jar:/opt/hbase/lib/jetty-http-9.3.19.v20170502.jar:/opt/hbase/lib/jetty-io-9.3.19.v20170502.jar:/opt/hbase/lib/jetty-jmx-9.3.19.v20170502.jar:/opt/hbase/lib/jetty-jsp-9.2.19.v20160908.jar:/opt/hbase/lib/jetty-schemas-3.1.M0.jar:/opt/hbase/lib/jetty-security-9.3.19.v20170502.jar:/opt/hbase/lib/jetty-server-9.3.19.v20170502.jar:/opt/hbase/lib/jetty-servlet-9.3.19.v20170502.jar:/opt/hbase/lib/jetty-sslengine-6.1.26.jar:/opt/hbase/lib/jetty-util-6.1.26.jar:/opt/hbase/lib/jetty-util-9.3.19.v20170502.jar:/opt/hbase/lib/jetty-util-ajax-9.3.19.v20170502.jar:/opt/hbase/lib/jetty-webapp-9.3.19.v20170502.jar:/opt/hbase/lib/jetty-xml-9.3.19.v20170502.jar:/opt/hbase/lib/joni-2.1.11.jar:/opt/hbase/lib/jsch-0.1.54.jar:/opt/hbase/lib/junit-4.12.jar:/opt/hbase/lib/leveldbjni-all-1.8.jar:/opt/hbase/lib/libthrift-0.9.3.jar:/opt/hbase/lib/log4j-1.2.17.jar:/opt/hbase/lib/metrics-core-3.2.1.jar:/opt/hbase/lib/netty-all-4.0.23.Final.jar:/opt/hbase/lib/org.eclipse.jdt.core-3.8.2.v20130121.jar:/opt/hbase/lib/osgi-resource-locator-1.0.1.jar:/opt/hbase/lib/paranamer-2.3.jar:/opt/hbase/lib/phoenix-5.0.0-HBase-2.0-client.jar:/opt/hbase/lib/phoenix-5.0.0-HBase-2.0-server.jar:/opt/hbase/lib/protobuf-java-2.5.0.jar:/opt/hbase/lib/remotecontent?filepath=com%2Fgoogle%2Fguava%2Fguava%2F13.0.1%2Fguava-13.0.1.jar:/opt/hbase/lib/slf4j-api-1.7.25.jar:/opt/hbase/lib/slf4j-log4j12-1.7.25.jar:/opt/hbase/lib/snappy-java-1.0.5.jar:/opt/hbase/lib/spymemcached-2.12.2.jar:/opt/hbase/lib/validation-api-1.1.0.Final.jar:/opt/hbase/lib/xmlenc-0.52.jar:/opt/hbase/lib/xz-1.0.jar:/opt/hbase/lib/zookeeper-3.4.10.jar::

>>>>>
>>>>>
>>>>> Does anyone know what might be causing this or any tips for 
>>>>> troubleshooting?
>>>>>
>>>>> Francis
>>>>>
>>>>>
>>>>
>>>>
>>>
>>>
>
>



Mime
View raw message