phoenix-user mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From Francis Chuang <francischu...@apache.org>
Subject Re: org.apache.phoenix.shaded.org.apache.thrift.TException: Unable to discover transaction service. -> TException: Unable to discover transaction service.
Date Thu, 27 Sep 2018 22:24:04 GMT
I tried updating my hbase-phoenix-all-in-one image to use HBase built 
with Hadoop 3. Unfortunately, it didn't work. I think this might be 
because Hadoop 3.0.0 is too new for tephra (which uses 2.2.0):

starting Query Server, logging to /tmp/phoenix/phoenix-root-queryserver.log
Thu Sep 27 02:50:56 UTC 2018 Starting tephra service on 
m401b01-phoenix.m401b01
Running class org.apache.tephra.TransactionServiceMain
SLF4J: Class path contains multiple SLF4J bindings.
SLF4J: Found binding in 
[jar:file:/opt/hbase/lib/phoenix-5.0.0-HBase-2.0-client.jar!/org/slf4j/impl/StaticLoggerBinder.class]
SLF4J: Found binding in 
[jar:file:/opt/hbase/lib/slf4j-log4j12-1.7.25.jar!/org/slf4j/impl/StaticLoggerBinder.class]
SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an 
explanation.
Exception in thread "main" java.lang.NoSuchMethodError: 
com.ctc.wstx.stax.WstxInputFactory.createSR(Lcom/ctc/wstx/api/ReaderConfig;Lcom/ctc/wstx/io/SystemId;Lcom/ctc/wstx/io/InputBootstrapper;ZZ)Lorg/codehaus/stax2/XMLStreamReader2;
     at org.apache.hadoop.conf.Configuration.parse(Configuration.java:2803)
     at org.apache.hadoop.conf.Configuration.parse(Configuration.java:2787)
     at 
org.apache.hadoop.conf.Configuration.loadResource(Configuration.java:2838)
     at 
org.apache.hadoop.conf.Configuration.loadResources(Configuration.java:2812)
     at 
org.apache.hadoop.conf.Configuration.getProps(Configuration.java:2689)
     at org.apache.hadoop.conf.Configuration.get(Configuration.java:1160)
     at 
org.apache.hadoop.conf.Configuration.getTrimmed(Configuration.java:1214)
     at 
org.apache.hadoop.conf.Configuration.getBoolean(Configuration.java:1620)
     at 
org.apache.hadoop.hbase.HBaseConfiguration.checkDefaultsVersion(HBaseConfiguration.java:66)
     at 
org.apache.hadoop.hbase.HBaseConfiguration.addHbaseResources(HBaseConfiguration.java:80)
     at 
org.apache.hadoop.hbase.HBaseConfiguration.create(HBaseConfiguration.java:94)
     at 
org.apache.hadoop.hbase.util.HBaseConfTool.main(HBaseConfTool.java:39)
SLF4J: Class path contains multiple SLF4J bindings.
SLF4J: Found binding in 
[jar:file:/opt/hbase/lib/phoenix-5.0.0-HBase-2.0-client.jar!/org/slf4j/impl/StaticLoggerBinder.class]
SLF4J: Found binding in 
[jar:file:/opt/hbase/lib/slf4j-log4j12-1.7.25.jar!/org/slf4j/impl/StaticLoggerBinder.class]
SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an 
explanation.
Exception in thread "main" java.lang.NoSuchMethodError: 
com.ctc.wstx.stax.WstxInputFactory.createSR(Lcom/ctc/wstx/api/ReaderConfig;Lcom/ctc/wstx/io/SystemId;Lcom/ctc/wstx/io/InputBootstrapper;ZZ)Lorg/codehaus/stax2/XMLStreamReader2;
     at org.apache.hadoop.conf.Configuration.parse(Configuration.java:2803)
     at org.apache.hadoop.conf.Configuration.parse(Configuration.java:2787)
     at 
org.apache.hadoop.conf.Configuration.loadResource(Configuration.java:2838)
     at 
org.apache.hadoop.conf.Configuration.loadResources(Configuration.java:2812)
     at 
org.apache.hadoop.conf.Configuration.getProps(Configuration.java:2689)
     at org.apache.hadoop.conf.Configuration.get(Configuration.java:1160)
     at 
org.apache.hadoop.conf.Configuration.getTrimmed(Configuration.java:1214)
     at 
org.apache.hadoop.conf.Configuration.getBoolean(Configuration.java:1620)
     at 
org.apache.hadoop.hbase.HBaseConfiguration.checkDefaultsVersion(HBaseConfiguration.java:66)
     at 
org.apache.hadoop.hbase.HBaseConfiguration.addHbaseResources(HBaseConfiguration.java:80)
     at 
org.apache.hadoop.hbase.HBaseConfiguration.create(HBaseConfiguration.java:94)
     at 
org.apache.hadoop.hbase.zookeeper.ZKServerTool.main(ZKServerTool.java:63)
running master, logging to 
/opt/hbase/logs/hbase--master-m401b01-phoenix.m401b01.out
SLF4J: Class path contains multiple SLF4J bindings.
SLF4J: Found binding in 
[jar:file:/opt/hbase/lib/phoenix-5.0.0-HBase-2.0-client.jar!/org/slf4j/impl/StaticLoggerBinder.class]
SLF4J: Found binding in 
[jar:file:/opt/hbase/lib/slf4j-log4j12-1.7.25.jar!/org/slf4j/impl/StaticLoggerBinder.class]
SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an 
explanation.
Exception in thread "main" java.lang.NoSuchMethodError: 
com.ctc.wstx.stax.WstxInputFactory.createSR(Lcom/ctc/wstx/api/ReaderConfig;Lcom/ctc/wstx/io/SystemId;Lcom/ctc/wstx/io/InputBootstrapper;ZZ)Lorg/codehaus/stax2/XMLStreamReader2;
     at org.apache.hadoop.conf.Configuration.parse(Configuration.java:2803)
     at org.apache.hadoop.conf.Configuration.parse(Configuration.java:2787)
     at 
org.apache.hadoop.conf.Configuration.loadResource(Configuration.java:2838)
     at 
org.apache.hadoop.conf.Configuration.loadResources(Configuration.java:2812)
     at 
org.apache.hadoop.conf.Configuration.getProps(Configuration.java:2689)
: running regionserver, logging to 
/opt/hbase/logs/hbase--regionserver-m401b01-phoenix.m401b01.out
: SLF4J: Class path contains multiple SLF4J bindings.
: SLF4J: Found binding in 
[jar:file:/opt/hbase/lib/phoenix-5.0.0-HBase-2.0-client.jar!/org/slf4j/impl/StaticLoggerBinder.class]
: SLF4J: Found binding in 
[jar:file:/opt/hbase/lib/slf4j-log4j12-1.7.25.jar!/org/slf4j/impl/StaticLoggerBinder.class]
: SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an 
explanation.
: Exception in thread "main" java.lang.NoSuchMethodError: 
com.ctc.wstx.stax.WstxInputFactory.createSR(Lcom/ctc/wstx/api/ReaderConfig;Lcom/ctc/wstx/io/SystemId;Lcom/ctc/wstx/io/InputBootstrapper;ZZ)Lorg/codehaus/stax2/XMLStreamReader2;
:     at org.apache.hadoop.conf.Configuration.parse(Configuration.java:2803)
:     at org.apache.hadoop.conf.Configuration.parse(Configuration.java:2787)
:     at 
org.apache.hadoop.conf.Configuration.loadResource(Configuration.java:2838)
:     at 
org.apache.hadoop.conf.Configuration.loadResources(Configuration.java:2812)
:     at 
org.apache.hadoop.conf.Configuration.getProps(Configuration.java:2689)

I built HBase with 2.8.5 and was able to resolve the hasKerberosKeyTab 
methodNotFound error. The only problem is that it took 4 hours to run an 
automated build on Docker cloud to build HBase and the build failed 
eventually. I think I am going to download the Hadoop jars
from maven, rather than build HBase.

On 27/09/2018 12:56 AM, Josh Elser wrote:
> If you're using HBase with Hadoop3, HBase should have Hadoop3 jars.
>
> Re-build HBase using the -Dhadoop.profile=3.0 (I think it is) CLI option.
>
> On 9/26/18 7:21 AM, Francis Chuang wrote:
>> Upon further investigation, it appears that this is because 
>> org.apache.hadoop.security.authentication.util.KerberosUtil.hasKerberosKeyTab 
>> is only available in Hadoop 2.8+. HBase ships with Hadoop 2.7.4 jars.
>>
>> I noticed that Hadoop was bumped from 2.7.4 to 3.0.0 a few months ago 
>> to fix PQS/Avatica issues: 
>> https://github.com/apache/phoenix/blame/master/pom.xml#L70
>>
>> I think this causes Phoenix to expect some things that are available 
>> in Hadoop 3.0.0, but are not present in HBase's Hadoop 2.7.4 jars.
>>
>> I think I can try and replace the hadoop-*.jar files in hbase/lib 
>> with the equivalent 2.8.5 versions, however I am not familiar with 
>> Java and the hadoop project, so I am not sure if this is going to 
>> introduce issues.
>>
>> On 26/09/2018 4:44 PM, Francis Chuang wrote:
>>> I wonder if this is because:
>>> - HBase's binary distribution ships with Hadoop 2.7.4 jars.
>>> - Phoenix 5.0.0 has Hadoop 3.0.0 declared in its pom.xml: 
>>> https://github.com/apache/phoenix/blob/8a819c6c3b4befce190c6ac759f744df511de61d/pom.xml#L70

>>>
>>> - Tephra has Hadoop 2.2.0 declared in its pom.xml: 
>>> https://github.com/apache/incubator-tephra/blob/master/pom.xml#L211
>>>
>>> On 26/09/2018 4:03 PM, Francis Chuang wrote:
>>>> Hi all,
>>>>
>>>> I am using Phoenix 5.0.0 with HBase 2.0.0. I am seeing errors while 
>>>> trying to create transactional tables using Phoenix.
>>>>
>>>> I am using my Phoenix + HBase all in one docker image available 
>>>> here: https://github.com/Boostport/hbase-phoenix-all-in-one
>>>>
>>>> This is the error: 
>>>> org.apache.phoenix.shaded.org.apache.thrift.TException: Unable to 
>>>> discover transaction service. -> TException: Unable to discover 
>>>> transaction service.
>>>>
>>>> I checked the tephra logs and got the following:
>>>>
>>>> Exception in thread "HDFSTransactionStateStorage STARTING" 
>>>> Exception in thread "ThriftRPCServer" 
>>>> com.google.common.util.concurrent.ExecutionError: 
>>>> java.lang.NoSuchMethodError: 
>>>> org.apache.hadoop.security.authentication.util.KerberosUtil.hasKerberosKeyTab(Ljavax/security/auth/Subject;)Z

>>>>
>>>>     at 
>>>> com.google.common.util.concurrent.Futures.wrapAndThrowUnchecked(Futures.java:1008)

>>>>
>>>>     at 
>>>> com.google.common.util.concurrent.Futures.getUnchecked(Futures.java:1001)

>>>>
>>>>     at 
>>>> com.google.common.util.concurrent.AbstractService.startAndWait(AbstractService.java:220)

>>>>
>>>>     at 
>>>> com.google.common.util.concurrent.AbstractIdleService.startAndWait(AbstractIdleService.java:106)

>>>>
>>>>     at 
>>>> org.apache.tephra.TransactionManager.doStart(TransactionManager.java:245)

>>>>
>>>>     at 
>>>> com.google.common.util.concurrent.AbstractService.start(AbstractService.java:170)

>>>>
>>>>     at 
>>>> com.google.common.util.concurrent.AbstractService.startAndWait(AbstractService.java:220)

>>>>
>>>>     at 
>>>> org.apache.tephra.distributed.TransactionServiceThriftHandler.init(TransactionServiceThriftHandler.java:249)

>>>>
>>>>     at 
>>>> org.apache.tephra.rpc.ThriftRPCServer.startUp(ThriftRPCServer.java:177) 
>>>>
>>>>     at 
>>>> com.google.common.util.concurrent.AbstractExecutionThreadService$1$1.run(AbstractExecutionThreadService.java:47)

>>>>
>>>>     at java.lang.Thread.run(Thread.java:748)
>>>> Caused by: java.lang.NoSuchMethodError: 
>>>> org.apache.hadoop.security.authentication.util.KerberosUtil.hasKerberosKeyTab(Ljavax/security/auth/Subject;)Z

>>>>
>>>>     at 
>>>> org.apache.hadoop.security.UserGroupInformation.<init>(UserGroupInformation.java:715)

>>>>
>>>>     at 
>>>> org.apache.hadoop.security.UserGroupInformation.loginUserFromSubject(UserGroupInformation.java:925)

>>>>
>>>>     at 
>>>> org.apache.hadoop.security.UserGroupInformation.getLoginUser(UserGroupInformation.java:873)

>>>>
>>>>     at 
>>>> org.apache.hadoop.security.UserGroupInformation.getCurrentUser(UserGroupInformation.java:740)

>>>>
>>>>     at 
>>>> org.apache.hadoop.fs.FileSystem$Cache$Key.<init>(FileSystem.java:3472)
>>>>     at 
>>>> org.apache.hadoop.fs.FileSystem$Cache.getUnique(FileSystem.java:3310)
>>>>     at 
>>>> org.apache.hadoop.fs.FileSystem.newInstance(FileSystem.java:529)
>>>>     at 
>>>> org.apache.tephra.persist.HDFSTransactionStateStorage.startUp(HDFSTransactionStateStorage.java:104)

>>>>
>>>>     at 
>>>> com.google.common.util.concurrent.AbstractIdleService$1$1.run(AbstractIdleService.java:43)

>>>>
>>>>     ... 1 more
>>>> 2018-09-26 04:31:11,290 INFO [leader-election-tx.service-leader] 
>>>> distributed.TransactionService 
>>>> (TransactionService.java:leader(115)) - Transaction Thrift Service 
>>>> didn't start on /0.0.0.0:15165
>>>> java.lang.NoSuchMethodError: 
>>>> org.apache.hadoop.security.authentication.util.KerberosUtil.hasKerberosKeyTab(Ljavax/security/auth/Subject;)Z

>>>>
>>>>     at 
>>>> org.apache.hadoop.security.UserGroupInformation.<init>(UserGroupInformation.java:715)

>>>>
>>>>     at 
>>>> org.apache.hadoop.security.UserGroupInformation.loginUserFromSubject(UserGroupInformation.java:925)

>>>>
>>>>     at 
>>>> org.apache.hadoop.security.UserGroupInformation.getLoginUser(UserGroupInformation.java:873)

>>>>
>>>>     at 
>>>> org.apache.hadoop.security.UserGroupInformation.getCurrentUser(UserGroupInformation.java:740)

>>>>
>>>>     at 
>>>> org.apache.hadoop.fs.FileSystem$Cache$Key.<init>(FileSystem.java:3472)
>>>>     at 
>>>> org.apache.hadoop.fs.FileSystem$Cache.getUnique(FileSystem.java:3310)
>>>>     at 
>>>> org.apache.hadoop.fs.FileSystem.newInstance(FileSystem.java:529)
>>>>     at 
>>>> org.apache.tephra.persist.HDFSTransactionStateStorage.startUp(HDFSTransactionStateStorage.java:104)

>>>>
>>>>     at 
>>>> com.google.common.util.concurrent.AbstractIdleService$1$1.run(AbstractIdleService.java:43)

>>>>
>>>>     at java.lang.Thread.run(Thread.java:748)
>>>>
>>>> I know that HBase ships with the Hadoop 2.7.4 jars and I was not 
>>>> able to find "hasKerberosKeyTab" grepping through the source code 
>>>> for hadoop 2.7.4. However, I checked the Hadoop 2.7.4 source files 
>>>> from the stack trace above and the line numbers do not match up.
>>>>
>>>> Interestingly, I only see this issue on my older machine (Core i7 
>>>> 920 with 12GB of RAM) and Gitlab's CI environment (a Google Cloud 
>>>> n1-standard-1 instance with 1vCPU and 3.75GB of RAM). I know 
>>>> Michael also encountered this problem while running the Phoenix 
>>>> tests for calcite-avatica-go on an older i5 machine from 2011.
>>>>
>>>> It does seem to be pretty weird that we are only seeing this on 
>>>> machines where the CPU is not very powerful.
>>>>
>>>> I also printed the classpath for tephra by doing:
>>>>
>>>> $ # export HBASE_CONF_DIR=/opt/hbase/conf
>>>> $ # export HBASE_CP=/opt/hbase/lib
>>>> $ # export HBASE_HOME=/opt/hbase
>>>> $ # /opt/hbase/bin/tephra classpath
>>>> /opt/hbase/bin/../lib/*:/opt/hbase/bin/../conf/:/opt/hbase/phoenix-client/target/*:/opt/hbase/conf:/usr/lib/jvm/java-1.8-openjdk/jre/lib/tools.jar:/opt/hbase:/opt/hbase/lib/aopalliance-1.0.jar:/opt/hbase/lib/aopalliance-repackaged-2.5.0-b32.jar:/opt/hbase/lib/apacheds-i18n-2.0.0-M15.jar:/opt/hbase/lib/apacheds-kerberos-codec-2.0.0-M15.jar:/opt/hbase/lib/api-asn1-api-1.0.0-M20.jar:/opt/hbase/lib/api-util-1.0.0-M20.jar:/opt/hbase/lib/asm-3.1.jar:/opt/hbase/lib/audience-annotations-0.5.0.jar:/opt/hbase/lib/avro-1.7.7.jar:/opt/hbase/lib/commons-beanutils-core-1.8.0.jar:/opt/hbase/lib/commons-cli-1.2.jar:/opt/hbase/lib/commons-codec-1.10.jar:/opt/hbase/lib/commons-collections-3.2.2.jar:/opt/hbase/lib/commons-compress-1.4.1.jar:/opt/hbase/lib/commons-configuration-1.6.jar:/opt/hbase/lib/commons-crypto-1.0.0.jar:/opt/hbase/lib/commons-daemon-1.0.13.jar:/opt/hbase/lib/commons-digester-1.8.jar:/opt/hbase/lib/commons-httpclient-3.1.jar:/opt/hbase/lib/commons-io-2.5.jar:/opt/hbase/lib/commons-lang-2.6.jar:/opt/hbase/lib/commons-lang3-3.6.jar:/opt/hbase/lib/commons-logging-1.2.jar:/opt/hbase/lib/commons-math3-3.6.1.jar:/opt/hbase/lib/commons-net-3.1.jar:/opt/hbase/lib/curator-client-4.0.0.jar:/opt/hbase/lib/curator-framework-4.0.0.jar:/opt/hbase/lib/curator-recipes-4.0.0.jar:/opt/hbase/lib/disruptor-3.3.6.jar:/opt/hbase/lib/findbugs-annotations-1.3.9-1.jar:/opt/hbase/lib/gson-2.2.4.jar:/opt/hbase/lib/guice-3.0.jar:/opt/hbase/lib/guice-servlet-3.0.jar:/opt/hbase/lib/hadoop-annotations-2.7.4.jar:/opt/hbase/lib/hadoop-auth-2.7.4.jar:/opt/hbase/lib/hadoop-client-2.7.4.jar:/opt/hbase/lib/hadoop-common-2.7.4-tests.jar:/opt/hbase/lib/hadoop-common-2.7.4.jar:/opt/hbase/lib/hadoop-distcp-2.7.4.jar:/opt/hbase/lib/hadoop-hdfs-2.7.4-tests.jar:/opt/hbase/lib/hadoop-hdfs-2.7.4.jar:/opt/hbase/lib/hadoop-mapreduce-client-app-2.7.4.jar:/opt/hbase/lib/hadoop-mapreduce-client-common-2.7.4.jar:/opt/hbase/lib/hadoop-mapreduce-client-core-2.7.4.jar:/opt/hbase/lib/hadoop-mapreduce-client-hs-2.7.4.jar:/opt/hbase/lib/hadoop-mapreduce-client-jobclient-2.7.4.jar:/opt/hbase/lib/hadoop-mapreduce-client-shuffle-2.7.4.jar:/opt/hbase/lib/hadoop-minicluster-2.7.4.jar:/opt/hbase/lib/hadoop-yarn-api-2.7.4.jar:/opt/hbase/lib/hadoop-yarn-client-2.7.4.jar:/opt/hbase/lib/hadoop-yarn-common-2.7.4.jar:/opt/hbase/lib/hadoop-yarn-server-applicationhistoryservice-2.7.4.jar:/opt/hbase/lib/hadoop-yarn-server-common-2.7.4.jar:/opt/hbase/lib/hadoop-yarn-server-nodemanager-2.7.4.jar:/opt/hbase/lib/hadoop-yarn-server-resourcemanager-2.7.4.jar:/opt/hbase/lib/hadoop-yarn-server-tests-2.7.4-tests.jar:/opt/hbase/lib/hadoop-yarn-server-web-proxy-2.7.4.jar:/opt/hbase/lib/hbase-annotations-2.0.0-tests.jar:/opt/hbase/lib/hbase-annotations-2.0.0.jar:/opt/hbase/lib/hbase-client-2.0.0.jar:/opt/hbase/lib/hbase-common-2.0.0-tests.jar:/opt/hbase/lib/hbase-common-2.0.0.jar:/opt/hbase/lib/hbase-endpoint-2.0.0.jar:/opt/hbase/lib/hbase-examples-2.0.0.jar:/opt/hbase/lib/hbase-external-blockcache-2.0.0.jar:/opt/hbase/lib/hbase-hadoop-compat-2.0.0-tests.jar:/opt/hbase/lib/hbase-hadoop-compat-2.0.0.jar:/opt/hbase/lib/hbase-hadoop2-compat-2.0.0-tests.jar:/opt/hbase/lib/hbase-hadoop2-compat-2.0.0.jar:/opt/hbase/lib/hbase-http-2.0.0.jar:/opt/hbase/lib/hbase-it-2.0.0-tests.jar:/opt/hbase/lib/hbase-it-2.0.0.jar:/opt/hbase/lib/hbase-mapreduce-2.0.0-tests.jar:/opt/hbase/lib/hbase-mapreduce-2.0.0.jar:/opt/hbase/lib/hbase-metrics-2.0.0.jar:/opt/hbase/lib/hbase-metrics-api-2.0.0.jar:/opt/hbase/lib/hbase-procedure-2.0.0.jar:/opt/hbase/lib/hbase-protocol-2.0.0.jar:/opt/hbase/lib/hbase-protocol-shaded-2.0.0.jar:/opt/hbase/lib/hbase-replication-2.0.0.jar:/opt/hbase/lib/hbase-resource-bundle-2.0.0.jar:/opt/hbase/lib/hbase-rest-2.0.0.jar:/opt/hbase/lib/hbase-rsgroup-2.0.0-tests.jar:/opt/hbase/lib/hbase-rsgroup-2.0.0.jar:/opt/hbase/lib/hbase-server-2.0.0-tests.jar:/opt/hbase/lib/hbase-server-2.0.0.jar:/opt/hbase/lib/hbase-shaded-miscellaneous-2.1.0.jar:/opt/hbase/lib/hbase-shaded-netty-2.1.0.jar:/opt/hbase/lib/hbase-shaded-protobuf-2.1.0.jar:/opt/hbase/lib/hbase-shell-2.0.0.jar:/opt/hbase/lib/hbase-testing-util-2.0.0.jar:/opt/hbase/lib/hbase-thrift-2.0.0.jar:/opt/hbase/lib/hbase-zookeeper-2.0.0-tests.jar:/opt/hbase/lib/hbase-zookeeper-2.0.0.jar:/opt/hbase/lib/hk2-api-2.5.0-b32.jar:/opt/hbase/lib/hk2-locator-2.5.0-b32.jar:/opt/hbase/lib/hk2-utils-2.5.0-b32.jar:/opt/hbase/lib/htrace-core-3.2.0-incubating.jar:/opt/hbase/lib/htrace-core4-4.2.0-incubating.jar:/opt/hbase/lib/httpclient-4.5.3.jar:/opt/hbase/lib/httpcore-4.4.6.jar:/opt/hbase/lib/jackson-annotations-2.9.0.jar:/opt/hbase/lib/jackson-core-2.9.2.jar:/opt/hbase/lib/jackson-core-asl-1.9.13.jar:/opt/hbase/lib/jackson-databind-2.9.2.jar:/opt/hbase/lib/jackson-jaxrs-1.8.3.jar:/opt/hbase/lib/jackson-jaxrs-base-2.9.2.jar:/opt/hbase/lib/jackson-jaxrs-json-provider-2.9.2.jar:/opt/hbase/lib/jackson-mapper-asl-1.9.13.jar:/opt/hbase/lib/jackson-module-jaxb-annotations-2.9.2.jar:/opt/hbase/lib/jackson-xc-1.8.3.jar:/opt/hbase/lib/jamon-runtime-2.4.1.jar:/opt/hbase/lib/java-xmlbuilder-0.4.jar:/opt/hbase/lib/javassist-3.20.0-GA.jar:/opt/hbase/lib/javax.annotation-api-1.2.jar:/opt/hbase/lib/javax.el-3.0.1-b08.jar:/opt/hbase/lib/javax.inject-2.5.0-b32.jar:/opt/hbase/lib/javax.servlet-api-3.1.0.jar:/opt/hbase/lib/javax.servlet.jsp-2.3.2.jar:/opt/hbase/lib/javax.servlet.jsp-api-2.3.1.jar:/opt/hbase/lib/javax.servlet.jsp.jstl-1.2.0.v201105211821.jar:/opt/hbase/lib/javax.servlet.jsp.jstl-1.2.2.jar:/opt/hbase/lib/javax.ws.rs-api-2.0.1.jar:/opt/hbase/lib/jaxb-api-2.2.12.jar:/opt/hbase/lib/jaxb-impl-2.2.3-1.jar:/opt/hbase/lib/jcodings-1.0.18.jar:/opt/hbase/lib/jersey-client-2.25.1.jar:/opt/hbase/lib/jersey-common-2.25.1.jar:/opt/hbase/lib/jersey-container-servlet-core-2.25.1.jar:/opt/hbase/lib/jersey-guava-2.25.1.jar:/opt/hbase/lib/jersey-media-jaxb-2.25.1.jar:/opt/hbase/lib/jersey-server-2.25.1.jar:/opt/hbase/lib/jets3t-0.9.0.jar:/opt/hbase/lib/jettison-1.3.8.jar:/opt/hbase/lib/jetty-6.1.26.jar:/opt/hbase/lib/jetty-http-9.3.19.v20170502.jar:/opt/hbase/lib/jetty-io-9.3.19.v20170502.jar:/opt/hbase/lib/jetty-jmx-9.3.19.v20170502.jar:/opt/hbase/lib/jetty-jsp-9.2.19.v20160908.jar:/opt/hbase/lib/jetty-schemas-3.1.M0.jar:/opt/hbase/lib/jetty-security-9.3.19.v20170502.jar:/opt/hbase/lib/jetty-server-9.3.19.v20170502.jar:/opt/hbase/lib/jetty-servlet-9.3.19.v20170502.jar:/opt/hbase/lib/jetty-sslengine-6.1.26.jar:/opt/hbase/lib/jetty-util-6.1.26.jar:/opt/hbase/lib/jetty-util-9.3.19.v20170502.jar:/opt/hbase/lib/jetty-util-ajax-9.3.19.v20170502.jar:/opt/hbase/lib/jetty-webapp-9.3.19.v20170502.jar:/opt/hbase/lib/jetty-xml-9.3.19.v20170502.jar:/opt/hbase/lib/joni-2.1.11.jar:/opt/hbase/lib/jsch-0.1.54.jar:/opt/hbase/lib/junit-4.12.jar:/opt/hbase/lib/leveldbjni-all-1.8.jar:/opt/hbase/lib/libthrift-0.9.3.jar:/opt/hbase/lib/log4j-1.2.17.jar:/opt/hbase/lib/metrics-core-3.2.1.jar:/opt/hbase/lib/netty-all-4.0.23.Final.jar:/opt/hbase/lib/org.eclipse.jdt.core-3.8.2.v20130121.jar:/opt/hbase/lib/osgi-resource-locator-1.0.1.jar:/opt/hbase/lib/paranamer-2.3.jar:/opt/hbase/lib/phoenix-5.0.0-HBase-2.0-client.jar:/opt/hbase/lib/phoenix-5.0.0-HBase-2.0-server.jar:/opt/hbase/lib/protobuf-java-2.5.0.jar:/opt/hbase/lib/remotecontent?filepath=com%2Fgoogle%2Fguava%2Fguava%2F13.0.1%2Fguava-13.0.1.jar:/opt/hbase/lib/slf4j-api-1.7.25.jar:/opt/hbase/lib/slf4j-log4j12-1.7.25.jar:/opt/hbase/lib/snappy-java-1.0.5.jar:/opt/hbase/lib/spymemcached-2.12.2.jar:/opt/hbase/lib/validation-api-1.1.0.Final.jar:/opt/hbase/lib/xmlenc-0.52.jar:/opt/hbase/lib/xz-1.0.jar:/opt/hbase/lib/zookeeper-3.4.10.jar::

>>>>
>>>>
>>>> Does anyone know what might be causing this or any tips for 
>>>> troubleshooting?
>>>>
>>>> Francis
>>>>
>>>>
>>>
>>>
>>
>>



Mime
View raw message