phoenix-user mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From 倪项菲<nixiangfei_...@chinamobile.com>
Subject error when using apache-phoenix-4.14.0-HBase-1.2-bin with hbase 1.2.6
Date Tue, 07 Aug 2018 01:10:50 GMT







Hi Experts,

    I am using HBase 1.2.6,the cluster is working good with HMaster HA,but
when we integrate phoenix with hbase,it failed,below are the steps

    1,download apache-phoenix-4.14.0-HBase-1.2-bin
from http://phoenix.apache.org,the copy the tar file to the HMaster and unzip the file

 
  2,copy phoenix-core-4.14.0-HBase-1.2.jar phoenix-4.14.0-HBase-1.2-server.jar to all HBase
nodes including HMaster and HRegionServer ,put them to hbasehome/lib,my path is /opt/hbase-1.2.6/lib

   3,restart hbase cluster

    4,then start to use phoenix,but it return below error:

 
    [apache@plat-ecloud01-bigdata-journalnode01 bin]$ ./sqlline.py plat-ecloud01-bigdata-zk01,plat-ecloud01-bigdata-zk02,plat-ecloud01-bigdata-zk03

Setting
property: [incremental, false]

Setting property: [isolation, TRANSACTION_READ_COMMITTED]

issuing:
!connect jdbc:phoenix:plat-ecloud01-bigdata-zk01 none none org.apache.phoenix.jdbc.PhoenixDriver

Connecting
to jdbc:phoenix:plat-ecloud01-bigdata-zk01,plat-ecloud01-bigdata-zk02,plat-ecloud01-bigdata-zk03

SLF4J:
Class path contains multiple SLF4J bindings.

SLF4J: Found binding in [jar:file:/opt/apache-phoenix-4.14.0-HBase-1.2-bin/phoenix-4.14.0-HBase-1.2-client.jar!/org/slf4j/impl/StaticLoggerBinder.class]

SLF4J:
Found binding in [jar:file:/opt/hadoop-2.7.6/share/hadoop/common/lib/slf4j-log4j12-1.7.10.jar!/org/slf4j/impl/StaticLoggerBinder.class]

SLF4J:
See http://www.slf4j.org/codes.html#multiple_bindings for an explanation.

18/08/06 18:40:08
WARN util.NativeCodeLoader: Unable to load native-hadoop library for your platform... using
builtin-java classes where applicable

Error: org.apache.hadoop.hbase.DoNotRetryIOException:
Unable to load configured region split policy 'org.apache.phoenix.schema.MetaDataSplitPolicy'
for table 'SYSTEM.CATALOG' Set hbase.table.sanity.checks to false at conf or table descriptor
if you want to bypass sanity checks

        at org.apache.hadoop.hbase.master.HMaster.warnOrThrowExceptionForFailure(HMaster.java:1754)

       at org.apache.hadoop.hbase.master.HMaster.sanityCheckTableDescriptor(HMaster.java:1615)

       at org.apache.hadoop.hbase.master.HMaster.createTable(HMaster.java:1541)

        at
org.apache.hadoop.hbase.master.MasterRpcServices.createTable(MasterRpcServices.java:463)

       at org.apache.hadoop.hbase.protobuf.generated.MasterProtos$MasterService$2.callBlockingMethod(MasterProtos.java:55682)

       at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2196)

        at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:112)

       at org.apache.hadoop.hbase.ipc.RpcExecutor.consumerLoop(RpcExecutor.java:133)

   
    at org.apache.hadoop.hbase.ipc.RpcExecutor$1.run(RpcExecutor.java:108)

        at java.lang.Thread.run(Thread.java:745)
(state=08000,code=101)

org.apache.phoenix.exception.PhoenixIOException: org.apache.hadoop.hbase.DoNotRetryIOException:
Unable to load configured region split policy 'org.apache.phoenix.schema.MetaDataSplitPolicy'
for table 'SYSTEM.CATALOG' Set hbase.table.sanity.checks to false at conf or table descriptor
if you want to bypass sanity checks

        at org.apache.hadoop.hbase.master.HMaster.warnOrThrowExceptionForFailure(HMaster.java:1754)

       at org.apache.hadoop.hbase.master.HMaster.sanityCheckTableDescriptor(HMaster.java:1615)

       at org.apache.hadoop.hbase.master.HMaster.createTable(HMaster.java:1541)

        at
org.apache.hadoop.hbase.master.MasterRpcServices.createTable(MasterRpcServices.java:463)

       at org.apache.hadoop.hbase.protobuf.generated.MasterProtos$MasterService$2.callBlockingMethod(MasterProtos.java:55682)

       at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2196)

        at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:112)

       at org.apache.hadoop.hbase.ipc.RpcExecutor.consumerLoop(RpcExecutor.java:133)

   
    at org.apache.hadoop.hbase.ipc.RpcExecutor$1.run(RpcExecutor.java:108)

        at java.lang.Thread.run(Thread.java:745)




       at org.apache.phoenix.util.ServerUtil.parseServerException(ServerUtil.java:144)

 
      at org.apache.phoenix.query.ConnectionQueryServicesImpl.ensureTableCreated(ConnectionQueryServicesImpl.java:1197)

       at org.apache.phoenix.query.ConnectionQueryServicesImpl.createTable(ConnectionQueryServicesImpl.java:1491)

       at org.apache.phoenix.schema.MetaDataClient.createTableInternal(MetaDataClient.java:2717)

       at org.apache.phoenix.schema.MetaDataClient.createTable(MetaDataClient.java:1114)

       at org.apache.phoenix.compile.CreateTableCompiler$1.execute(CreateTableCompiler.java:192)

       at org.apache.phoenix.jdbc.PhoenixStatement$2.call(PhoenixStatement.java:408)

   
    at org.apache.phoenix.jdbc.PhoenixStatement$2.call(PhoenixStatement.java:391)

      
 at org.apache.phoenix.call.CallRunner.run(CallRunner.java:53)

        at org.apache.phoenix.jdbc.PhoenixStatement.executeMutation(PhoenixStatement.java:389)

       at org.apache.phoenix.jdbc.PhoenixStatement.executeMutation(PhoenixStatement.java:378)

       at org.apache.phoenix.jdbc.PhoenixStatement.executeUpdate(PhoenixStatement.java:1806)

       at org.apache.phoenix.query.ConnectionQueryServicesImpl$12.call(ConnectionQueryServicesImpl.java:2528)

       at org.apache.phoenix.query.ConnectionQueryServicesImpl$12.call(ConnectionQueryServicesImpl.java:2491)

       at org.apache.phoenix.util.PhoenixContextExecutor.call(PhoenixContextExecutor.java:76)

       at org.apache.phoenix.query.ConnectionQueryServicesImpl.init(ConnectionQueryServicesImpl.java:2491)

       at org.apache.phoenix.jdbc.PhoenixDriver.getConnectionQueryServices(PhoenixDriver.java:255)

       at org.apache.phoenix.jdbc.PhoenixEmbeddedDriver.createConnection(PhoenixEmbeddedDriver.java:150)

       at org.apache.phoenix.jdbc.PhoenixDriver.connect(PhoenixDriver.java:221)

        at
sqlline.DatabaseConnection.connect(DatabaseConnection.java:157)

        at sqlline.DatabaseConnection.getConnection(DatabaseConnection.java:203)

       at sqlline.Commands.connect(Commands.java:1064)

        at sqlline.Commands.connect(Commands.java:996)

       at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)

        at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)

       at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)

       at java.lang.reflect.Method.invoke(Method.java:498)

        at sqlline.ReflectiveCommandHandler.execute(ReflectiveCommandHandler.java:38)

       at sqlline.SqlLine.dispatch(SqlLine.java:809)

        at sqlline.SqlLine.initArgs(SqlLine.java:588)

       at sqlline.SqlLine.begin(SqlLine.java:661)

        at sqlline.SqlLine.start(SqlLine.java:398)

       at sqlline.SqlLine.main(SqlLine.java:291)

Caused by: org.apache.hadoop.hbase.DoNotRetryIOException:
org.apache.hadoop.hbase.DoNotRetryIOException: Unable to load configured region split policy
'org.apache.phoenix.schema.MetaDataSplitPolicy' for table 'SYSTEM.CATALOG' Set hbase.table.sanity.checks
to false at conf or table descriptor if you want to bypass sanity checks




      I searched
from internet,but got no help.

      Any help will be highly appreciated!






Mime
View raw message