phoenix-user mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From F21 <f21.gro...@gmail.com>
Subject Re: 回复: 回复: 回复: Can query server run with hadoop ha mode?
Date Thu, 08 Sep 2016 09:56:45 GMT
Glad you got it working! :)

Cheers,
Francis

On 8/09/2016 7:11 PM, zengbaitang wrote:
>
> I found the reason ,  because i have not set the env : HADOOP_CONF_DIR 
> while i set the env, the problem solved .
> Thank you F21, thank you very much!
>
>
> ------------------ 原始邮件 ------------------
> *发件人:* "F21";<f21.groups@gmail.com>;
> *发送时间:* 2016年9月8日(星期四) 下午3:33
> *收件人:* "user"<user@phoenix.apache.org>;
> *主题:* Re: 回复: 回复: Can query server run with hadoop ha mode?
>
> From the response of your curl, it appears that the query server is 
> started correctly and running. The next bit to check is to see if it 
> can talk to the HBase servers properly.
>
> Add phoenix.queryserver.serialization to the hbase-site.xml for the 
> query server and set the value to JSON.
>
> Then try and send a CatalogsRequest to the query server using curl or 
> wget.
> See here for how to set up the request 
> https://calcite.apache.org/docs/avatica_json_reference.html#catalogsrequest
>
> Before sending the CatalogsRequest, remember to send an 
> OpenConnectionRequest first: 
> https://calcite.apache.org/docs/avatica_json_reference.html#openconnectionrequest
>
> In your case, the `info` key of the OpenConnectionRequest can be omitted.
>
> Cheers,
> Francis
>
> On 8/09/2016 4:12 PM, zengbaitang wrote:
>> yes, the query server run on one of the regionservers
>>
>> and exec curl 'http://tnode02:8765' the terminal returns :
>> <HTML>
>> <HEAD>
>> <TITLE>Error 404 - Not Found</TITLE>
>> <BODY>
>> <H2>Error 404 - Not Found.</H2>
>> No context on this server matched or handled this 
>> request.<BR>Contexts known to this server are: <ul></ul><hr><a

>> href="http://eclipse.org/jetty"><img border=0 
>> src="/favicon.ico"/></a>&nbsp;<a 
>> href="http://eclipse.org/jetty">Powered by Jetty:// Java Web 
>> Server</a><hr/>
>>
>> </BODY>
>> </HTML>
>>
>>
>>
>> ------------------ 原始邮件 ------------------
>> *发件人:* "F21";<f21.groups@gmail.com>;
>> *发送时间:* 2016年9月8日(星期四) 下午2:01
>> *收件人:* "user"<user@phoenix.apache.org>;
>> *主题:* Re: 回复: Can query server run with hadoop ha mode?
>>
>> Your logs do not seem to show any errors.
>>
>> You mentioned that you have 2 hbase-site.xml. Are the Phoenix query 
>> servers running on the same machine as the HBase servers? If not, the 
>> hbase-site.xml for the phoenix query servers also needs the zookeeper 
>> configuration.
>>
>> Did you also try to use curl or wget to get 
>> http://your-phoenix-query-server:8765 to see if there's a response?
>>
>> Cheers,
>> Francis
>>
>> On 8/09/2016 3:54 PM, zengbaitang wrote:
>>> hi F21 ,  I am sure hbase-site.xml was configured properly ,
>>>
>>> here is my *hbase-site.xml (hbase side)*:
>>> <configuration>
>>>     <property>
>>>         <name>hbase.rootdir</name>
>>> <value>hdfs://stage-cluster/hbase</value>
>>>     </property>
>>>
>>>     <property>
>>> <name>hbase.cluster.distributed</name>
>>>         <value>true</value>
>>>     </property>
>>>     <property>
>>> <name>hbase.zookeeper.quorum</name>
>>> <value>tnode01,tnode02,tnode03</value>
>>>     </property>
>>>     <property>
>>> <name>zookeeper.znode.parent</name>
>>>         <value>/hbase</value>
>>>     </property>
>>>     <property>
>>>         <name>dfs.support.append</name>
>>>         <value>true</value>
>>>     </property>
>>>     <property>
>>> <name>zookeeper.session.timeout</name>
>>>         <value>180000</value>
>>>     </property>
>>>     <property>
>>>         <name>hbase.rpc.timeout</name>
>>>         <value>120000</value>
>>>     </property>
>>>     <property>
>>> <name>hbase.hregion.memstore.flush.size</name>
>>>         <value>67108864</value>
>>>     </property>
>>>     <property>
>>> <name>hfile.block.cache.size</name>
>>>         <value>0.1</value>
>>>     </property>
>>>
>>>     <!-- phoenix conf -->
>>>     <property>
>>> <name>phoenix.schema.isNamespaceMappingEnabled</name>
>>>         <value>true</value>
>>>     </property>

>>>
>>>     <property>
>>> <name>hbase.regionserver.wal.codec</name>
>>> <value>org.apache.hadoop.hbase.regionserver.wal.IndexedWALEditCodec</value>
>>>     </property>
>>>
>>>     <property>
>>> <name>hbase.region.server.rpc.scheduler.factory.class</name>
>>> <value>org.apache.hadoop.hbase.ipc.PhoenixRpcSchedulerFactory</value>
>>>         <description>Factory to create the Phoenix RPC Scheduler 
>>> that uses separate queues for index and metadata updates</description>
>>>     </property>
>>>
>>>     <property>
>>> <name>hbase.rpc.controllerfactory.class</name>
>>> <value>org.apache.hadoop.hbase.ipc.controller.ServerRpcControllerFactory</value>
>>>         <description>Factory to create the Phoenix RPC Scheduler 
>>> that uses separate queues for index and metadata updates</description>
>>>     </property>
>>>
>>>
>>> </configuration>
>>>
>>> *and the following is phoenix side hbase-site.xml*
>>> <configuration>
>>>   <property>
>>> <name>hbase.regionserver.wal.codec</name>
>>> <value>org.apache.hadoop.hbase.regionserver.wal.IndexedWALEditCodec</value>
>>>   </property>
>>>
>>>   <property>
>>> <name>phoenix.schema.isNamespaceMappingEnabled</name>
>>>     <value>true</value>
>>>   </property>
>>>
>>> </configuration>
>>>
>>> *and the following is query server log*
>>> *
>>> *
>>> 2016-09-08 13:33:03,218 INFO 
>>> org.apache.phoenix.queryserver.server.Main: 
>>> env:PATH=/usr/lib64/qt-3.3/bin:/usr/local/bin:/bin:/usr/bin:/usr/local/sbin:/usr/sbin:/sbin:/home/hadoop/bin:/usr/local/hadoop-2.7.1/bin:/usr/local/hbase-1.1.2/bin:/usr/local/apache-hive-1.2.1-bin/bin:/usr/local/sqoop-1.4.6.bin__hadoop-2.0.4-alpha/bin
>>> 2016-09-08 13:33:03,219 INFO 
>>> org.apache.phoenix.queryserver.server.Main: env:HISTCONTROL=ignoredups
>>> 2016-09-08 13:33:03,219 INFO 
>>> org.apache.phoenix.queryserver.server.Main: 
>>> env:HCAT_HOME=/usr/local/apache-hive-1.2.1-bin/hcatalog
>>> 2016-09-08 13:33:03,220 INFO 
>>> org.apache.phoenix.queryserver.server.Main: env:HISTSIZE=1000
>>> 2016-09-08 13:33:03,220 INFO 
>>> org.apache.phoenix.queryserver.server.Main: 
>>> env:JAVA_HOME=/usr/local/java/latest
>>> 2016-09-08 13:33:03,220 INFO 
>>> org.apache.phoenix.queryserver.server.Main: env:TERM=xterm
>>> 2016-09-08 13:33:03,220 INFO 
>>> org.apache.phoenix.queryserver.server.Main: 
>>> env:XFILESEARCHPATH=/usr/dt/app-defaults/%L/Dt
>>> 2016-09-08 13:33:03,220 INFO 
>>> org.apache.phoenix.queryserver.server.Main: env:LANG=en_US.UTF-8
>>> 2016-09-08 13:33:03,220 INFO 
>>> org.apache.phoenix.queryserver.server.Main: env:G_BROKEN_FILENAMES=1
>>> 2016-09-08 13:33:03,220 INFO 
>>> org.apache.phoenix.queryserver.server.Main: env:SELINUX_LEVEL_REQUESTED=
>>> 2016-09-08 13:33:03,221 INFO 
>>> org.apache.phoenix.queryserver.server.Main: env:SELINUX_ROLE_REQUESTED=
>>> 2016-09-08 13:33:03,221 INFO 
>>> org.apache.phoenix.queryserver.server.Main: 
>>> env:MAIL=/var/spool/mail/hadoop
>>> 2016-09-08 13:33:03,221 INFO 
>>> org.apache.phoenix.queryserver.server.Main: env:LOGNAME=hadoop
>>> 2016-09-08 13:33:03,221 INFO 
>>> org.apache.phoenix.queryserver.server.Main: 
>>> env:PWD=/usr/local/apache-phoenix-4.8.0-HBase-1.1-bin/bin
>>> 2016-09-08 13:33:03,221 INFO 
>>> org.apache.phoenix.queryserver.server.Main: 
>>> env:KYLIN_HOME=/usr/local/apache-kylin-1.5.1-bin
>>> 2016-09-08 13:33:03,221 INFO 
>>> org.apache.phoenix.queryserver.server.Main: env:_=./queryserver.py
>>> 2016-09-08 13:33:03,221 INFO 
>>> org.apache.phoenix.queryserver.server.Main: 
>>> env:LESSOPEN=|/usr/bin/lesspipe.sh %s
>>> 2016-09-08 13:33:03,222 INFO 
>>> org.apache.phoenix.queryserver.server.Main: env:SHELL=/bin/bash
>>> 2016-09-08 13:33:03,222 INFO 
>>> org.apache.phoenix.queryserver.server.Main: 
>>> env:SELINUX_USE_CURRENT_RANGE=
>>> 2016-09-08 13:33:03,222 INFO 
>>> org.apache.phoenix.queryserver.server.Main: 
>>> env:QTINC=/usr/lib64/qt-3.3/include
>>> 2016-09-08 13:33:03,222 INFO 
>>> org.apache.phoenix.queryserver.server.Main: env:CVS_RSH=ssh
>>> 2016-09-08 13:33:03,222 INFO 
>>> org.apache.phoenix.queryserver.server.Main: env:SSH_TTY=/dev/pts/0
>>> 2016-09-08 13:33:03,222 INFO 
>>> org.apache.phoenix.queryserver.server.Main: 
>>> env:SSH_CLIENT=172.18.100.27 51441 22
>>> 2016-09-08 13:33:03,223 INFO 
>>> org.apache.phoenix.queryserver.server.Main: 
>>> env:HIVE_HOME=/usr/local/apache-hive-1.2.1-bin
>>> 2016-09-08 13:33:03,223 INFO 
>>> org.apache.phoenix.queryserver.server.Main: 
>>> env:OLDPWD=/usr/local/hadoop-2.7.1/etc/hadoop
>>> 2016-09-08 13:33:03,223 INFO 
>>> org.apache.phoenix.queryserver.server.Main: env:USER=hadoop
>>> 2016-09-08 13:33:03,223 INFO 
>>> org.apache.phoenix.queryserver.server.Main: 
>>> env:SSH_ASKPASS=/usr/libexec/openssh/gnome-ssh-askpass
>>> 2016-09-08 13:33:03,223 INFO 
>>> org.apache.phoenix.queryserver.server.Main: 
>>> env:SSH_CONNECTION=172.18.100.27 51441 172.23.201.49 22
>>> 2016-09-08 13:33:03,223 INFO 
>>> org.apache.phoenix.queryserver.server.Main: env:HOSTNAME=tnode02
>>> 2016-09-08 13:33:03,223 INFO 
>>> org.apache.phoenix.queryserver.server.Main: env:QTDIR=/usr/lib64/qt-3.3
>>> 2016-09-08 13:33:03,224 INFO 
>>> org.apache.phoenix.queryserver.server.Main: 
>>> env:NLSPATH=/usr/dt/lib/nls/msg/%L/%N.cat
>>> 2016-09-08 13:33:03,224 INFO 
>>> org.apache.phoenix.queryserver.server.Main: 
>>> env:HADOOP_HOME=/usr/local/hadoop-2.7.1
>>> 2016-09-08 13:33:03,224 INFO 
>>> org.apache.phoenix.queryserver.server.Main: 
>>> env:HBASE_HOME=/usr/local/hbase-1.1.2
>>> 2016-09-08 13:33:03,224 INFO 
>>> org.apache.phoenix.queryserver.server.Main: 
>>> env:LS_COLORS=rs=0:di=01;34:ln=01;36:mh=00:pi=40;33:so=01;35:do=01;35:bd=40;33;01:cd=40;33;01:or=40;31;01:mi=01;05;37;41:su=37;41:sg=30;43:ca=30;41:tw=30;42:ow=34;42:st=37;44:ex=01;32:*.tar=01;31:*.tgz=01;31:*.arj=01;31:*.taz=01;31:*.lzh=01;31:*.lzma=01;31:*.tlz=01;31:*.txz=01;31:*.zip=01;31:*.z=01;31:*.Z=01;31:*.dz=01;31:*.gz=01;31:*.lz=01;31:*.xz=01;31:*.bz2=01;31:*.tbz=01;31:*.tbz2=01;31:*.bz=01;31:*.tz=01;31:*.deb=01;31:*.rpm=01;31:*.jar=01;31:*.rar=01;31:*.ace=01;31:*.zoo=01;31:*.cpio=01;31:*.7z=01;31:*.rz=01;31:*.jpg=01;35:*.jpeg=01;35:*.gif=01;35:*.bmp=01;35:*.pbm=01;35:*.pgm=01;35:*.ppm=01;35:*.tga=01;35:*.xbm=01;35:*.xpm=01;35:*.tif=01;35:*.tiff=01;35:*.png=01;35:*.svg=01;35:*.svgz=01;35:*.mng=01;35:*.pcx=01;35:*.mov=01;35:*.mpg=01;35:*.mpeg=01;35:*.m2v=01;35:*.mkv=01;35:*.ogm=01;35:*.mp4=01;35:*.m4v=01;35:*.mp4v=01;35:*.vob=01;35:*.qt=01;35:*.nuv=01;35:*.wmv=01;35:*.asf=01;35:*.rm=01;35:*.rmvb=01;35:*.flc=01;35:*.avi=01;35:*.fli=01;35:*.flv=01;35:*.gl=01;35:*.dl=01;35:*.xcf=01;35:*.xwd=01;35:*.yuv=01;35:*.cgm=01;35:*.emf=01;35:*.axv=01;35:*.anx=01;35:*.ogv=01;35:*.ogx=01;35:*.aac=01;36:*.au=01;36:*.flac=01;36:*.mid=01;36:*.midi=01;36:*.mka=01;36:*.mp3=01;36:*.mpc=01;36:*.ogg=01;36:*.ra=01;36:*.wav=01;36:*.axa=01;36:*.oga=01;36:*.spx=01;36:*.xspf=01;36:
>>> 2016-09-08 13:33:03,225 INFO 
>>> org.apache.phoenix.queryserver.server.Main: 
>>> env:QTLIB=/usr/lib64/qt-3.3/lib
>>> 2016-09-08 13:33:03,225 INFO 
>>> org.apache.phoenix.queryserver.server.Main: env:HOME=/home/hadoop
>>> 2016-09-08 13:33:03,225 INFO 
>>> org.apache.phoenix.queryserver.server.Main: env:SHLVL=1
>>> 2016-09-08 13:33:03,225 INFO 
>>> org.apache.phoenix.queryserver.server.Main: 
>>> env:ZOOKEEPER_HOME=/usr/local/zookeeper-3.4.8
>>> 2016-09-08 13:33:03,228 INFO 
>>> org.apache.phoenix.queryserver.server.Main: vmName=Java HotSpot(TM) 
>>> 64-Bit Server VM, vmVendor=Oracle Corporation, vmVersion=25.65-b01
>>> 2016-09-08 13:33:03,229 INFO 
>>> org.apache.phoenix.queryserver.server.Main: 
>>> vmInputArguments=[-Dproc_phoenixserver, 
>>> -Dlog4j.configuration=file:/usr/local/apache-phoenix-4.8.0-HBase-1.1-bin/bin/log4j.properties,

>>> -Dpsql.root.logger=INFO,DRFA, -Dpsql.log.dir=/var/log/hbase/logs, 
>>> -Dpsql.log.file=hadoop-queryserver.log]
>>> 2016-09-08 13:33:03,444 WARN 
>>> org.apache.hadoop.util.NativeCodeLoader: Unable to load 
>>> native-hadoop library for your platform... using builtin-java 
>>> classes where applicable
>>> 2016-09-08 13:33:03,709 INFO 
>>> org.apache.calcite.avatica.metrics.MetricsSystemLoader: No metrics 
>>> implementation available on classpath. Using No-op implementation
>>> 2016-09-08 13:33:03,736 INFO 
>>> org.apache.phoenix.shaded.org.eclipse.jetty.util.log: Logging 
>>> initialized @1458ms
>>> 2016-09-08 13:33:04,129 INFO 
>>> org.apache.phoenix.shaded.org.eclipse.jetty.server.Server: 
>>> jetty-9.2.z-SNAPSHOT
>>> 2016-09-08 13:33:04,194 INFO 
>>> org.apache.phoenix.shaded.org.eclipse.jetty.server.ServerConnector: 
>>> Started ServerConnector@131ef10{HTTP/1.1}{0.0.0.0:8765}
>>> 2016-09-08 13:33:04,195 INFO 
>>> org.apache.phoenix.shaded.org.eclipse.jetty.server.Server: Started 
>>> @1922ms
>>> 2016-09-08 13:33:04,195 INFO 
>>> org.apache.calcite.avatica.server.HttpServer: Service listening on 
>>> port 8765.
>>> 2016-09-08 13:33:36,903 INFO 
>>> org.apache.phoenix.shaded.org.apache.zookeeper.ZooKeeper: Client 
>>> environment:zookeeper.version=3.4.6-1569965, built on 02/20/2014 
>>> 09:09 GMT
>>> 2016-09-08 13:33:36,903 INFO 
>>> org.apache.phoenix.shaded.org.apache.zookeeper.ZooKeeper: Client 
>>> environment:host.name=tnode02
>>> 2016-09-08 13:33:36,903 INFO 
>>> org.apache.phoenix.shaded.org.apache.zookeeper.ZooKeeper: Client 
>>> environment:java.version=1.8.0_65
>>> 2016-09-08 13:33:36,903 INFO 
>>> org.apache.phoenix.shaded.org.apache.zookeeper.ZooKeeper: Client 
>>> environment:java.vendor=Oracle Corporation
>>> 2016-09-08 13:33:36,903 INFO 
>>> org.apache.phoenix.shaded.org.apache.zookeeper.ZooKeeper: Client 
>>> environment:java.home=/usr/local/java/jdk1.8.0_65/jre
>>> 2016-09-08 13:33:36,903 INFO 
>>> org.apache.phoenix.shaded.org.apache.zookeeper.ZooKeeper: Client 
>>> environment:java.class.path=/usr/local/hbase-1.1.2/conf:/etc/hadoop/conf:/usr/local/apache-phoenix-4.8.0-HBase-1.1-bin/bin/../phoenix-4.8.0-HBase-1.1-client.jar:/usr/local/apache-phoenix-4.8.0-HBase-1.1-bin/bin/../phoenix-4.8.0-HBase-1.1-queryserver.jar
>>> 2016-09-08 13:33:36,903 INFO 
>>> org.apache.phoenix.shaded.org.apache.zookeeper.ZooKeeper: Client 
>>> environment:java.library.path=/usr/java/packages/lib/amd64:/usr/lib64:/lib64:/lib:/usr/lib
>>> 2016-09-08 13:33:36,903 INFO 
>>> org.apache.phoenix.shaded.org.apache.zookeeper.ZooKeeper: Client 
>>> environment:java.io.tmpdir=/tmp
>>> 2016-09-08 13:33:36,903 INFO 
>>> org.apache.phoenix.shaded.org.apache.zookeeper.ZooKeeper: Client 
>>> environment:java.compiler=<NA>
>>> 2016-09-08 13:33:36,903 INFO 
>>> org.apache.phoenix.shaded.org.apache.zookeeper.ZooKeeper: Client 
>>> environment:os.name=Linux
>>> 2016-09-08 13:33:36,903 INFO 
>>> org.apache.phoenix.shaded.org.apache.zookeeper.ZooKeeper: Client 
>>> environment:os.arch=amd64
>>> 2016-09-08 13:33:36,903 INFO 
>>> org.apache.phoenix.shaded.org.apache.zookeeper.ZooKeeper: Client 
>>> environment:os.version=2.6.32-431.el6.x86_64
>>> 2016-09-08 13:33:36,903 INFO 
>>> org.apache.phoenix.shaded.org.apache.zookeeper.ZooKeeper: Client 
>>> environment:user.name=hadoop
>>> 2016-09-08 13:33:36,904 INFO 
>>> org.apache.phoenix.shaded.org.apache.zookeeper.ZooKeeper: Client 
>>> environment:user.home=/home/hadoop
>>> 2016-09-08 13:33:36,904 INFO 
>>> org.apache.phoenix.shaded.org.apache.zookeeper.ZooKeeper: Client 
>>> environment:user.dir=/
>>> 2016-09-08 13:33:36,904 INFO 
>>> org.apache.phoenix.shaded.org.apache.zookeeper.ZooKeeper: Initiating 
>>> client connection, 
>>> connectString=tnode01:2181,tnode02:2181,tnode03:2181 
>>> sessionTimeout=180000 watcher=hconnection-0x47e590f30x0, 
>>> quorum=tnode01:2181,tnode02:2181,tnode03:2181, baseZNode=/hbase
>>> 2016-09-08 13:33:36,925 INFO 
>>> org.apache.phoenix.shaded.org.apache.zookeeper.ClientCnxn: Opening 
>>> socket connection to server tnode02/172.23.201.49:2181. Will not 
>>> attempt to authenticate using SASL (unknown error)
>>> 2016-09-08 13:33:36,927 INFO 
>>> org.apache.phoenix.shaded.org.apache.zookeeper.ClientCnxn: Socket 
>>> connection established to tnode02/172.23.201.49:2181, initiating session
>>> 2016-09-08 13:33:36,950 INFO 
>>> org.apache.phoenix.shaded.org.apache.zookeeper.ClientCnxn: Session 
>>> establishment complete on server tnode02/172.23.201.49:2181, 
>>> sessionid = 0x25702951a6e001b, negotiated timeout = 40000
>>>
>>>
>>>
>>> ------------------ 原始邮件 ------------------
>>> *发件人:* "F21";<f21.groups@gmail.com>;
>>> *发送时间:* 2016年9月8日(星期四) 中午11:58
>>> *收件人:* "user"<user@phoenix.apache.org>;
>>> *主题:* Re: Can query server run with hadoop ha mode?
>>>
>>> I have a test cluster running HDFS in HA mode with HBase + Phoenix 
>>> on docker running successfully.
>>>
>>> Can you check if you have a properly configured hbase-site.xml that 
>>> is available to your phoenix query server? Make sure 
>>> hbase.zookeeper.quorum and zookeeper.znode.parent is present. If 
>>> zookeeper does not run on 2181, you will also need 
>>> hbase.zookeeper.property.clientPort.
>>>
>>> As a quick test, can you wget or curl 
>>> http://your-phoenix-server:8765 to see if it has any response? 
>>> Finally, if you could post the logs from the query server, that 
>>> would be great too.
>>>
>>> Cheers,
>>> Francis
>>>
>>>
>>> On 8/09/2016 12:55 PM, zengbaitang wrote:
>>>> I have a hadoop ha cluster and hbase, and  have installed phoenix.
>>>>
>>>> I try to use query server today , I start the queryserver and then 
>>>> I exec the following command
>>>>
>>>> ./sqlline-thin.py http://tnode02:8765 sel.sql
>>>>
>>>> the terminal responds the following error , and the *stage-cluster* 
>>>> is the value of  hadoop dfs.nameservices ,
>>>> how to solve this error?
>>>>
>>>> AvaticaClientRuntimeException: Remote driver error: 
>>>> RuntimeException: java.sql.SQLException: ERROR 103 (08004): Unable 
>>>> to establish connection. -> SQLException: ERROR 103 (08004): Unable 
>>>> to establish connection. -> IOException: 
>>>> java.lang.reflect.InvocationTargetException -> 
>>>> InvocationTargetException: (null exception message) -> 
>>>> ExceptionInInitializerError: (null exception message) -> 
>>>> IllegalArgumentException: java.net.UnknownHostException: 
>>>> stage-cluster -> UnknownHostException: stage-cluster. Error -1 
>>>> (00000) null
>>>>
>>>> java.lang.RuntimeException: java.sql.SQLException: ERROR 103 
>>>> (08004): Unable to establish connection.
>>>>         at 
>>>> org.apache.calcite.avatica.jdbc.JdbcMeta.openConnection(JdbcMeta.java:619)
>>>>         at 
>>>> org.apache.calcite.avatica.remote.LocalService.apply(LocalService.java:299)
>>>>         at 
>>>> org.apache.calcite.avatica.remote.Service$OpenConnectionRequest.accept(Service.java:1748)
>>>>         at 
>>>> org.apache.calcite.avatica.remote.Service$OpenConnectionRequest.accept(Service.java:1728)
>>>>         at 
>>>> org.apache.calcite.avatica.remote.AbstractHandler.apply(AbstractHandler.java:95)
>>>>         at 
>>>> org.apache.calcite.avatica.remote.ProtobufHandler.apply(ProtobufHandler.java:46)
>>>>         at 
>>>> org.apache.calcite.avatica.server.AvaticaProtobufHandler.handle(AvaticaProtobufHandler.java:124)
>>>>         at 
>>>> org.apache.phoenix.shaded.org.eclipse.jetty.server.handler.HandlerList.handle(HandlerList.java:52)
>>>>         at 
>>>> org.apache.phoenix.shaded.org.eclipse.jetty.server.handler.HandlerWrapper.handle(HandlerWrapper.java:97)
>>>>         at 
>>>> org.apache.phoenix.shaded.org.eclipse.jetty.server.Server.handle(Server.java:499)
>>>>         at 
>>>> org.apache.phoenix.shaded.org.eclipse.jetty.server.HttpChannel.handle(HttpChannel.java:311)
>>>>         at 
>>>> org.apache.phoenix.shaded.org.eclipse.jetty.server.HttpConnection.onFillable(HttpConnection.java:257)
>>>>         at 
>>>> org.apache.phoenix.shaded.org.eclipse.jetty.io.AbstractConnection$2.run(AbstractConnection.java:544)
>>>>         at 
>>>> org.apache.phoenix.shaded.org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:635)
>>>>         at 
>>>> org.apache.phoenix.shaded.org.eclipse.jetty.util.thread.QueuedThreadPool$3.run(QueuedThreadPool.java:555)
>>>>         at java.lang.Thread.run(Thread.java:745)
>>>> Caused by: java.sql.SQLException: ERROR 103 (08004): Unable to 
>>>> establish connection.
>>>>         at 
>>>> org.apache.phoenix.exception.SQLExceptionCode$Factory$1.newException(SQLExceptionCode.java:454)
>>>>         at 
>>>> org.apache.phoenix.exception.SQLExceptionInfo.buildException(SQLExceptionInfo.java:145)
>>>>         at 
>>>> org.apache.phoenix.query.ConnectionQueryServicesImpl.openConnection(ConnectionQueryServicesImpl.java:393)
>>>>         at 
>>>> org.apache.phoenix.query.ConnectionQueryServicesImpl.access$300(ConnectionQueryServicesImpl.java:219)
>>>>         at 
>>>> org.apache.phoenix.query.ConnectionQueryServicesImpl$13.call(ConnectionQueryServicesImpl.java:2321)
>>>>         at 
>>>> org.apache.phoenix.query.ConnectionQueryServicesImpl$13.call(ConnectionQueryServicesImpl.java:2300)
>>>>         at 
>>>> org.apache.phoenix.util.PhoenixContextExecutor.call(PhoenixContextExecutor.java:78)
>>>>         at 
>>>> org.apache.phoenix.query.ConnectionQueryServicesImpl.init(ConnectionQueryServicesImpl.java:2300)
>>>>         at 
>>>> org.apache.phoenix.jdbc.PhoenixDriver.getConnectionQueryServices(PhoenixDriver.java:231)
>>>>         at 
>>>> org.apache.phoenix.jdbc.PhoenixEmbeddedDriver.createConnection(PhoenixEmbeddedDriver.java:144)
>>>>         at 
>>>> org.apache.phoenix.jdbc.PhoenixDriver.connect(PhoenixDriver.java:202)
>>>>         at java.sql.DriverManager.getConnection(DriverManager.java:664)
>>>>         at java.sql.DriverManager.getConnection(DriverManager.java:208)
>>>>         at 
>>>> org.apache.calcite.avatica.jdbc.JdbcMeta.openConnection(JdbcMeta.java:616)
>>>>         ... 15 more
>>>> Caused by: java.io.IOException: 
>>>> java.lang.reflect.InvocationTargetException
>>>>         at 
>>>> org.apache.hadoop.hbase.client.ConnectionFactory.createConnection(ConnectionFactory.java:240)
>>>>         at 
>>>> org.apache.hadoop.hbase.client.ConnectionManager.createConnection(ConnectionManager.java:421)
>>>>         at 
>>>> org.apache.hadoop.hbase.client.ConnectionManager.createConnectionInternal(ConnectionManager.java:330)
>>>>         at 
>>>> org.apache.hadoop.hbase.client.HConnectionManager.createConnection(HConnectionManager.java:144)
>>>>         at 
>>>> org.apache.phoenix.query.HConnectionFactory$HConnectionFactoryImpl.createConnection(HConnectionFactory.java:47)
>>>>         at 
>>>> org.apache.phoenix.query.ConnectionQueryServicesImpl.openConnection(ConnectionQueryServicesImpl.java:391)
>>>>         ... 26 more
>>>> Caused by: java.lang.reflect.InvocationTargetException
>>>>         at 
>>>> sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
>>>>         at 
>>>> sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)
>>>>         at 
>>>> sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
>>>>         at 
>>>> java.lang.reflect.Constructor.newInstance(Constructor.java:422)
>>>>         at 
>>>> org.apache.hadoop.hbase.client.ConnectionFactory.createConnection(ConnectionFactory.java:238)
>>>>         ... 31 more
>>>> Caused by: java.lang.ExceptionInInitializerError
>>>>         at 
>>>> org.apache.hadoop.hbase.ClusterId.parseFrom(ClusterId.java:64)
>>>>         at 
>>>> org.apache.hadoop.hbase.zookeeper.ZKClusterId.readClusterIdZNode(ZKClusterId.java:75)
>>>>         at 
>>>> org.apache.hadoop.hbase.client.ZooKeeperRegistry.getClusterId(ZooKeeperRegistry.java:105)
>>>>         at 
>>>> org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.retrieveClusterId(ConnectionManager.java:880)
>>>>         at 
>>>> org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.<init>(ConnectionManager.java:636)
>>>>         ... 36 more
>>>> Caused by: java.lang.IllegalArgumentException: 
>>>> java.net.UnknownHostException: stage-cluster
>>>>         at 
>>>> org.apache.hadoop.security.SecurityUtil.buildTokenService(SecurityUtil.java:378)
>>>>         at 
>>>> org.apache.hadoop.hdfs.NameNodeProxies.createNonHAProxy(NameNodeProxies.java:310)
>>>>         at 
>>>> org.apache.hadoop.hdfs.NameNodeProxies.createProxy(NameNodeProxies.java:176)
>>>>         at org.apache.hadoop.hdfs.DFSClient.<init>(DFSClient.java:678)
>>>>         at org.apache.hadoop.hdfs.DFSClient.<init>(DFSClient.java:619)
>>>>         at 
>>>> org.apache.hadoop.hdfs.DistributedFileSystem.initialize(DistributedFileSystem.java:149)
>>>>         at 
>>>> org.apache.hadoop.fs.FileSystem.createFileSystem(FileSystem.java:2653)
>>>>         at 
>>>> org.apache.hadoop.fs.FileSystem.access$200(FileSystem.java:92)
>>>>         at 
>>>> org.apache.hadoop.fs.FileSystem$Cache.getInternal(FileSystem.java:2687)
>>>>         at 
>>>> org.apache.hadoop.fs.FileSystem$Cache.get(FileSystem.java:2669)
>>>>         at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:371)
>>>>         at org.apache.hadoop.fs.Path.getFileSystem(Path.java:295)
>>>>         at 
>>>> org.apache.hadoop.hbase.util.DynamicClassLoader.initTempDir(DynamicClassLoader.java:118)
>>>>         at 
>>>> org.apache.hadoop.hbase.util.DynamicClassLoader.<init>(DynamicClassLoader.java:98)
>>>>         at 
>>>> org.apache.hadoop.hbase.protobuf.ProtobufUtil.<clinit>(ProtobufUtil.java:241)
>>>>         ... 41 more
>>>> Caused by: java.net.UnknownHostException: *stage-cluster*
>>>>         ... 56 more
>>>>
>>>>
>>>>         at 
>>>> org.apache.calcite.avatica.remote.Service$ErrorResponse.toException(Service.java:2453)
>>>>         at 
>>>> org.apache.calcite.avatica.remote.RemoteProtobufService._apply(RemoteProtobufService.java:61)
>>>>         at 
>>>> org.apache.calcite.avatica.remote.ProtobufService.apply(ProtobufService.java:81)
>>>>         at 
>>>> org.apache.calcite.avatica.remote.Driver.connect(Driver.java:175)
>>>>         at 
>>>> sqlline.DatabaseConnection.connect(DatabaseConnection.java:157)
>>>>         at 
>>>> sqlline.DatabaseConnection.getConnection(DatabaseConnection.java:203)
>>>>         at sqlline.Commands.connect(Commands.java:1064)
>>>>         at sqlline.Commands.connect(Commands.java:996)
>>>>         at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>>>>         at 
>>>> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
>>>>         at 
>>>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>>>>         at java.lang.reflect.Method.invoke(Method.java:497)
>>>>         at 
>>>> sqlline.ReflectiveCommandHandler.execute(ReflectiveCommandHandler.java:36)
>>>>         at sqlline.SqlLine.dispatch(SqlLine.java:803)
>>>>         at sqlline.SqlLine.initArgs(SqlLine.java:588)
>>>>         at sqlline.SqlLine.begin(SqlLine.java:656)
>>>>         at sqlline.SqlLine.start(SqlLine.java:398)
>>>>         at sqlline.SqlLine.main(SqlLine.java:292)
>>>>         at 
>>>> org.apache.phoenix.queryserver.client.SqllineWrapper.main(SqllineWrapper.java:83)
>>>>
>>>>
>>>
>>
>


Mime
View raw message