I didn't see Phoenix as option anywhere in Ambari and I found this http://docs.hortonworks.com/HDPDocuments/HDP2/HDP-2.1.3/bk_installing_manually_book/content/rpm-chap-phoenix.html

Did I miss something in Ambari for Phoenix?


On Mon, Sep 1, 2014 at 11:01 PM, James Taylor <jamestaylor@apache.org> wrote:
Hi Vikas,
Glad you got it working. Just curious - why did you install Phoenix
via yum when the HDP 2.1 already comes pre-installed with Phoenix?
Thanks,
James

On Mon, Sep 1, 2014 at 10:16 AM, Vikas Agarwal <vikas@infoobjects.com> wrote:
> Yes, I am using HDP 2.1 and installed Phoenix via yum and it installed 4.0.0
> of phoenix. Added symlink to the phoenix-core*jar into /usr/lib/hbase/lib
> and I did restart after these changes.
>
> However, now, I am able to connect with HBase. Somehow, SYSTEM.CATALOG was
> got created in one of my earlier attempts to connect phoenix to HBase and it
> didn't contained the required region server information for phoenix to work
> on. So, I deleted this table and hope phoenix to recreate it and it did and
> connection also worked after that. :)
>
>
> On Mon, Sep 1, 2014 at 6:54 PM, Nicolas Maillard <nmaillard@hortonworks.com>
> wrote:
>>
>> Hello
>> Just to be sure you are using HDP 2.1 and you gotten the phoenix jars from
>> the page you listed and have put the phoenix jar in all on the hbase nodes
>> in the lib directory and restarted the whole hbase service.
>> If so could you also paste the line you use to start sqlline.
>>
>>
>> On Mon, Sep 1, 2014 at 12:43 PM, Vikas Agarwal <vikas@infoobjects.com>
>> wrote:
>>>
>>> Hi,
>>>
>>> We have installed Hadoop cluster using Hortonworks distribution and
>>> trying to connect the Phoenix with HBase. However, even after following the
>>> steps mentioned here, we are not able to connect Phoenix to HBase.
>>>
>>> When I am trying to test the connection using sqlline.py, the command
>>> hangs the control and I am even not able to do Ctrl-C or Ctrl-Z. After
>>> sometime (around 10 min) it throws same exception as with psql.py,
>>> complaining mismatch of phoenix jars.
>>>
>>> When I am trying to test the connection using Phoenix's psql.py command,
>>> following set of exceptions are coming:
>>>
>>> 14/09/01 05:41:25 WARN util.NativeCodeLoader: Unable to load
>>> native-hadoop library for your platform... using builtin-java classes where
>>> applicable
>>> 14/09/01 05:41:26 WARN util.DynamicClassLoader: Failed to identify the fs
>>> of dir hdfs://hdp.ambari:8020/apps/hbase/data/lib, ignored
>>> java.io.IOException: No FileSystem for scheme: hdfs
>>> at
>>> org.apache.hadoop.fs.FileSystem.getFileSystemClass(FileSystem.java:2385)
>>> at org.apache.hadoop.fs.FileSystem.createFileSystem(FileSystem.java:2392)
>>> at org.apache.hadoop.fs.FileSystem.access$200(FileSystem.java:89)
>>> at
>>> org.apache.hadoop.fs.FileSystem$Cache.getInternal(FileSystem.java:2431)
>>> at org.apache.hadoop.fs.FileSystem$Cache.get(FileSystem.java:2413)
>>> at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:368)
>>> at org.apache.hadoop.fs.Path.getFileSystem(Path.java:296)
>>> at
>>> org.apache.hadoop.hbase.util.DynamicClassLoader.<init>(DynamicClassLoader.java:104)
>>> at
>>> org.apache.hadoop.hbase.protobuf.ProtobufUtil.<clinit>(ProtobufUtil.java:217)
>>> at org.apache.hadoop.hbase.ClusterId.parseFrom(ClusterId.java:64)
>>> at
>>> org.apache.hadoop.hbase.zookeeper.ZKClusterId.readClusterIdZNode(ZKClusterId.java:75)
>>> at
>>> org.apache.hadoop.hbase.client.ZooKeeperRegistry.getClusterId(ZooKeeperRegistry.java:86)
>>> at
>>> org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.retrieveClusterId(ConnectionManager.java:853)
>>> at
>>> org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.<init>(ConnectionManager.java:657)
>>> at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
>>> at
>>> sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:57)
>>> at
>>> sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
>>> at java.lang.reflect.Constructor.newInstance(Constructor.java:526)
>>> at
>>> org.apache.hadoop.hbase.client.ConnectionManager.createConnection(ConnectionManager.java:419)
>>> at
>>> org.apache.hadoop.hbase.client.ConnectionManager.createConnectionInternal(ConnectionManager.java:314)
>>> at
>>> org.apache.hadoop.hbase.client.HConnectionManager.createConnection(HConnectionManager.java:291)
>>> at
>>> org.apache.phoenix.query.HConnectionFactory$HConnectionFactoryImpl.createConnection(HConnectionFactory.java:47)
>>> at
>>> org.apache.phoenix.query.ConnectionQueryServicesImpl.openConnection(ConnectionQueryServicesImpl.java:252)
>>> at
>>> org.apache.phoenix.query.ConnectionQueryServicesImpl.init(ConnectionQueryServicesImpl.java:1447)
>>> at
>>> org.apache.phoenix.jdbc.PhoenixDriver.getConnectionQueryServices(PhoenixDriver.java:131)
>>> at
>>> org.apache.phoenix.jdbc.PhoenixEmbeddedDriver.connect(PhoenixEmbeddedDriver.java:112)
>>> at java.sql.DriverManager.getConnection(DriverManager.java:571)
>>> at java.sql.DriverManager.getConnection(DriverManager.java:187)
>>> at org.apache.phoenix.util.PhoenixRuntime.main(PhoenixRuntime.java:197)
>>>
>>> java.sql.SQLException: ERROR 2006 (INT08): Incompatible jars detected
>>> between client and server. Ensure that phoenix.jar is put on the classpath
>>> of HBase in every region server: ERROR 1102 (XCL02): Cannot get all table
>>> regions
>>> at
>>> org.apache.phoenix.exception.SQLExceptionCode$Factory$1.newException(SQLExceptionCode.java:309)
>>> at
>>> org.apache.phoenix.exception.SQLExceptionInfo.buildException(SQLExceptionInfo.java:133)
>>> at
>>> org.apache.phoenix.query.ConnectionQueryServicesImpl.checkClientServerCompatibility(ConnectionQueryServicesImpl.java:932)
>>> at
>>> org.apache.phoenix.query.ConnectionQueryServicesImpl.ensureTableCreated(ConnectionQueryServicesImpl.java:831)
>>> at
>>> org.apache.phoenix.query.ConnectionQueryServicesImpl.createTable(ConnectionQueryServicesImpl.java:1058)
>>> at
>>> org.apache.phoenix.schema.MetaDataClient.createTableInternal(MetaDataClient.java:1156)
>>> at
>>> org.apache.phoenix.schema.MetaDataClient.createTable(MetaDataClient.java:422)
>>> at
>>> org.apache.phoenix.compile.CreateTableCompiler$2.execute(CreateTableCompiler.java:183)
>>> at
>>> org.apache.phoenix.jdbc.PhoenixStatement.executeMutation(PhoenixStatement.java:226)
>>> at
>>> org.apache.phoenix.jdbc.PhoenixStatement.executeUpdate(PhoenixStatement.java:908)
>>> at
>>> org.apache.phoenix.query.ConnectionQueryServicesImpl.init(ConnectionQueryServicesImpl.java:1453)
>>> at
>>> org.apache.phoenix.jdbc.PhoenixDriver.getConnectionQueryServices(PhoenixDriver.java:131)
>>> at
>>> org.apache.phoenix.jdbc.PhoenixEmbeddedDriver.connect(PhoenixEmbeddedDriver.java:112)
>>> at java.sql.DriverManager.getConnection(DriverManager.java:571)
>>> at java.sql.DriverManager.getConnection(DriverManager.java:187)
>>> at org.apache.phoenix.util.PhoenixRuntime.main(PhoenixRuntime.java:197)
>>> Caused by: java.sql.SQLException: ERROR 1102 (XCL02): Cannot get all
>>> table regions
>>> at
>>> org.apache.phoenix.exception.SQLExceptionCode$Factory$1.newException(SQLExceptionCode.java:309)
>>> at
>>> org.apache.phoenix.exception.SQLExceptionInfo.buildException(SQLExceptionInfo.java:133)
>>> at
>>> org.apache.phoenix.query.ConnectionQueryServicesImpl.getAllTableRegions(ConnectionQueryServicesImpl.java:425)
>>> at
>>> org.apache.phoenix.query.ConnectionQueryServicesImpl.checkClientServerCompatibility(ConnectionQueryServicesImpl.java:887)
>>> ... 13 more
>>> Caused by: org.apache.hadoop.hbase.client.NoServerForRegionException: No
>>> server address listed in hbase:meta for region
>>> SYSTEM.CATALOG,,1408684006641.0d1ea455127dd4af6a806574b1f42a91. containing
>>> row
>>> at
>>> org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.locateRegionInMeta(ConnectionManager.java:1334)
>>> at
>>> org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.locateRegion(ConnectionManager.java:1128)
>>> at
>>> org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.relocateRegion(ConnectionManager.java:1097)
>>> at
>>> org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.relocateRegion(ConnectionManager.java:1084)
>>> at
>>> org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.getRegionLocation(ConnectionManager.java:904)
>>> at
>>> org.apache.phoenix.query.ConnectionQueryServicesImpl.getAllTableRegions(ConnectionQueryServicesImpl.java:411)
>>> ... 14 more
>>>
>>> --
>>> Regards,
>>> Vikas Agarwal
>>> 91 – 9928301411
>>>
>>> InfoObjects, Inc.
>>> Execution Matters
>>> http://www.infoobjects.com
>>> 2041 Mission College Boulevard, #280
>>> Santa Clara, CA 95054
>>> +1 (408) 988-2000 Work
>>> +1 (408) 716-2726 Fax
>>
>>
>>
>> CONFIDENTIALITY NOTICE
>> NOTICE: This message is intended for the use of the individual or entity
>> to which it is addressed and may contain information that is confidential,
>> privileged and exempt from disclosure under applicable law. If the reader of
>> this message is not the intended recipient, you are hereby notified that any
>> printing, copying, dissemination, distribution, disclosure or forwarding of
>> this communication is strictly prohibited. If you have received this
>> communication in error, please contact the sender immediately and delete it
>> from your system. Thank You.
>
>
>
>
> --
> Regards,
> Vikas Agarwal
> 91 – 9928301411
>
> InfoObjects, Inc.
> Execution Matters
> http://www.infoobjects.com
> 2041 Mission College Boulevard, #280
> Santa Clara, CA 95054
> +1 (408) 988-2000 Work
> +1 (408) 716-2726 Fax



--
Regards,
Vikas Agarwal
91 – 9928301411

InfoObjects, Inc. 
Execution Matters
http://www.infoobjects.com 
2041 Mission College Boulevard, #280 
Santa Clara, CA 95054
+1 (408) 988-2000 Work
+1 (408) 716-2726 Fax