Hello friends..,

I am very new to Apache Phoenix and i just started running sample phoenix spark example in spark 1.6 version. it was successful and now i want to run this example in spark version 2.0.0. Is phoenix provides support for spark-2.0.0?

previously i used this command:

DataFrame fromPhx = context.read().format("org.apache.phoenix.spark")

.options(ImmutableMap.of("driver", "org.apache.phoenix.jdbc.PhoenixDriver", "zkUrl",

"jdbc:phoenix:localhost:2181", "table", "SAMPLE"))

.load();


In spark 2.0.0:


org.apache.spark.sql.Dataset<Row> df  = spark.read().format("org.apache.phoenix.spark")

.options(ImmutableMap.of("driver", "org.apache.phoenix.jdbc.PhoenixDriver", "zkUrl",

"jdbc:phoenix:localhost:2181", "table", "SAMPLE"))

.load();


This is correct or i need to change any code?


please help me out.