livy-user mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From Lian Jiang <jiangok2...@gmail.com>
Subject specify external jar for livy2.spark interpreter
Date Fri, 17 Aug 2018 19:49:08 GMT
Hi,

I have a hadoop cluster installed using HDP3.0 and ambari 2.7 blueprint. I
have copied oracle oci hdfs connector jar (downloaded from
https://github.com/oracle/oci-hdfs-connector/releases) to
/usr/hdp/3.0.0.0-1634/livy2/repl_2.11-jars/
on namenodes and datanodes by following
https://docs.hortonworks.com/HDPDocuments/HDP3/HDP-3.0.0/using-zeppelin/zeppelin-using.pdf
.

When using livy2.spark interpreter in Zeppelin 0.8.0 (which use yarn
cluster mode), I got mixed story.

This works which means the oci connector hdfs jar is imported successfully.
%livy2.spark
import com.oracle.bmc.hdfs._

This failed:
%livy2.spark
val df = spark.read.parquet("oci://mybucket@mytenancy/my.parquet")

org.apache.hadoop.fs.UnsupportedFileSystemException: No FileSystem for
scheme "oci" at
org.apache.hadoop.fs.FileSystem.getFileSystemClass(FileSystem.java:3332) at
org.apache.hadoop.fs.FileSystem.createFileSystem(FileSystem.java:3352) at
org.apache.hadoop.fs.FileSystem.access$200(FileSystem.java:124) at
org.apache.hadoop.fs.FileSystem$Cache.getInternal(FileSystem.java:3403) at
org.apache.hadoop.fs.FileSystem$Cache.get(FileSystem.java:3371) at
org.apache.hadoop.fs.FileSystem.get(FileSystem.java:477) at
org.apache.hadoop.fs.Path.getFileSystem(Path.java:361) at
org.apache.spark.sql.execution.datasources.DataSource$.org$apache$spark$sql$execution$datasources$DataSource$$checkAndGlobPathIfNecessary(DataSource.scala:705)
at
org.apache.spark.sql.execution.datasources.DataSource$$anonfun$15.apply(DataSource.scala:389)
at
org.apache.spark.sql.execution.datasources.DataSource$$anonfun$15.apply(DataSource.scala:389)
at
scala.collection.TraversableLike$$anonfun$flatMap$1.apply(TraversableLike.scala:241)
at
scala.collection.TraversableLike$$anonfun$flatMap$1.apply(TraversableLike.scala:241)
at scala.collection.immutable.List.foreach(List.scala:381) at
scala.collection.TraversableLike$class.flatMap(TraversableLike.scala:241)
at scala.collection.immutable.List.flatMap(List.scala:344) at
org.apache.spark.sql.execution.datasources.DataSource.resolveRelation(DataSource.scala:388)
at
org.apache.spark.sql.DataFrameReader.loadV1Source(DataFrameReader.scala:239)
at org.apache.spark.sql.DataFrameReader.load(DataFrameReader.scala:227) at
org.apache.spark.sql.DataFrameReader.parquet(DataFrameReader.scala:622) at
org.apache.spark.sql.DataFrameReader.parquet(DataFrameReader.scala:606) ...
60 elided

Any idea? Appreciate very much!

Mime
View raw message