phoenix-user mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From Roberto Coluccio <>
Subject Use Phoenix hints with Spark Integration [main use case: block cache disable]
Date Wed, 30 Aug 2017 11:11:41 GMT
Hello folks,

I'm facing the issue of disabling adding to the block cache records I'm 
selecting from my Spark application when reading as DataFrame (e.g. 
sqlContext.phoenixTableAsDataFrame(myTable, myColumns, myPredicate, 
myZkUrl, myConf).

I know I can force the no cache on a query basis when issuing SQL 
queries leveraging the /*+ NO_CACHE */ hint.
I know I can disable the caching at a table-specific or colum-family 
specific basis through an ALTER TABLE HBase shell command.

What I don't know is how to do so when leveraging Phoenix-Spark APIs. I 
think my problem can be stated as a more general purpose question: *how 
can Phoenix hints be specified when using Phoenix-Spark APIs?

*For my specific use case, I tried to push within a Configuration object 
the property /hfile.block.cache.size=0/ before creating the DataFrame 
but I realized records resulting from the underneath scan where still 

Thank you in advance for your help.

Best regards,

View raw message