phoenix-user mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From James Taylor <>
Subject Re: Phoenix UPSERTS SELECT for data transformation
Date Wed, 28 Jan 2015 04:51:45 GMT
Hi Jaime,

Would it be possible to see a few examples of the kind of
transformations you're doing? I think the tool you use depends on
whether you're transforming all of your data or a smaller subset. It
also depends on the complexity of the transformation. If you're
transforming every row of data in your HBase table to create a new row
in a different HBase table, Phoenix is not going to be an ideal
choice. Also, if you're transforming data such that you need a new
built-in function for each kind of transformation, Phoenix would not
be the right choice.

Have you seen our map-reduce[1] and pig integration[2] support? Pig is
very good at ETL. You may be able to leverage Pig to do the
transformation such that the resulting table is queryable through
Phoenix as well.

HTH. Thanks,



On Tue, Jan 27, 2015 at 11:50 AM, Jaime Solano <> wrote:
> Hi guys,
> The company I work for wants to use Phoenix for Data Transformation.
> Basically, the idea is to denormalize and include additional calculated data
> to a details table, by using UPSERT SELECT statements (joins with other
> tables and specific functions). This has proven to be challenging, since the
> SQL language sometimes is not enough, leading us to try implement our own
> built-in Phoenix functions (following this post:
> I feel this is not the right direction, and maybe we should be using other
> tools like Pig, MR or Storm (for Near-Real Time).
> What are your thoughts about this? Would you recommend Phoenix for complex
> Data transformation? What a re the drawbacks you see in this approach?
> Thanks in advance,
> -Jaime

View raw message