phoenix-user mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From Gaurav Agarwal <gaurav130...@gmail.com>
Subject Re: Phoenix map reduce
Date Wed, 02 Sep 2015 03:35:13 GMT
thanks for the reply.

On Wed, Sep 2, 2015 at 12:48 AM, Jean-Marc Spaggiari <
jean-marc@spaggiari.org> wrote:

> copytable will start a MR job and will do the copy in parallele, which is
> good. But it's still going to do a lot of puts on the destination cluster
> which will trigger flushs and compactions. If it's easy for you to send
> your csv file there I think it will be more efficient, even if copyTbale
> can solve your issue.
>
> JM
>
> 2015-09-01 15:01 GMT-04:00 Gaurav Agarwal <gaurav130403@gmail.com>:
>
>> In this will copyTable command of hbase is good to use or transfer the
>> csv file on the other side and bulkload from there which one is good
>> according to performance
>>
>> On Wed, Sep 2, 2015 at 12:23 AM, Jean-Marc Spaggiari <
>> jean-marc@spaggiari.org> wrote:
>>
>>> Hi Gaurav,
>>>
>>> bulk load bypass the WAL, that's correct. It's true for Phoenix, it's
>>> true for HBase (outside of Phoenix).
>>>
>>> If you have replication activated, you will have to bulkload the data
>>> into the 2 clusters. Transfert your csv files on the other side too and
>>> bulkload from there.
>>>
>>> JM
>>>
>>> 2015-09-01 14:51 GMT-04:00 Gaurav Agarwal <gaurav130403@gmail.com>:
>>>
>>>> Hello
>>>>
>>>> We are using phoenix Map reduce CSV uploader to load data into HBASe .
>>>> I read documentation on Phoenix site, it will only create HFLE no WAL logs
>>>> will be created.Please confirm understanding is correct or wrong
>>>>
>>>> We have to use HBASe replication across cluster for Master Master
>>>> scenario. Will the replication work in that scenario or do we need to use
>>>> Copy Table to replicate ?
>>>>
>>>> thanks
>>>>
>>>
>>>
>>
>

Mime
View raw message