in case of csv try presplitting the table into some x regions and disable the WAL for bulk loading

Cheers,
Yeshwanth


On Wed, Sep 17, 2014 at 8:11 PM, Gabriel Reid <gabriel.reid@gmail.com> wrote:
Hi Poonam,

Could you say a bit about how you're doing this bulk load? Are you
loading from a CSV file, or making JDBC calls yourself (or something
else)?

Are you aware of the CSV bulk loader
(http://phoenix.apache.org/bulk_dataload.html)? That will probably
give you the best performance for a large load of data.

- Gabriel


On Wed, Sep 17, 2014 at 3:44 PM, Poonam Ligade
<poonam.v.ligade@gmail.com> wrote:
> Hi,
>
> I am trying to load around 15GB of data in hbase
> Files are split in around 1.3GB. I have 2 regionservers and 1 master running
> on seperate machines with 4 GB physical RAM each.
> Heapsize For Hmaster and regionserver is 3 GB each
>
> How can I configure HBase to handle these writes faster.
>
> Currently I have below settings
>
> hbase.regionserver.handler.count                         60
> hbase.client.write.buffer                                      8388608
> hbase.hregion.memstore.flush.size                     134217728
> hbase.regionserver.global.memstore.lowerLimit    0.95
>
> hbase.regionserver.global.memstore.upperLimit    0.4
>
> hbase.hregion.max.filesize                                 3294967296
>
>
>
> Regards,
> Poonam