Could you say a bit about how you're doing this bulk load? Are you
loading from a CSV file, or making JDBC calls yourself (or something
Are you aware of the CSV bulk loader
(http://phoenix.apache.org/bulk_dataload.html)? That will probably
give you the best performance for a large load of data.
On Wed, Sep 17, 2014 at 3:44 PM, Poonam Ligade
> I am trying to load around 15GB of data in hbase
> Files are split in around 1.3GB. I have 2 regionservers and 1 master running
> on seperate machines with 4 GB physical RAM each.
> Heapsize For Hmaster and regionserver is 3 GB each
> How can I configure HBase to handle these writes faster.
> Currently I have below settings
> hbase.regionserver.handler.count 60
> hbase.client.write.buffer 8388608
> hbase.hregion.memstore.flush.size 134217728
> hbase.regionserver.global.memstore.lowerLimit 0.95
> hbase.regionserver.global.memstore.upperLimit 0.4
> hbase.hregion.max.filesize 3294967296