hadoop-common-commits mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From Apache Wiki <wikidi...@apache.org>
Subject [Hadoop Wiki] Update of "Hbase/FAQ" by JeffHammerbacher
Date Mon, 17 May 2010 17:47:32 GMT
Dear Wiki user,

You have subscribed to a wiki page or wiki category on "Hadoop Wiki" for change notification.

The "Hbase/FAQ" page has been changed by JeffHammerbacher.


   1. [[#16|Can HBase development be done on windows?]]
   1. [[#17|Please explain HBase version numbering?]]
   1. [[#18|What version of Hadoop do I need to run HBase?]]
-  1. [[#10|Any other troubleshooting pointers for me?]]
+  1. [[#19|Any other troubleshooting pointers for me?]]
   1. [[#20|Are there any schema design examples?]]
   1. [[#21|How do I add/remove a node?]]
   1. [[#22|Why do servers have start codes?]]
+  1. [[#23|What is the maximum recommended cell size?]]
  == Answers ==
@@ -224, +225 @@

  If a region server crashes and recovers, it cannot be given work until its lease times out.
If the lease is identified only by an IP address and port number, then that server can't do
any progress until the lease times out. A start code is added so that the restarted server
can begin doing work immediately upon recovery. For more, see https://issues.apache.org/jira/browse/HBASE-1156.
+ '''23. <<Anchor(23)>> What is the maximum recommended cell size?'''
+ A rough rule of thumb, with little empirical validation, is to keep the data in HDFS and
store pointers to the data in HBase if you expect the cell size to be consistently above 10
MB. If you do expect large cell values and you still plan to use HBase for the storage of
cell contents, you'll want to increase the block size and the maximum region size for the
table to keep the index size reasonable and the split frequency acceptable.

View raw message