Return-Path: X-Original-To: apmail-hbase-dev-archive@www.apache.org Delivered-To: apmail-hbase-dev-archive@www.apache.org Received: from mail.apache.org (hermes.apache.org [140.211.11.3]) by minotaur.apache.org (Postfix) with SMTP id A173E11431 for ; Tue, 22 Jul 2014 05:47:44 +0000 (UTC) Received: (qmail 78496 invoked by uid 500); 22 Jul 2014 05:47:43 -0000 Delivered-To: apmail-hbase-dev-archive@hbase.apache.org Received: (qmail 78413 invoked by uid 500); 22 Jul 2014 05:47:43 -0000 Mailing-List: contact dev-help@hbase.apache.org; run by ezmlm Precedence: bulk List-Help: List-Unsubscribe: List-Post: List-Id: Reply-To: dev@hbase.apache.org Delivered-To: mailing list dev@hbase.apache.org Received: (qmail 78401 invoked by uid 99); 22 Jul 2014 05:47:43 -0000 Received: from nike.apache.org (HELO nike.apache.org) (192.87.106.230) by apache.org (qpsmtpd/0.29) with ESMTP; Tue, 22 Jul 2014 05:47:43 +0000 X-ASF-Spam-Status: No, hits=-0.7 required=5.0 tests=RCVD_IN_DNSWL_LOW,SPF_PASS X-Spam-Check-By: apache.org Received-SPF: pass (nike.apache.org: domain of fancyerii@gmail.com designates 209.85.217.174 as permitted sender) Received: from [209.85.217.174] (HELO mail-lb0-f174.google.com) (209.85.217.174) by apache.org (qpsmtpd/0.29) with ESMTP; Tue, 22 Jul 2014 05:47:40 +0000 Received: by mail-lb0-f174.google.com with SMTP id c11so5433981lbj.5 for ; Mon, 21 Jul 2014 22:47:16 -0700 (PDT) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=gmail.com; s=20120113; h=mime-version:in-reply-to:references:date:message-id:subject:from:to :content-type; bh=dcUFfXSl2U+4xe6/RtDEFVRihm5C6rsWmMkErHisoK0=; b=DkIGRWcIGUY6vhFNEVnFF2vtrABTLgh7N7QBOAQuC57sy8zdAjWmr6gdBwzLiSEF6b 91cVbuWbXrDBrUipNQyJ+h4szv9aBHNMXdVY1Wh7oPbgyrZy5bM2j4t6lXvM1aWtlVRP UdibwU91l/MyzBcKtP8159zHJ4CBDVUKPTcaHt3o/rMo0QK80YsDDu8Qk6uNRVLcg3MG kVkPQ8ioTgcMLqZJtmaWm4YEGuLNT0HP90LwcetRrCoyy9Pky0lrJN/7sE/jvzD6P/d1 K1NHwdM/HvQZj5+SvXE/GA80hbI9BvPj2lpi9Iejqyjvrrsj0ddYmk+xLX/NKm+G0LK3 Myuw== MIME-Version: 1.0 X-Received: by 10.152.185.104 with SMTP id fb8mr32075583lac.64.1406008036419; Mon, 21 Jul 2014 22:47:16 -0700 (PDT) Received: by 10.112.151.200 with HTTP; Mon, 21 Jul 2014 22:47:16 -0700 (PDT) In-Reply-To: References: Date: Tue, 22 Jul 2014 13:47:16 +0800 Message-ID: Subject: Re: how to do parallel scanning in map reduce using hbase as input? From: Li Li To: dev@hbase.apache.org Content-Type: text/plain; charset=UTF-8 X-Virus-Checked: Checked by ClamAV on apache.org sorry. I have not finished it. 1. yes, I have 20 concurrent running mappers. 2. I can't add more mappers because I set io.sort.mb to 500mb and if I set 8 mappers, it hit oov exception and load average is high 3. fast mapper only use 1 minute. following is the statistics HBase Counters REMOTE_RPC_CALLS 0 RPC_CALLS 523 RPC_RETRIES 0 NOT_SERVING_REGION_EXCEPTION 0 NUM_SCANNER_RESTARTS 0 MILLIS_BETWEEN_NEXTS 62,415 BYTES_IN_RESULTS 1,380,694,667 BYTES_IN_REMOTE_RESULTS 0 REGIONS_SCANNED 1 REMOTE_RPC_RETRIES 0 FileSystemCounters FILE_BYTES_READ 120,508,552 HDFS_BYTES_READ 176 FILE_BYTES_WRITTEN 241,000,600 File Input Format Counters Bytes Read 0 Map-Reduce Framework Map output materialized bytes 120,448,992 Combine output records 0 Map input records 5,208,607 Physical memory (bytes) snapshot 965,730,304 Spilled Records 10,417,214 Map output bytes 282,122,973 CPU time spent (ms) 82,610 Total committed heap usage (bytes) 1,061,158,912 Virtual memory (bytes) snapshot 1,681,047,552 Combine input records 0 Map output records 5,208,607 SPLIT_RAW_BYTES 176 slow mapper cost 25 minutes HBase Counters REMOTE_RPC_CALLS 0 RPC_CALLS 2,268 RPC_RETRIES 0 NOT_SERVING_REGION_EXCEPTION 0 NUM_SCANNER_RESTARTS 0 MILLIS_BETWEEN_NEXTS 907,402 BYTES_IN_RESULTS 9,459,568,932 BYTES_IN_REMOTE_RESULTS 0 REGIONS_SCANNED 1 REMOTE_RPC_RETRIES 0 FileSystemCounters FILE_BYTES_READ 2,274,832,004 HDFS_BYTES_READ 161 FILE_BYTES_WRITTEN 3,770,108,961 File Input Format Counters Bytes Read 0 Map-Reduce Framework Map output materialized bytes 1,495,451,997 Combine output records 0 Map input records 22,659,551 Physical memory (bytes) snapshot 976,842,752 Spilled Records 57,085,847 Map output bytes 3,348,373,811 CPU time spent (ms) 1,134,640 Total committed heap usage (bytes) 945,291,264 Virtual memory (bytes) snapshot 1,699,991,552 Combine input records 0 Map output records 22,644,687 SPLIT_RAW_BYTES 161 4. I have about 11 billion rows and it takes 1.3TB(hdfs usage) and the replication factor is 2 5. for block information, one column family file: Name Type Size Replication Block Size Modification Time Permission Owner Group b8297e0a415a4ddc811009e70aa30371 file 195.43 MB 2 64 MB 2014-07-22 10:16 rw-r--r-- hadoop supergroup dea1d498ec6d46ea84ad35ea6cc3cf6e file 5.12 GB 2 64 MB 2014-07-20 20:24 rw-r--r-- hadoop supergroup ee01947bad6f450d89bd71be84d9d60a file 2.68 MB 2 64 MB 2014-07-22 13:18 rw-r--r-- hadoop supergroup another example 1923bdcf47ed40879ec4a2f6d314167e file 729.43 MB 2 64 MB 2014-07-18 20:32 rw-r--r-- hadoop supergroup 532d56af4457492194c5336f1f1d8359 file 372.27 MB 2 64 MB 2014-07-21 20:55 rw-r--r-- hadoop supergroup 55e92aef7b754059be9fc7e4692832ec file 117.45 MB 2 64 MB 2014-07-22 13:19 rw-r--r-- hadoop supergroup c927509f280a4cb3bc5c6db2feea5c16 file 7.87 GB 2 64 MB 2014-07-12 06:55 rw-r--r-- hadoop supergroup 6. I have only one column family for this table 7. each row has less than 10 columns 8. region info in web ui ServerName Num. Stores Num. Storefiles Storefile Size Uncompressed Storefile Size Index Size Bloom Size mphbase1,60020,1405730850512 46 103 126528m 126567mb 94993k 329266k mphbase2,60020,1405730850549 45 100 157746m 157789mb 117250k 432066k mphbase3,60020,1405730850546 46 46 53592m 53610mb 42858k 110748k mphbase4,60020,1405730850585 43 101 109790m 109827mb 83236k 295068k mphbase5,60020,1405730850652 41 81 89073m 89099mb 66622k 243354k 9. url_db has 84 regions On Tue, Jul 22, 2014 at 1:32 PM, Li Li wrote: > 1. yes, I have 20 concurrent running mappers. > 2. I can't add more mappers because I set io.sort.mb to 500mb and if I > set 8 mappers, it hit oov exception and load average is high > 3. fast mapper only use 1 minute. following is the statistics > HBase Counters > REMOTE_RPC_CALLS 0 > RPC_CALLS 523 > RPC_RETRIES 0 > NOT_SERVING_REGION_EXCEPTION 0 > NUM_SCANNER_RESTARTS 0 > MILLIS_BETWEEN_NEXTS 62,415 > BYTES_IN_RESULTS 1,380,694,667 > BYTES_IN_REMOTE_RESULTS 0 > REGIONS_SCANNED 1 > REMOTE_RPC_RETRIES 0 > > FileSystemCounters > FILE_BYTES_READ 120,508,552 > HDFS_BYTES_READ 176 > FILE_BYTES_WRITTEN 241,000,600 > > File Input Format Counters > Bytes Read 0 > > Map-Reduce Framework > Map output materialized bytes 120,448,992 > Combine output records 0 > Map input records 5,208,607 > Physical memory (bytes) snapshot 965,730,304 > Spilled Records 10,417,214 > Map output bytes 282,122,973 > CPU time spent (ms) 82,610 > Total committed heap usage (bytes) 1,061,158,912 > Virtual memory (bytes) snapshot 1,681,047,552 > Combine input records 0 > Map output records 5,208,607 > SPLIT_RAW_BYTES 176 > > > On Tue, Jul 22, 2014 at 12:11 PM, Stack wrote: >> How many regions now? >> >> You still have 20 concurrent mappers running? Are your machines loaded w/ >> 4 map tasks on each? Can you up the number of concurrent mappers? Can you >> get an idea of your scan rates? Are all map tasks scanning at same rate? >> Does one task lag the others? Do you emit stats on each map task such as >> rows processed? Can you figure your bottleneck? Are you seeking disk all >> the time? Anything else running while this big scan is going on? How big >> are your cells? Do you have one or more column families? How many columns? >> >> For average region size, do du on the hdfs region directories and then sum >> and divide by region count. >> >> St.Ack >> >> >> On Mon, Jul 21, 2014 at 7:30 PM, Li Li wrote: >> >>> anyone could help? now I have about 1.1 billion nodes and it takes 2 >>> hours to finish a map reduce job. >>> >>> ---------- Forwarded message ---------- >>> From: Li Li >>> Date: Thu, Jun 26, 2014 at 3:34 PM >>> Subject: how to do parallel scanning in map reduce using hbase as input? >>> To: user@hbase.apache.org >>> >>> >>> my table has about 700 million rows and about 80 regions. each task >>> tracker is configured with 4 mappers and 4 reducers at the same time. >>> The hadoop/hbase cluster has 5 nodes so at the same time, it has 20 >>> mappers running. it takes more than an hour to finish mapper stage. >>> The hbase cluster's load is very low, about 2,000 request per second. >>> I think one mapper for a region is too small. How can I run more than >>> one mapper for a region so that it can take full advantage of >>> computing resources? >>>