hbase-user mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From "Daniel Leffel" <daniel.lef...@gmail.com>
Subject Re: Bug in 0.1.3 scanner?
Date Tue, 22 Jul 2008 19:43:01 GMT
Breakthrough in the mystery?

The key that it fails on just happens to be the last key in the region....

Is opening the scanner on the new region the problem?



On Tue, Jul 22, 2008 at 12:18 PM, Daniel Leffel <daniel.leffel@gmail.com>wrote:

> OK. So, I didn't notice the IOE subclassing, so that explained the fact I
> wasn't catching it.
>
> However, somethign is still definitely going on. The scanner at some point
> gets "stuck". See the next's being called:
> 2008-07-22 13:34:18,274 DEBUG
> org.apache.hadoop.hbase.filter.WhileMatchRowFilter: FilterNotNull on
> cols:{hierarcy:immediate_parents=[B@26c1186f}.  Result = false
> 2008-07-22 13:34:18,284 DEBUG
> org.apache.hadoop.hbase.filter.WhileMatchRowFilter: this.filterAllRemaining
> is now: false
> 2008-07-22 13:34:18,285 DEBUG
> org.apache.hadoop.hbase.filter.WhileMatchRowFilter: FilterNotNull on
> cols:{hierarcy:immediate_parents=[B@2db6235b}.  Result = false
> 2008-07-22 13:34:18,292 DEBUG org.apache.hadoop.hbase.filter.StopRowFilter:
> Filter result for rowKey: 1121874.  Result: false
> 2008-07-22 13:34:18,292 DEBUG
> org.apache.hadoop.hbase.filter.WhileMatchRowFilter: this.filterAllRemaining
> is now: false
> 2008-07-22 13:34:18,292 DEBUG
> org.apache.hadoop.hbase.filter.WhileMatchRowFilter: Filter on
> rowKey:1121874.  Result = false
> 2008-07-22 13:34:18,292 DEBUG org.apache.hadoop.hbase.filter.StopRowFilter:
> Filter result for rowKey: 1121874.  Result: false
> 2008-07-22 13:34:18,292 DEBUG
> org.apache.hadoop.hbase.filter.WhileMatchRowFilter: this.filterAllRemaining
> is now: false
> 2008-07-22 13:34:18,292 DEBUG
> org.apache.hadoop.hbase.filter.WhileMatchRowFilter: Filter on
> rowKey:1121874, colKey: hierarcy:immediate_parents, data: [B@7d5b1773.
> Result = false
> 2008-07-22 13:34:18,324 DEBUG org.apache.hadoop.hbase.filter.StopRowFilter:
> Filter result for rowKey: 11218740.  Result: false
> 2008-07-22 13:34:18,324 DEBUG
> org.apache.hadoop.hbase.filter.WhileMatchRowFilter: this.filterAllRemaining
> is now: false
> 2008-07-22 13:34:18,325 DEBUG
> org.apache.hadoop.hbase.filter.WhileMatchRowFilter: Filter on
> rowKey:11218740.  Result = false
> 2008-07-22 13:34:18,325 DEBUG org.apache.hadoop.hbase.filter.StopRowFilter:
> Filter result for rowKey: 11218740.  Result: false
> 2008-07-22 13:34:18,325 DEBUG
> org.apache.hadoop.hbase.filter.WhileMatchRowFilter: this.filterAllRemaining
> is now: false
> 2008-07-22 13:34:18,325 DEBUG
> org.apache.hadoop.hbase.filter.WhileMatchRowFilter: Filter on
> rowKey:11218740, colKey: hierarcy:immediate_parents, data: [B@b05236.
> Result = false
> 2008-07-22 13:34:18,334 DEBUG
> org.apache.hadoop.hbase.filter.WhileMatchRowFilter: this.filterAllRemaining
> is now: false
> 2008-07-22 13:34:18,334 DEBUG
> org.apache.hadoop.hbase.filter.WhileMatchRowFilter: FilterNotNull on
> cols:{hierarcy:immediate_parents=[B@7d5b1773}.  Result = false
> 2008-07-22 13:34:18,744 DEBUG
> org.apache.hadoop.hbase.filter.WhileMatchRowFilter: Successfully read a
> sub-filter of type: org.apache.hadoop.hbase.filter.StopRowFilter
> 2008-07-22 13:34:18,745 DEBUG
> org.apache.hadoop.hbase.filter.WhileMatchRowFilter: Successfully read a
> sub-filter of type: org.apache.hadoop.hbase.filter.StopRowFilter
> 2008-07-22 13:34:18,745 DEBUG
> org.apache.hadoop.hbase.filter.WhileMatchRowFilter: Resetting.
> 2008-07-22 13:34:18,758 DEBUG org.apache.hadoop.hbase.filter.StopRowFilter:
> Filter result for rowKey: 11391639.  Result: false
> 2008-07-22 13:34:18,758 DEBUG
> org.apache.hadoop.hbase.filter.WhileMatchRowFilter: this.filterAllRemaining
> is now: false
> 2008-07-22 13:34:18,758 DEBUG
> org.apache.hadoop.hbase.filter.WhileMatchRowFilter: Filter on
> rowKey:11391639.  Result = false
> 2008-07-22 13:34:18,759 DEBUG org.apache.hadoop.hbase.filter.StopRowFilter:
> Filter result for rowKey: 11391639.  Result: false
> 2008-07-22 13:34:18,759 DEBUG
> org.apache.hadoop.hbase.filter.WhileMatchRowFilter: this.filterAllRemaining
> is now: false
> 2008-07-22 13:34:18,759 DEBUG
> org.apache.hadoop.hbase.filter.WhileMatchRowFilter: Filter on
> rowKey:11391639, colKey: hierarcy:immediate_parents, data: [B@273d1402.
> Result = false
> 2008-07-22 13:34:18,779 DEBUG org.apache.hadoop.hbase.filter.StopRowFilter:
> Filter result for rowKey: 1139164.  Result: false
> 2008-07-22 13:34:18,779 DEBUG
> org.apache.hadoop.hbase.filter.WhileMatchRowFilter: this.filterAllRemaining
> is now: false
> 2008-07-22 13:34:18,779 DEBUG
> org.apache.hadoop.hbase.filter.WhileMatchRowFilter: Filter on
> rowKey:1139164.  Result = false
> 2008-07-22 13:34:18,779 DEBUG org.apache.hadoop.hbase.filter.StopRowFilter:
> Filter result for rowKey: 1139164.  Result: false
> 2008-07-22 13:34:18,779 DEBUG
> org.apache.hadoop.hbase.filter.WhileMatchRowFilter: this.filterAllRemaining
> is now: false
> 2008-07-22 13:34:18,779 DEBUG
> org.apache.hadoop.hbase.filter.WhileMatchRowFilter: Filter on
> rowKey:1139164, colKey: hierarcy:immediate_parents, data: [B@25dd9891.
> Result = false
> 2008-07-22 13:34:18,782 DEBUG
> org.apache.hadoop.hbase.filter.WhileMatchRowFilter: this.filterAllRemaining
> is now: false
> 2008-07-22 13:34:18,783 DEBUG
> org.apache.hadoop.hbase.filter.WhileMatchRowFilter: FilterNotNull on
> cols:{hierarcy:immediate_parents=[B@273d1402}.  Result = false
>
>
> Then at some point, next just hangs and the next entry in the log is
> expiration:
> 2008-07-22 13:44:18,015 INFO org.apache.hadoop.hbase.Leases:
> regionserver/0.0.0.0:60020.leaseChecker lease expired
> -4780003109149242775/-4780003109149242775
> 2008-07-22 13:44:18,015 INFO org.apache.hadoop.hbase.HRegionServer: Scanner
> -4780003109149242775 lease expired
> 2008-07-22 13:44:18,015 INFO org.apache.hadoop.hbase.Leases:
> regionserver/0.0.0.0:60020.leaseChecker lease expired
> -8543632393752004827/-8543632393752004827
> 2008-07-22 13:44:18,015 INFO org.apache.hadoop.hbase.HRegionServer: Scanner
> -8543632393752004827 lease expired
> 2008-07-22 13:44:18,015 INFO org.apache.hadoop.hbase.Leases:
> regionserver/0.0.0.0:60020.leaseChecker lease expired
> -3834469434163616397/-3834469434163616397
> 2008-07-22 13:44:18,015 INFO org.apache.hadoop.hbase.HRegionServer: Scanner
> -3834469434163616397 lease expired
> 2008-07-22 13:44:28,024 INFO org.apache.hadoop.hbase.Leases:
> regionserver/0.0.0.0:60020.leaseChecker lease expired
> -6363251876420626143/-6363251876420626143
> 2008-07-22 13:44:28,024 INFO org.apache.hadoop.hbase.HRegionServer: Scanner
> -6363251876420626143 lease expired
> 2008-07-22 13:44:28,024 INFO org.apache.hadoop.hbase.Leases:
> regionserver/0.0.0.0:60020.leaseChecker lease expired
> -4332402553799574820/-4332402553799574820
> 2008-07-22 13:44:28,024 INFO org.apache.hadoop.hbase.HRegionServer: Scanner
> -4332402553799574820 lease expired
> 2008-07-22 13:44:28,024 INFO org.apache.hadoop.hbase.Leases:
> regionserver/0.0.0.0:60020.leaseChecker lease expired
> 7541512102009958038/7541512102009958038
> 2008-07-22 13:44:28,024 INFO org.apache.hadoop.hbase.HRegionServer: Scanner
> 7541512102009958038 lease expired
> 2008-07-22 13:44:28,025 INFO org.apache.hadoop.hbase.Leases:
> regionserver/0.0.0.0:60020.leaseChecker lease expired
> -6673102874800967750/-6673102874800967750
> 2008-07-22 13:44:28,025 INFO org.apache.hadoop.hbase.HRegionServer: Scanner
> -6673102874800967750 lease expired
> 2008-07-22 13:44:28,025 INFO org.apache.hadoop.hbase.Leases:
> regionserver/0.0.0.0:60020.leaseChecker lease expired
> 9143807894343107328/9143807894343107328
> 2008-07-22 13:44:28,025 INFO org.apache.hadoop.hbase.HRegionServer: Scanner
> 9143807894343107328 lease expired
>
> Now interestingly, the map task doesn't know about the expired lease until
> a number of minutes later when it finally tries to call next:
> 2008-07-22 14:54:29,062 INFO org.apache.hadoop.ipc.Server: IPC Server
> handler 8 on 60020, call next(-4780003109149242775) from
> 10.252.242.159:60900: error:
> org.apache.hadoop.hbase.UnknownScannerException: Name: -4780003109149242775
> org.apache.hadoop.hbase.UnknownScannerException: Name: -4780003109149242775
>         at
> org.apache.hadoop.hbase.HRegionServer.next(HRegionServer.java:1506)
>         at sun.reflect.GeneratedMethodAccessor3.invoke(Unknown Source)
>         at
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
>         at java.lang.reflect.Method.invoke(Method.java:597)
>         at
> org.apache.hadoop.hbase.ipc.HbaseRPC$Server.call(HbaseRPC.java:413)
>         at org.apache.hadoop.ipc.Server$Handler.run(Server.java:901)
> 2008-07-22 14:54:29,070 INFO org.apache.hadoop.ipc.Server: IPC Server
> handler 3 on 60020, call close(-4780003109149242775) from
> 10.252.242.159:60900: error:
> org.apache.hadoop.hbase.UnknownScannerException: -4780003109149242775
> org.apache.hadoop.hbase.UnknownScannerException: -4780003109149242775
>         at
> org.apache.hadoop.hbase.HRegionServer.close(HRegionServer.java:1616)
>         at sun.reflect.GeneratedMethodAccessor11.invoke(Unknown Source)
>         at
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
>         at java.lang.reflect.Method.invoke(Method.java:597)
>         at
> org.apache.hadoop.hbase.ipc.HbaseRPC$Server.call(HbaseRPC.java:413)
>         at org.apache.hadoop.ipc.Server$Handler.run(Server.java:901)
>
>
> Even more interestingly is that the other scanners for the other tasks were
> also "frozen" during the time and their leases expire too.
>
> Inside the iteration, it's pretty simple stuff happening - string
> manipulation mostly and a put into a bdb - takes <10ms.
>
> The only other suspect piece of code is a reporter.progress() which I had
> to put or else the task tracker thinks the task is dead. I've not had
> trouble with reporter.progress() to date...
>
> Any thoughts?
>
>
> On Tue, Jul 22, 2008 at 11:22 AM, stack <stack@duboce.net> wrote:
>
>> Hey D:
>>
>> TableInputFormat#TableRecordReader#next looks like it lets out
>> IOExceptions.  UnknownScannerException subclasses IOE so I'd think it should
>> bubble up into your map task.
>>
>> Paste in the exception you found in your logs.  That might help figure
>> whats going on.
>>
>> St.Ack
>>
>>
>> Daniel Leffel wrote:
>>
>>> I've had a little bit of weird behavior.
>>>
>>> I am opening a scanner in the configure method of a Map task to load a
>>> simple little in-memory map (I'd love to this with in-memory column
>>> stores,
>>> but that's another story ;-).
>>>
>>> Anyway, the scanner iterates over 4 million rows or so and I've verified
>>> next is being called multiple times per second. However, at the same
>>> record
>>> every time (after around 6 minutes of iteration), the scanner throws an
>>> UnknownScannerException in the log (I see it in the log), but rather than
>>> throw the exception to the Map task, next just returns false and the loop
>>> ends (making it exceptionally difficult to detect if the scanner finished
>>> scanning or not). Increasing the regionserver timeout has most definitely
>>> solved the issue, but shouldn't calling next reset the timeout anyway?
>>>
>>> Such a frustrating problem!
>>>
>>>
>>>
>>
>>
>

Mime
  • Unnamed multipart/alternative (inline, None, 0 bytes)
View raw message