hbase-user mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From "Daniel Leffel" <daniel.lef...@gmail.com>
Subject Re: Bug in 0.1.3 scanner?
Date Tue, 22 Jul 2008 19:18:21 GMT
OK. So, I didn't notice the IOE subclassing, so that explained the fact I
wasn't catching it.

However, somethign is still definitely going on. The scanner at some point
gets "stuck". See the next's being called:
2008-07-22 13:34:18,274 DEBUG
org.apache.hadoop.hbase.filter.WhileMatchRowFilter: FilterNotNull on
cols:{hierarcy:immediate_parents=[B@26c1186f}.  Result = false
2008-07-22 13:34:18,284 DEBUG
org.apache.hadoop.hbase.filter.WhileMatchRowFilter: this.filterAllRemaining
is now: false
2008-07-22 13:34:18,285 DEBUG
org.apache.hadoop.hbase.filter.WhileMatchRowFilter: FilterNotNull on
cols:{hierarcy:immediate_parents=[B@2db6235b}.  Result = false
2008-07-22 13:34:18,292 DEBUG org.apache.hadoop.hbase.filter.StopRowFilter:
Filter result for rowKey: 1121874.  Result: false
2008-07-22 13:34:18,292 DEBUG
org.apache.hadoop.hbase.filter.WhileMatchRowFilter: this.filterAllRemaining
is now: false
2008-07-22 13:34:18,292 DEBUG
org.apache.hadoop.hbase.filter.WhileMatchRowFilter: Filter on
rowKey:1121874.  Result = false
2008-07-22 13:34:18,292 DEBUG org.apache.hadoop.hbase.filter.StopRowFilter:
Filter result for rowKey: 1121874.  Result: false
2008-07-22 13:34:18,292 DEBUG
org.apache.hadoop.hbase.filter.WhileMatchRowFilter: this.filterAllRemaining
is now: false
2008-07-22 13:34:18,292 DEBUG
org.apache.hadoop.hbase.filter.WhileMatchRowFilter: Filter on
rowKey:1121874, colKey: hierarcy:immediate_parents, data: [B@7d5b1773.
Result = false
2008-07-22 13:34:18,324 DEBUG org.apache.hadoop.hbase.filter.StopRowFilter:
Filter result for rowKey: 11218740.  Result: false
2008-07-22 13:34:18,324 DEBUG
org.apache.hadoop.hbase.filter.WhileMatchRowFilter: this.filterAllRemaining
is now: false
2008-07-22 13:34:18,325 DEBUG
org.apache.hadoop.hbase.filter.WhileMatchRowFilter: Filter on
rowKey:11218740.  Result = false
2008-07-22 13:34:18,325 DEBUG org.apache.hadoop.hbase.filter.StopRowFilter:
Filter result for rowKey: 11218740.  Result: false
2008-07-22 13:34:18,325 DEBUG
org.apache.hadoop.hbase.filter.WhileMatchRowFilter: this.filterAllRemaining
is now: false
2008-07-22 13:34:18,325 DEBUG
org.apache.hadoop.hbase.filter.WhileMatchRowFilter: Filter on
rowKey:11218740, colKey: hierarcy:immediate_parents, data: [B@b05236.
Result = false
2008-07-22 13:34:18,334 DEBUG
org.apache.hadoop.hbase.filter.WhileMatchRowFilter: this.filterAllRemaining
is now: false
2008-07-22 13:34:18,334 DEBUG
org.apache.hadoop.hbase.filter.WhileMatchRowFilter: FilterNotNull on
cols:{hierarcy:immediate_parents=[B@7d5b1773}.  Result = false
2008-07-22 13:34:18,744 DEBUG
org.apache.hadoop.hbase.filter.WhileMatchRowFilter: Successfully read a
sub-filter of type: org.apache.hadoop.hbase.filter.StopRowFilter
2008-07-22 13:34:18,745 DEBUG
org.apache.hadoop.hbase.filter.WhileMatchRowFilter: Successfully read a
sub-filter of type: org.apache.hadoop.hbase.filter.StopRowFilter
2008-07-22 13:34:18,745 DEBUG
org.apache.hadoop.hbase.filter.WhileMatchRowFilter: Resetting.
2008-07-22 13:34:18,758 DEBUG org.apache.hadoop.hbase.filter.StopRowFilter:
Filter result for rowKey: 11391639.  Result: false
2008-07-22 13:34:18,758 DEBUG
org.apache.hadoop.hbase.filter.WhileMatchRowFilter: this.filterAllRemaining
is now: false
2008-07-22 13:34:18,758 DEBUG
org.apache.hadoop.hbase.filter.WhileMatchRowFilter: Filter on
rowKey:11391639.  Result = false
2008-07-22 13:34:18,759 DEBUG org.apache.hadoop.hbase.filter.StopRowFilter:
Filter result for rowKey: 11391639.  Result: false
2008-07-22 13:34:18,759 DEBUG
org.apache.hadoop.hbase.filter.WhileMatchRowFilter: this.filterAllRemaining
is now: false
2008-07-22 13:34:18,759 DEBUG
org.apache.hadoop.hbase.filter.WhileMatchRowFilter: Filter on
rowKey:11391639, colKey: hierarcy:immediate_parents, data: [B@273d1402.
Result = false
2008-07-22 13:34:18,779 DEBUG org.apache.hadoop.hbase.filter.StopRowFilter:
Filter result for rowKey: 1139164.  Result: false
2008-07-22 13:34:18,779 DEBUG
org.apache.hadoop.hbase.filter.WhileMatchRowFilter: this.filterAllRemaining
is now: false
2008-07-22 13:34:18,779 DEBUG
org.apache.hadoop.hbase.filter.WhileMatchRowFilter: Filter on
rowKey:1139164.  Result = false
2008-07-22 13:34:18,779 DEBUG org.apache.hadoop.hbase.filter.StopRowFilter:
Filter result for rowKey: 1139164.  Result: false
2008-07-22 13:34:18,779 DEBUG
org.apache.hadoop.hbase.filter.WhileMatchRowFilter: this.filterAllRemaining
is now: false
2008-07-22 13:34:18,779 DEBUG
org.apache.hadoop.hbase.filter.WhileMatchRowFilter: Filter on
rowKey:1139164, colKey: hierarcy:immediate_parents, data: [B@25dd9891.
Result = false
2008-07-22 13:34:18,782 DEBUG
org.apache.hadoop.hbase.filter.WhileMatchRowFilter: this.filterAllRemaining
is now: false
2008-07-22 13:34:18,783 DEBUG
org.apache.hadoop.hbase.filter.WhileMatchRowFilter: FilterNotNull on
cols:{hierarcy:immediate_parents=[B@273d1402}.  Result = false


Then at some point, next just hangs and the next entry in the log is
expiration:
2008-07-22 13:44:18,015 INFO org.apache.hadoop.hbase.Leases:
regionserver/0.0.0.0:60020.leaseChecker lease expired
-4780003109149242775/-4780003109149242775
2008-07-22 13:44:18,015 INFO org.apache.hadoop.hbase.HRegionServer: Scanner
-4780003109149242775 lease expired
2008-07-22 13:44:18,015 INFO org.apache.hadoop.hbase.Leases:
regionserver/0.0.0.0:60020.leaseChecker lease expired
-8543632393752004827/-8543632393752004827
2008-07-22 13:44:18,015 INFO org.apache.hadoop.hbase.HRegionServer: Scanner
-8543632393752004827 lease expired
2008-07-22 13:44:18,015 INFO org.apache.hadoop.hbase.Leases:
regionserver/0.0.0.0:60020.leaseChecker lease expired
-3834469434163616397/-3834469434163616397
2008-07-22 13:44:18,015 INFO org.apache.hadoop.hbase.HRegionServer: Scanner
-3834469434163616397 lease expired
2008-07-22 13:44:28,024 INFO org.apache.hadoop.hbase.Leases:
regionserver/0.0.0.0:60020.leaseChecker lease expired
-6363251876420626143/-6363251876420626143
2008-07-22 13:44:28,024 INFO org.apache.hadoop.hbase.HRegionServer: Scanner
-6363251876420626143 lease expired
2008-07-22 13:44:28,024 INFO org.apache.hadoop.hbase.Leases:
regionserver/0.0.0.0:60020.leaseChecker lease expired
-4332402553799574820/-4332402553799574820
2008-07-22 13:44:28,024 INFO org.apache.hadoop.hbase.HRegionServer: Scanner
-4332402553799574820 lease expired
2008-07-22 13:44:28,024 INFO org.apache.hadoop.hbase.Leases:
regionserver/0.0.0.0:60020.leaseChecker lease expired
7541512102009958038/7541512102009958038
2008-07-22 13:44:28,024 INFO org.apache.hadoop.hbase.HRegionServer: Scanner
7541512102009958038 lease expired
2008-07-22 13:44:28,025 INFO org.apache.hadoop.hbase.Leases:
regionserver/0.0.0.0:60020.leaseChecker lease expired
-6673102874800967750/-6673102874800967750
2008-07-22 13:44:28,025 INFO org.apache.hadoop.hbase.HRegionServer: Scanner
-6673102874800967750 lease expired
2008-07-22 13:44:28,025 INFO org.apache.hadoop.hbase.Leases:
regionserver/0.0.0.0:60020.leaseChecker lease expired
9143807894343107328/9143807894343107328
2008-07-22 13:44:28,025 INFO org.apache.hadoop.hbase.HRegionServer: Scanner
9143807894343107328 lease expired

Now interestingly, the map task doesn't know about the expired lease until a
number of minutes later when it finally tries to call next:
2008-07-22 14:54:29,062 INFO org.apache.hadoop.ipc.Server: IPC Server
handler 8 on 60020, call next(-4780003109149242775) from
10.252.242.159:60900: error:
org.apache.hadoop.hbase.UnknownScannerException: Name: -4780003109149242775
org.apache.hadoop.hbase.UnknownScannerException: Name: -4780003109149242775
        at
org.apache.hadoop.hbase.HRegionServer.next(HRegionServer.java:1506)
        at sun.reflect.GeneratedMethodAccessor3.invoke(Unknown Source)
        at
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
        at java.lang.reflect.Method.invoke(Method.java:597)
        at
org.apache.hadoop.hbase.ipc.HbaseRPC$Server.call(HbaseRPC.java:413)
        at org.apache.hadoop.ipc.Server$Handler.run(Server.java:901)
2008-07-22 14:54:29,070 INFO org.apache.hadoop.ipc.Server: IPC Server
handler 3 on 60020, call close(-4780003109149242775) from
10.252.242.159:60900: error:
org.apache.hadoop.hbase.UnknownScannerException: -4780003109149242775
org.apache.hadoop.hbase.UnknownScannerException: -4780003109149242775
        at
org.apache.hadoop.hbase.HRegionServer.close(HRegionServer.java:1616)
        at sun.reflect.GeneratedMethodAccessor11.invoke(Unknown Source)
        at
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
        at java.lang.reflect.Method.invoke(Method.java:597)
        at
org.apache.hadoop.hbase.ipc.HbaseRPC$Server.call(HbaseRPC.java:413)
        at org.apache.hadoop.ipc.Server$Handler.run(Server.java:901)


Even more interestingly is that the other scanners for the other tasks were
also "frozen" during the time and their leases expire too.

Inside the iteration, it's pretty simple stuff happening - string
manipulation mostly and a put into a bdb - takes <10ms.

The only other suspect piece of code is a reporter.progress() which I had to
put or else the task tracker thinks the task is dead. I've not had trouble
with reporter.progress() to date...

Any thoughts?

On Tue, Jul 22, 2008 at 11:22 AM, stack <stack@duboce.net> wrote:

> Hey D:
>
> TableInputFormat#TableRecordReader#next looks like it lets out
> IOExceptions.  UnknownScannerException subclasses IOE so I'd think it should
> bubble up into your map task.
>
> Paste in the exception you found in your logs.  That might help figure
> whats going on.
>
> St.Ack
>
>
> Daniel Leffel wrote:
>
>> I've had a little bit of weird behavior.
>>
>> I am opening a scanner in the configure method of a Map task to load a
>> simple little in-memory map (I'd love to this with in-memory column
>> stores,
>> but that's another story ;-).
>>
>> Anyway, the scanner iterates over 4 million rows or so and I've verified
>> next is being called multiple times per second. However, at the same
>> record
>> every time (after around 6 minutes of iteration), the scanner throws an
>> UnknownScannerException in the log (I see it in the log), but rather than
>> throw the exception to the Map task, next just returns false and the loop
>> ends (making it exceptionally difficult to detect if the scanner finished
>> scanning or not). Increasing the regionserver timeout has most definitely
>> solved the issue, but shouldn't calling next reset the timeout anyway?
>>
>> Such a frustrating problem!
>>
>>
>>
>
>

Mime
  • Unnamed multipart/alternative (inline, None, 0 bytes)
View raw message