carbondata-user mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From Ravindra Pesala <ravi.pes...@gmail.com>
Subject Re: Re: get error when load tpcds data catalog_returns
Date Thu, 24 Aug 2017 09:51:13 GMT
Hi,

It seems like a bug in 1.1.1 version, can you try out on the latest master
branch once.

Regards,
Ravindra.

On 24 August 2017 at 14:52, lk_hadoop <lk_hadoop@163.com> wrote:

> @Ravindra   carbondata1.1.1  spark2.1.0  yarn 2.7.3 and
> catalog_returns_1_4.dat size is 5.5G
>
> Container: container_1499044620720_0086_01_000007 on XX_57375
> ============================================================
> =====================
> LogType:stderr
> Log Upload Time:Thu Aug 24 16:03:58 +0800 2017
> LogLength:47927
> Log Contents:
> SLF4J: Class path contains multiple SLF4J bindings.
> SLF4J: Found binding in [jar:file:/fs/data10/yarn/
> usercache/hadoop/filecache/209/carbondata.tar.gz/
> carbonlib/carbondata_2.11-1.1.1-shade-hadoop2.7.2.jar!/org/
> slf4j/impl/StaticLoggerBinder.class]
> SLF4J: Found binding in [jar:file:/fs/data11/yarn/
> usercache/hadoop/filecache/210/__spark_libs__
> 5076531087226320409.zip/slf4j-log4j12-1.7.16.jar!/org/slf4j/
> impl/StaticLoggerBinder.class]
> SLF4J: Found binding in [jar:file:/home/hadoop/dmp/
> hadoop-2.7.3/share/hadoop/common/lib/slf4j-log4j12-1.7.
> 10.jar!/org/slf4j/impl/StaticLoggerBinder.class]
> SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an
> explanation.
> SLF4J: Actual binding is of type [org.slf4j.impl.Log4jLoggerFactory]
> 17/08/24 15:57:13 INFO executor.CoarseGrainedExecutorBackend: Started
> daemon with process name: 84197@kafka03
> 17/08/24 15:57:13 INFO util.SignalUtils: Registered signal handler for TERM
> 17/08/24 15:57:13 INFO util.SignalUtils: Registered signal handler for HUP
> 17/08/24 15:57:13 INFO util.SignalUtils: Registered signal handler for INT
> 17/08/24 15:57:13 INFO spark.SecurityManager: Changing view acls to: hadoop
> 17/08/24 15:57:13 INFO spark.SecurityManager: Changing modify acls to:
> hadoop
> 17/08/24 15:57:13 INFO spark.SecurityManager: Changing view acls groups
> to:
> 17/08/24 15:57:13 INFO spark.SecurityManager: Changing modify acls groups
> to:
> 17/08/24 15:57:13 INFO spark.SecurityManager: SecurityManager:
> authentication disabled; ui acls disabled; users  with view permissions:
> Set(hadoop); groups with view permissions: Set(); users  with modify
> permissions: Set(hadoop); groups with modify permissions: Set()
> 17/08/24 15:57:14 INFO client.TransportClientFactory: Successfully created
> connection to /192.168.0.142:51695 after 82 ms (0 ms spent in bootstraps)
> 17/08/24 15:57:14 INFO spark.SecurityManager: Changing view acls to: hadoop
> 17/08/24 15:57:14 INFO spark.SecurityManager: Changing modify acls to:
> hadoop
> 17/08/24 15:57:14 INFO spark.SecurityManager: Changing view acls groups
> to:
> 17/08/24 15:57:14 INFO spark.SecurityManager: Changing modify acls groups
> to:
> 17/08/24 15:57:14 INFO spark.SecurityManager: SecurityManager:
> authentication disabled; ui acls disabled; users  with view permissions:
> Set(hadoop); groups with view permissions: Set(); users  with modify
> permissions: Set(hadoop); groups with modify permissions: Set()
> 17/08/24 15:57:14 INFO client.TransportClientFactory: Successfully created
> connection to /192.168.0.142:51695 after 1 ms (0 ms spent in bootstraps)
> 17/08/24 15:57:14 INFO storage.DiskBlockManager: Created local directory
> at /fs/data10/yarn/usercache/hadoop/appcache/application_
> 1499044620720_0086/blockmgr-ffa868ad-0ce0-43c5-90be-60ae014d8863
> 17/08/24 15:57:14 INFO storage.DiskBlockManager: Created local directory
> at /fs/data11/yarn/usercache/hadoop/appcache/application_
> 1499044620720_0086/blockmgr-2fbb8304-06b4-4843-88be-ee7b497c92be
> 17/08/24 15:57:14 INFO storage.DiskBlockManager: Created local directory
> at /fs/data12/yarn/usercache/hadoop/appcache/application_
> 1499044620720_0086/blockmgr-9c04bab5-d599-4fdc-8a62-d8767037b18a
> 17/08/24 15:57:14 INFO memory.MemoryStore: MemoryStore started with
> capacity 47.8 GB
> 17/08/24 15:57:14 INFO executor.CoarseGrainedExecutorBackend: Connecting
> to driver: spark://CoarseGrainedScheduler@192.168.0.142:51695
> 17/08/24 15:57:14 INFO executor.CoarseGrainedExecutorBackend:
> Successfully registered with driver
> 17/08/24 15:57:14 INFO executor.Executor: Starting executor ID 6 on host
> xxx
> 17/08/24 15:57:14 INFO util.Utils: Successfully started service
> 'org.apache.spark.network.netty.NettyBlockTransferService' on port 60309.
> 17/08/24 15:57:14 INFO netty.NettyBlockTransferService: Server created on
> xxx
> 17/08/24 15:57:14 INFO storage.BlockManager: Using
> org.apache.spark.storage.RandomBlockReplicationPolicy for block
> replication policy
> 17/08/24 15:57:14 INFO storage.BlockManagerMaster: Registering
> BlockManager BlockManagerId(6, xxx, 60309, None)
> 17/08/24 15:57:14 INFO storage.BlockManagerMaster: Registered BlockManager
> BlockManagerId(6, xxx, 60309, None)
> 17/08/24 15:57:14 INFO storage.BlockManager: Initialized BlockManager:
> BlockManagerId(6, xxx, 60309, None)
> 17/08/24 15:57:14 INFO executor.Executor: Using REPL class URI: spark://
> 192.168.0.142:51695/classes
> 17/08/24 15:59:21 INFO executor.CoarseGrainedExecutorBackend: Got
> assigned task 3
> 17/08/24 15:59:21 INFO executor.CoarseGrainedExecutorBackend: Got
> assigned task 9
> 17/08/24 15:59:21 INFO executor.CoarseGrainedExecutorBackend: Got
> assigned task 15
> 17/08/24 15:59:21 INFO executor.CoarseGrainedExecutorBackend: Got
> assigned task 21
> 17/08/24 15:59:21 INFO executor.Executor: Running task 8.0 in stage 0.0
> (TID 9)
> 17/08/24 15:59:21 INFO executor.Executor: Running task 20.0 in stage 0.0
> (TID 21)
> 17/08/24 15:59:21 INFO executor.Executor: Running task 14.0 in stage 0.0
> (TID 15)
> 17/08/24 15:59:21 INFO executor.Executor: Running task 3.0 in stage 0.0
> (TID 3)
> 17/08/24 15:59:21 INFO broadcast.TorrentBroadcast: Started reading
> broadcast variable 1
> 17/08/24 15:59:21 INFO client.TransportClientFactory: Successfully created
> connection to /192.168.0.142:48571 after 1 ms (0 ms spent in bootstraps)
> 17/08/24 15:59:21 INFO memory.MemoryStore: Block broadcast_1_piece0 stored
> as bytes in memory (estimated size 11.0 KB, free 47.8 GB)
> 17/08/24 15:59:21 INFO broadcast.TorrentBroadcast: Reading broadcast
> variable 1 took 108 ms
> 17/08/24 15:59:21 INFO memory.MemoryStore: Block broadcast_1 stored as
> values in memory (estimated size 29.6 KB, free 47.8 GB)
> 17/08/24 15:59:21 INFO util.CarbonProperties: Executor task launch
> worker-0 Property file path: /fs/data10/yarn/usercache/
> hadoop/appcache/application_1499044620720_0086/container_
> 1499044620720_0086_01_000007/carbon.properties
> 17/08/24 15:59:21 INFO util.CarbonProperties: Executor task launch
> worker-0 ------Using Carbon.properties --------
> 17/08/24 15:59:21 INFO util.CarbonProperties: Executor task launch
> worker-0 {carbon.graph.rowset.size=100000, carbon.enable.quick.filter=false,
> carbon.number.of.cores=4, carbon.sort.file.buffer.size=20,
> carbon.number.of.cores.while.compacting=2, carbon.compaction.level.threshold=4,3,
> carbon.lock.type=HDFSLOCK, carbon.number.of.cores.while.loading=6,
> carbon.badRecords.location=/opt/Carbon/Spark/badrecords,
> carbon.sort.size=500000, carbon.inmemory.record.size=120000,
> carbon.enableXXHash=true, carbon.ddl.base.hdfs.url=hdfs://kafka01:9000/opt/data,
> carbon.major.compaction.size=1024, carbon.storelocation=hdfs://
> kafka01:9000/Opt/CarbonStore}
> 17/08/24 15:59:21 INFO util.CarbonProperties: Executor task launch
> worker-0 Carbon Current data file version: V3
> 17/08/24 15:59:21 INFO util.CarbonProperties: Executor task launch
> worker-0 Executor start up wait time: 5
> 17/08/24 15:59:21 INFO util.CarbonProperties: Executor task launch
> worker-0 Blocklet Size Configured value is "64
> 17/08/24 15:59:21 INFO rdd.NewHadoopRDD: Input split:
> hdfs://kafka01:9000/tpcds/source/catalog_returns/catalog_returns_1_4.dat:
> 5368709120+268435456
> 17/08/24 15:59:21 INFO rdd.NewHadoopRDD: Input split:
> hdfs://kafka01:9000/tpcds/source/catalog_returns/catalog_returns_1_4.dat:
> 2147483648+268435456
> 17/08/24 15:59:21 INFO rdd.NewHadoopRDD: Input split:
> hdfs://kafka01:9000/tpcds/source/catalog_returns/catalog_returns_1_4.dat:
> 3758096384+268435456
> 17/08/24 15:59:21 INFO rdd.NewHadoopRDD: Input split:
> hdfs://kafka01:9000/tpcds/source/catalog_returns/catalog_returns_1_4.dat:
> 805306368+268435456
> 17/08/24 15:59:21 INFO broadcast.TorrentBroadcast: Started reading
> broadcast variable 0
> 17/08/24 15:59:21 INFO client.TransportClientFactory: Successfully created
> connection to kafka01.youedata.com/192.168.0.140:37515 after 2 ms (0 ms
> spent in bootstraps)
> 17/08/24 15:59:21 INFO memory.MemoryStore: Block broadcast_0_piece0 stored
> as bytes in memory (estimated size 24.1 KB, free 47.8 GB)
> 17/08/24 15:59:21 INFO broadcast.TorrentBroadcast: Reading broadcast
> variable 0 took 35 ms
> 17/08/24 15:59:21 INFO memory.MemoryStore: Block broadcast_0 stored as
> values in memory (estimated size 424.1 KB, free 47.8 GB)
> 17/08/24 15:59:22 INFO client.TransportClientFactory: Successfully created
> connection to /192.168.0.142:51695 after 2 ms (0 ms spent in bootstraps)
> 17/08/24 15:59:22 INFO codegen.CodeGenerator: Code generated in 310.137373
> ms
> 17/08/24 15:59:22 INFO codegen.CodeGenerator: Code generated in 19.689183
> ms
> 17/08/24 15:59:23 INFO codegen.CodeGenerator: Code generated in 16.974435
> ms
> 17/08/24 15:59:24 INFO codegen.CodeGenerator: Code generated in 174.032802
> ms
> 17/08/24 15:59:39 INFO executor.Executor: Finished task 20.0 in stage 0.0
> (TID 21). 2515 bytes result sent to driver
> 17/08/24 15:59:39 INFO executor.Executor: Finished task 14.0 in stage 0.0
> (TID 15). 1803 bytes result sent to driver
> 17/08/24 15:59:39 INFO executor.Executor: Finished task 3.0 in stage 0.0
> (TID 3). 2515 bytes result sent to driver
> 17/08/24 15:59:39 INFO executor.Executor: Finished task 8.0 in stage 0.0
> (TID 9). 1716 bytes result sent to driver
> 17/08/24 16:01:21 INFO executor.CoarseGrainedExecutorBackend: Got
> assigned task 24
> 17/08/24 16:01:21 INFO executor.Executor: Running task 0.0 in stage 2.0
> (TID 24)
> 17/08/24 16:01:21 INFO spark.MapOutputTrackerWorker: Updating epoch to 1
> and clearing cache
> 17/08/24 16:01:21 INFO broadcast.TorrentBroadcast: Started reading
> broadcast variable 4
> 17/08/24 16:01:21 INFO memory.MemoryStore: Block broadcast_4_piece0 stored
> as bytes in memory (estimated size 8.4 KB, free 47.8 GB)
> 17/08/24 16:01:21 INFO broadcast.TorrentBroadcast: Reading broadcast
> variable 4 took 7 ms
> 17/08/24 16:01:21 INFO memory.MemoryStore: Block broadcast_4 stored as
> values in memory (estimated size 24.9 KB, free 47.8 GB)
> 17/08/24 16:01:21 INFO broadcast.TorrentBroadcast: Started reading
> broadcast variable 3
> 17/08/24 16:01:21 INFO memory.MemoryStore: Block broadcast_3_piece0 stored
> as bytes in memory (estimated size 46.5 KB, free 47.8 GB)
> 17/08/24 16:01:21 INFO broadcast.TorrentBroadcast: Reading broadcast
> variable 3 took 7 ms
> 17/08/24 16:01:21 INFO memory.MemoryStore: Block broadcast_3 stored as
> values in memory (estimated size 842.7 KB, free 47.8 GB)
> 17/08/24 16:01:21 INFO rdd.NewCarbonDataLoadRDD: Input split:
> kafka03.youedata.com
> 17/08/24 16:01:21 INFO rdd.NewCarbonDataLoadRDD: The Block Count in this
> node :28
> 17/08/24 16:01:21 WARN util.CarbonDataProcessorUtil: [Executor task launch
> worker-4][partitionID:default_carbon_catalog_returns2_
> 03f3c47c-cd02-4774-ac7f-d7a2aa97c7ef] sort scope is set to LOCAL_SORT
> 17/08/24 16:01:21 INFO newflow.AbstractDataLoadProcessorStep: Thread-7
> Rows processed in step Input Processor : 0
> 17/08/24 16:01:21 INFO newflow.AbstractDataLoadProcessorStep: Thread-8
> Rows processed in step Data Converter : 0
> 17/08/24 16:01:21 INFO newflow.AbstractDataLoadProcessorStep: Thread-9
> Rows processed in step Sort Processor : 0
> 17/08/24 16:01:21 INFO newflow.AbstractDataLoadProcessorStep: Thread-10
> Rows processed in step Data Writer : 0
> 17/08/24 16:01:22 INFO cache.CacheProvider: [Executor task launch
> worker-4][partitionID:default_carbon_catalog_returns2_
> 03f3c47c-cd02-4774-ac7f-d7a2aa97c7ef] Executor LRU cache size not
> configured. Initializing with driver LRU cache size.
> 17/08/24 16:01:22 INFO cache.CarbonLRUCache: [Executor task launch
> worker-4][partitionID:default_carbon_catalog_returns2_
> 03f3c47c-cd02-4774-ac7f-d7a2aa97c7ef] LRU cache size not configured.
> Therefore default behavior will be considered and no LRU based eviction of
> columns will be done
> 17/08/24 16:01:31 INFO newflow.AbstractDataLoadProcessorStep: Thread-7
> Rows processed in step Input Processor : 0
> 17/08/24 16:01:31 INFO newflow.AbstractDataLoadProcessorStep: Thread-8
> Rows processed in step Data Converter : 0
> 17/08/24 16:01:31 INFO newflow.AbstractDataLoadProcessorStep: Thread-9
> Rows processed in step Sort Processor : 0
> 17/08/24 16:01:31 INFO newflow.AbstractDataLoadProcessorStep: Thread-10
> Rows processed in step Data Writer : 0
> 17/08/24 16:01:36 WARN util.CarbonDataProcessorUtil: [Executor task launch
> worker-4][partitionID:default_carbon_catalog_returns2_
> 03f3c47c-cd02-4774-ac7f-d7a2aa97c7ef] batch sort size is set to 0
> 17/08/24 16:01:36 INFO sortdata.SortParameters: [Executor task launch
> worker-4][partitionID:default_carbon_catalog_returns2_
> 03f3c47c-cd02-4774-ac7f-d7a2aa97c7ef] Sort size for table: 500000
> 17/08/24 16:01:36 INFO sortdata.SortParameters: [Executor task launch
> worker-4][partitionID:default_carbon_catalog_returns2_
> 03f3c47c-cd02-4774-ac7f-d7a2aa97c7ef] Number of intermediate file to be
> merged: 20
> 17/08/24 16:01:36 INFO sortdata.SortParameters: [Executor task launch
> worker-4][partitionID:default_carbon_catalog_returns2_
> 03f3c47c-cd02-4774-ac7f-d7a2aa97c7ef] File Buffer Size: 1048576
> 17/08/24 16:01:36 INFO sortdata.SortParameters: [Executor task launch
> worker-4][partitionID:default_carbon_catalog_returns2_
> 03f3c47c-cd02-4774-ac7f-d7a2aa97c7ef] temp file location/fs/data10/yarn/
> usercache/hadoop/appcache/application_1499044620720_
> 0086/container_1499044620720_0086_01_000007/tmp/
> 4688236258085316/0/default/carbon_catalog_returns2/Fact/
> Part0/Segment_0/0/sortrowtmp
> 17/08/24 16:01:36 WARN util.CarbonDataProcessorUtil: [Executor task launch
> worker-4][partitionID:default_carbon_catalog_returns2_
> 03f3c47c-cd02-4774-ac7f-d7a2aa97c7ef] sort scope is set to LOCAL_SORT
> 17/08/24 16:01:36 INFO newflow.DataLoadExecutor: [Executor task launch
> worker-4][partitionID:default_carbon_catalog_returns2_
> 03f3c47c-cd02-4774-ac7f-d7a2aa97c7ef] Data Loading is started for table
> carbon_catalog_returns2
> 17/08/24 16:01:54 INFO newflow.AbstractDataLoadProcessorStep: Thread-10
> Rows processed in step Data Writer : 0
> 17/08/24 16:01:54 INFO newflow.AbstractDataLoadProcessorStep: Thread-7
> Rows processed in step Input Processor : 489000
> 17/08/24 16:01:54 INFO newflow.AbstractDataLoadProcessorStep: Thread-9
> Rows processed in step Sort Processor : 487000
> 17/08/24 16:01:54 INFO newflow.AbstractDataLoadProcessorStep: Thread-8
> Rows processed in step Data Converter : 487000
> 17/08/24 16:01:57 INFO sortdata.SortDataRows: pool-16-thread-2 Time taken
> to sort and write sort temp file /fs/data10/yarn/usercache/
> hadoop/appcache/application_1499044620720_0086/container_
> 1499044620720_0086_01_000007/tmp/4688236258085316/0/
> default/carbon_catalog_returns2/Fact/Part0/Segment_0/
> 0/sortrowtmp/carbon_catalog_returns24688269866804135.sorttemp is: 3069
> 17/08/24 16:01:57 INFO sortdata.SortDataRows: pool-16-thread-1 Time taken
> to sort and write sort temp file /fs/data10/yarn/usercache/
> hadoop/appcache/application_1499044620720_0086/container_
> 1499044620720_0086_01_000007/tmp/4688236258085316/0/
> default/carbon_catalog_returns2/Fact/Part0/Segment_0/
> 0/sortrowtmp/carbon_catalog_returns24688269560609898.sorttemp is: 3512
> 17/08/24 16:01:58 INFO sortdata.SortDataRows: pool-16-thread-3 Time taken
> to sort and write sort temp file /fs/data10/yarn/usercache/
> hadoop/appcache/application_1499044620720_0086/container_
> 1499044620720_0086_01_000007/tmp/4688236258085316/0/
> default/carbon_catalog_returns2/Fact/Part0/Segment_0/
> 0/sortrowtmp/carbon_catalog_returns24688270205697874.sorttemp is: 2751
> 17/08/24 16:02:00 INFO sortdata.SortDataRows: pool-16-thread-2 Time taken
> to sort and write sort temp file /fs/data10/yarn/usercache/
> hadoop/appcache/application_1499044620720_0086/container_
> 1499044620720_0086_01_000007/tmp/4688236258085316/0/
> default/carbon_catalog_returns2/Fact/Part0/Segment_0/
> 0/sortrowtmp/carbon_catalog_returns24688272980143789.sorttemp is: 2966
> 17/08/24 16:02:01 INFO sortdata.SortDataRows: pool-16-thread-1 Time taken
> to sort and write sort temp file /fs/data10/yarn/usercache/
> hadoop/appcache/application_1499044620720_0086/container_
> 1499044620720_0086_01_000007/tmp/4688236258085316/0/
> default/carbon_catalog_returns2/Fact/Part0/Segment_0/
> 0/sortrowtmp/carbon_catalog_returns24688273622524925.sorttemp is: 2815
> 17/08/24 16:02:02 INFO sortdata.SortDataRows: pool-16-thread-3 Time taken
> to sort and write sort temp file /fs/data10/yarn/usercache/
> hadoop/appcache/application_1499044620720_0086/container_
> 1499044620720_0086_01_000007/tmp/4688236258085316/0/
> default/carbon_catalog_returns2/Fact/Part0/Segment_0/
> 0/sortrowtmp/carbon_catalog_returns24688274076536424.sorttemp is: 2925
> 17/08/24 16:02:04 INFO sortdata.SortDataRows: pool-16-thread-2 Time taken
> to sort and write sort temp file /fs/data10/yarn/usercache/
> hadoop/appcache/application_1499044620720_0086/container_
> 1499044620720_0086_01_000007/tmp/4688236258085316/0/
> default/carbon_catalog_returns2/Fact/Part0/Segment_0/
> 0/sortrowtmp/carbon_catalog_returns24688275840260663.sorttemp is: 3445
> 17/08/24 16:02:04 INFO newflow.AbstractDataLoadProcessorStep: Thread-10
> Rows processed in step Data Writer : 0
> 17/08/24 16:02:04 INFO newflow.AbstractDataLoadProcessorStep: Thread-7
> Rows processed in step Input Processor : 5045000
> 17/08/24 16:02:04 INFO newflow.AbstractDataLoadProcessorStep: Thread-9
> Rows processed in step Sort Processor : 5042000
> 17/08/24 16:02:04 INFO newflow.AbstractDataLoadProcessorStep: Thread-8
> Rows processed in step Data Converter : 5043000
> 17/08/24 16:02:04 INFO sortdata.SortDataRows: pool-16-thread-1 Time taken
> to sort and write sort temp file /fs/data10/yarn/usercache/
> hadoop/appcache/application_1499044620720_0086/container_
> 1499044620720_0086_01_000007/tmp/4688236258085316/0/
> default/carbon_catalog_returns2/Fact/Part0/Segment_0/
> 0/sortrowtmp/carbon_catalog_returns24688276446874268.sorttemp is: 3305
> 17/08/24 16:02:05 INFO sortdata.SortDataRows: pool-16-thread-3 Time taken
> to sort and write sort temp file /fs/data10/yarn/usercache/
> hadoop/appcache/application_1499044620720_0086/container_
> 1499044620720_0086_01_000007/tmp/4688236258085316/0/
> default/carbon_catalog_returns2/Fact/Part0/Segment_0/
> 0/sortrowtmp/carbon_catalog_returns24688277415523789.sorttemp is: 3297
> 17/08/24 16:02:08 INFO sortdata.SortDataRows: pool-16-thread-2 Time taken
> to sort and write sort temp file /fs/data10/yarn/usercache/
> hadoop/appcache/application_1499044620720_0086/container_
> 1499044620720_0086_01_000007/tmp/4688236258085316/0/
> default/carbon_catalog_returns2/Fact/Part0/Segment_0/
> 0/sortrowtmp/carbon_catalog_returns24688279224093804.sorttemp is: 4084
> 17/08/24 16:02:08 INFO sortdata.SortDataRows: pool-16-thread-1 Time taken
> to sort and write sort temp file /fs/data10/yarn/usercache/
> hadoop/appcache/application_1499044620720_0086/container_
> 1499044620720_0086_01_000007/tmp/4688236258085316/0/
> default/carbon_catalog_returns2/Fact/Part0/Segment_0/
> 0/sortrowtmp/carbon_catalog_returns24688279682044167.sorttemp is: 3909
> 17/08/24 16:02:09 INFO sortdata.SortDataRows: pool-16-thread-3 Time taken
> to sort and write sort temp file /fs/data10/yarn/usercache/
> hadoop/appcache/application_1499044620720_0086/container_
> 1499044620720_0086_01_000007/tmp/4688236258085316/0/
> default/carbon_catalog_returns2/Fact/Part0/Segment_0/
> 0/sortrowtmp/carbon_catalog_returns24688280238732307.sorttemp is: 3877
> 17/08/24 16:02:10 INFO sortdata.SortDataRows: pool-16-thread-2 Time taken
> to sort and write sort temp file /fs/data10/yarn/usercache/
> hadoop/appcache/application_1499044620720_0086/container_
> 1499044620720_0086_01_000007/tmp/4688236258085316/0/
> default/carbon_catalog_returns2/Fact/Part0/Segment_0/
> 0/sortrowtmp/carbon_catalog_returns24688283131136222.sorttemp is: 2580
> 17/08/24 16:02:11 INFO sortdata.SortDataRows: pool-16-thread-1 Time taken
> to sort and write sort temp file /fs/data10/yarn/usercache/
> hadoop/appcache/application_1499044620720_0086/container_
> 1499044620720_0086_01_000007/tmp/4688236258085316/0/
> default/carbon_catalog_returns2/Fact/Part0/Segment_0/
> 0/sortrowtmp/carbon_catalog_returns24688283724960116.sorttemp is: 2806
> 17/08/24 16:02:12 INFO sortdata.SortDataRows: pool-16-thread-3 Time taken
> to sort and write sort temp file /fs/data10/yarn/usercache/
> hadoop/appcache/application_1499044620720_0086/container_
> 1499044620720_0086_01_000007/tmp/4688236258085316/0/
> default/carbon_catalog_returns2/Fact/Part0/Segment_0/
> 0/sortrowtmp/carbon_catalog_returns24688284161309296.sorttemp is: 2868
> 17/08/24 16:02:14 INFO newflow.AbstractDataLoadProcessorStep: Thread-10
> Rows processed in step Data Writer : 0
> 17/08/24 16:02:14 INFO newflow.AbstractDataLoadProcessorStep: Thread-7
> Rows processed in step Input Processor : 9505000
> 17/08/24 16:02:14 INFO newflow.AbstractDataLoadProcessorStep: Thread-9
> Rows processed in step Sort Processor : 9499000
> 17/08/24 16:02:14 INFO newflow.AbstractDataLoadProcessorStep: Thread-8
> Rows processed in step Data Converter : 9505000
> 17/08/24 16:02:14 INFO sortdata.SortDataRows: pool-16-thread-2 Time taken
> to sort and write sort temp file /fs/data10/yarn/usercache/
> hadoop/appcache/application_1499044620720_0086/container_
> 1499044620720_0086_01_000007/tmp/4688236258085316/0/
> default/carbon_catalog_returns2/Fact/Part0/Segment_0/
> 0/sortrowtmp/carbon_catalog_returns24688285856830704.sorttemp is: 3782
> 17/08/24 16:02:15 INFO sortdata.SortDataRows: pool-16-thread-1 Time taken
> to sort and write sort temp file /fs/data10/yarn/usercache/
> hadoop/appcache/application_1499044620720_0086/container_
> 1499044620720_0086_01_000007/tmp/4688236258085316/0/
> default/carbon_catalog_returns2/Fact/Part0/Segment_0/
> 0/sortrowtmp/carbon_catalog_returns24688286523938421.sorttemp is: 3657
> 17/08/24 16:02:15 INFO sortdata.SortDataRows: pool-16-thread-3 Time taken
> to sort and write sort temp file /fs/data10/yarn/usercache/
> hadoop/appcache/application_1499044620720_0086/container_
> 1499044620720_0086_01_000007/tmp/4688236258085316/0/
> default/carbon_catalog_returns2/Fact/Part0/Segment_0/
> 0/sortrowtmp/carbon_catalog_returns24688287952674573.sorttemp is: 3760
> 17/08/24 16:02:17 INFO sortdata.SortDataRows: pool-16-thread-2 Time taken
> to sort and write sort temp file /fs/data10/yarn/usercache/
> hadoop/appcache/application_1499044620720_0086/container_
> 1499044620720_0086_01_000007/tmp/4688236258085316/0/
> default/carbon_catalog_returns2/Fact/Part0/Segment_0/
> 0/sortrowtmp/carbon_catalog_returns24688289648034842.sorttemp is: 2804
> 17/08/24 16:02:18 INFO sortdata.SortDataRows: pool-16-thread-1 Time taken
> to sort and write sort temp file /fs/data10/yarn/usercache/
> hadoop/appcache/application_1499044620720_0086/container_
> 1499044620720_0086_01_000007/tmp/4688236258085316/0/
> default/carbon_catalog_returns2/Fact/Part0/Segment_0/
> 0/sortrowtmp/carbon_catalog_returns24688290214867296.sorttemp is: 3682
> 17/08/24 16:02:19 INFO sortdata.SortDataRows: pool-16-thread-3 Time taken
> to sort and write sort temp file /fs/data10/yarn/usercache/
> hadoop/appcache/application_1499044620720_0086/container_
> 1499044620720_0086_01_000007/tmp/4688236258085316/0/
> default/carbon_catalog_returns2/Fact/Part0/Segment_0/
> 0/sortrowtmp/carbon_catalog_returns24688290712710427.sorttemp is: 3387
> 17/08/24 16:02:19 INFO sortdata.SortDataRows: [Executor task launch
> worker-4][partitionID:default_carbon_catalog_returns2_
> 03f3c47c-cd02-4774-ac7f-d7a2aa97c7ef] File based sorting will be used
> 17/08/24 16:02:21 INFO sortdata.SortDataRows: pool-16-thread-2 Time taken
> to sort and write sort temp file /fs/data10/yarn/usercache/
> hadoop/appcache/application_1499044620720_0086/container_
> 1499044620720_0086_01_000007/tmp/4688236258085316/0/
> default/carbon_catalog_returns2/Fact/Part0/Segment_0/
> 0/sortrowtmp/carbon_catalog_returns24688293177154616.sorttemp is: 3535
> 17/08/24 16:02:21 INFO sortdata.SortDataRows: pool-16-thread-1 Time taken
> to sort and write sort temp file /fs/data10/yarn/usercache/
> hadoop/appcache/application_1499044620720_0086/container_
> 1499044620720_0086_01_000007/tmp/4688236258085316/0/
> default/carbon_catalog_returns2/Fact/Part0/Segment_0/
> 0/sortrowtmp/carbon_catalog_returns24688293732092549.sorttemp is: 2696
> 17/08/24 16:02:22 INFO impl.ParallelReadMergeSorterImpl: [Executor task
> launch worker-4][partitionID:default_carbon_catalog_returns2_
> 03f3c47c-cd02-4774-ac7f-d7a2aa97c7ef] Record Processed For table:
> carbon_catalog_returns2
> 17/08/24 16:02:22 INFO store.SingleThreadFinalSortFilesMerger: [Executor
> task launch worker-4][partitionID:default_carbon_catalog_returns2_
> 03f3c47c-cd02-4774-ac7f-d7a2aa97c7ef] Number of temp file: 24
> 17/08/24 16:02:22 INFO store.SingleThreadFinalSortFilesMerger: [Executor
> task launch worker-4][partitionID:default_carbon_catalog_returns2_
> 03f3c47c-cd02-4774-ac7f-d7a2aa97c7ef] File Buffer Size: 873813
> 17/08/24 16:02:22 INFO store.SingleThreadFinalSortFilesMerger: [Executor
> task launch worker-4][partitionID:default_carbon_catalog_returns2_
> 03f3c47c-cd02-4774-ac7f-d7a2aa97c7ef] Started adding first record from
> each file
> 17/08/24 16:02:22 INFO store.SingleThreadFinalSortFilesMerger: [Executor
> task launch worker-4][partitionID:default_carbon_catalog_returns2_
> 03f3c47c-cd02-4774-ac7f-d7a2aa97c7ef] Heap Size24
> 17/08/24 16:02:22 INFO store.CarbonFactDataHandlerColumnar: [Executor
> task launch worker-4][partitionID:default_carbon_catalog_returns2_
> 03f3c47c-cd02-4774-ac7f-d7a2aa97c7ef] Initializing writer executors
> 17/08/24 16:02:22 INFO store.CarbonFactDataHandlerColumnar: [Executor
> task launch worker-4][partitionID:default_carbon_catalog_returns2_
> 03f3c47c-cd02-4774-ac7f-d7a2aa97c7ef] Number of rows per column blocklet
> 32000
> 17/08/24 16:02:22 INFO writer.AbstractFactDataWriter: [Executor task
> launch worker-4][partitionID:default_carbon_catalog_returns2_
> 03f3c47c-cd02-4774-ac7f-d7a2aa97c7ef] Total file size: 1073741824 and
> dataBlock Size: 966367642
> 17/08/24 16:02:22 INFO store.CarbonFactDataHandlerColumnar: [Executor
> task launch worker-4][partitionID:default_carbon_catalog_returns2_
> 03f3c47c-cd02-4774-ac7f-d7a2aa97c7ef] Total Number Of records added to
> store: 32000
> 17/08/24 16:02:22 INFO store.CarbonFactDataHandlerColumnar: [Executor
> task launch worker-4][partitionID:default_carbon_catalog_returns2_
> 03f3c47c-cd02-4774-ac7f-d7a2aa97c7ef] Total Number Of records added to
> store: 64000
> 17/08/24 16:02:22 INFO store.CarbonFactDataHandlerColumnar: [Executor
> task launch worker-4][partitionID:default_carbon_catalog_returns2_
> 03f3c47c-cd02-4774-ac7f-d7a2aa97c7ef] Total Number Of records added to
> store: 96000
> 17/08/24 16:02:22 INFO store.CarbonFactDataHandlerColumnar: [Executor
> task launch worker-4][partitionID:default_carbon_catalog_returns2_
> 03f3c47c-cd02-4774-ac7f-d7a2aa97c7ef] Total Number Of records added to
> store: 128000
> 17/08/24 16:02:22 INFO store.CarbonFactDataHandlerColumnar: [Executor
> task launch worker-4][partitionID:default_carbon_catalog_returns2_
> 03f3c47c-cd02-4774-ac7f-d7a2aa97c7ef] Total Number Of records added to
> store: 160000
> 17/08/24 16:02:22 INFO store.CarbonFactDataHandlerColumnar: [Executor
> task launch worker-4][partitionID:default_carbon_catalog_returns2_
> 03f3c47c-cd02-4774-ac7f-d7a2aa97c7ef] Total Number Of records added to
> store: 192000
> 17/08/24 16:02:22 INFO store.CarbonFactDataHandlerColumnar:
> pool-43-thread-5 Number Of records processed: 32000
> 17/08/24 16:02:22 INFO store.CarbonFactDataHandlerColumnar:
> pool-43-thread-6 Number Of records processed: 32000
> 17/08/24 16:02:22 INFO store.CarbonFactDataHandlerColumnar:
> pool-43-thread-4 Number Of records processed: 32000
> 17/08/24 16:02:22 INFO store.CarbonFactDataHandlerColumnar:
> pool-43-thread-3 Number Of records processed: 32000
> 17/08/24 16:02:22 INFO store.CarbonFactDataHandlerColumnar:
> pool-43-thread-1 Number Of records processed: 32000
> 17/08/24 16:02:22 INFO store.CarbonFactDataHandlerColumnar: [Executor
> task launch worker-4][partitionID:default_carbon_catalog_returns2_
> 03f3c47c-cd02-4774-ac7f-d7a2aa97c7ef] Total Number Of records added to
> store: 224000
> 17/08/24 16:02:22 INFO store.CarbonFactDataHandlerColumnar:
> pool-43-thread-2 Number Of records processed: 32000
> 17/08/24 16:02:22 INFO store.CarbonFactDataHandlerColumnar: [Executor
> task launch worker-4][partitionID:default_carbon_catalog_returns2_
> 03f3c47c-cd02-4774-ac7f-d7a2aa97c7ef] Total Number Of records added to
> store: 256000
> 17/08/24 16:02:23 INFO store.CarbonFactDataHandlerColumnar: [Executor
> task launch worker-4][partitionID:default_carbon_catalog_returns2_
> 03f3c47c-cd02-4774-ac7f-d7a2aa97c7ef] Total Number Of records added to
> store: 288000
> 17/08/24 16:02:23 INFO store.CarbonFactDataHandlerColumnar: [Executor
> task launch worker-4][partitionID:default_carbon_catalog_returns2_
> 03f3c47c-cd02-4774-ac7f-d7a2aa97c7ef] Total Number Of records added to
> store: 320000
> 17/08/24 16:02:23 INFO store.CarbonFactDataHandlerColumnar: [Executor
> task launch worker-4][partitionID:default_carbon_catalog_returns2_
> 03f3c47c-cd02-4774-ac7f-d7a2aa97c7ef] Total Number Of records added to
> store: 352000
> 17/08/24 16:02:23 INFO store.CarbonFactDataHandlerColumnar: [Executor
> task launch worker-4][partitionID:default_carbon_catalog_returns2_
> 03f3c47c-cd02-4774-ac7f-d7a2aa97c7ef] Total Number Of records added to
> store: 384000
> 17/08/24 16:02:23 INFO store.CarbonFactDataHandlerColumnar:
> pool-43-thread-4 Number Of records processed: 32000
> 17/08/24 16:02:23 INFO store.CarbonFactDataHandlerColumnar:
> pool-43-thread-3 Number Of records processed: 32000
> 17/08/24 16:02:23 INFO store.CarbonFactDataHandlerColumnar:
> pool-43-thread-1 Number Of records processed: 32000
> 17/08/24 16:02:23 INFO store.CarbonFactDataHandlerColumnar:
> pool-43-thread-5 Number Of records processed: 32000
> 17/08/24 16:02:23 INFO store.CarbonFactDataHandlerColumnar: [Executor
> task launch worker-4][partitionID:default_carbon_catalog_returns2_
> 03f3c47c-cd02-4774-ac7f-d7a2aa97c7ef] Total Number Of records added to
> store: 416000
> 17/08/24 16:02:23 INFO store.CarbonFactDataHandlerColumnar:
> pool-43-thread-6 Number Of records processed: 32000
> 17/08/24 16:02:23 INFO store.CarbonFactDataHandlerColumnar:
> pool-43-thread-2 Number Of records processed: 32000
> 17/08/24 16:02:23 INFO store.CarbonFactDataHandlerColumnar: [Executor
> task launch worker-4][partitionID:default_carbon_catalog_returns2_
> 03f3c47c-cd02-4774-ac7f-d7a2aa97c7ef] Total Number Of records added to
> store: 448000
> 17/08/24 16:02:23 INFO store.CarbonFactDataHandlerColumnar: [Executor
> task launch worker-4][partitionID:default_carbon_catalog_returns2_
> 03f3c47c-cd02-4774-ac7f-d7a2aa97c7ef] Total Number Of records added to
> store: 480000
> 17/08/24 16:02:23 INFO store.CarbonFactDataHandlerColumnar: [Executor
> task launch worker-4][partitionID:default_carbon_catalog_returns2_
> 03f3c47c-cd02-4774-ac7f-d7a2aa97c7ef] Total Number Of records added to
> store: 512000
> 17/08/24 16:02:23 INFO store.CarbonFactDataHandlerColumnar: [Executor
> task launch worker-4][partitionID:default_carbon_catalog_returns2_
> 03f3c47c-cd02-4774-ac7f-d7a2aa97c7ef] Total Number Of records added to
> store: 544000
> 17/08/24 16:02:23 INFO store.CarbonFactDataHandlerColumnar: [Executor
> task launch worker-4][partitionID:default_carbon_catalog_returns2_
> 03f3c47c-cd02-4774-ac7f-d7a2aa97c7ef] Total Number Of records added to
> store: 576000
> 17/08/24 16:02:23 INFO store.CarbonFactDataHandlerColumnar:
> pool-43-thread-4 Number Of records processed: 32000
> 17/08/24 16:02:23 INFO store.CarbonFactDataHandlerColumnar: [Executor
> task launch worker-4][partitionID:default_carbon_catalog_returns2_
> 03f3c47c-cd02-4774-ac7f-d7a2aa97c7ef] Total Number Of records added to
> store: 608000
> 17/08/24 16:02:23 INFO store.CarbonFactDataHandlerColumnar:
> pool-43-thread-3 Number Of records processed: 32000
> 17/08/24 16:02:23 INFO store.CarbonFactDataHandlerColumnar: [Executor
> task launch worker-4][partitionID:default_carbon_catalog_returns2_
> 03f3c47c-cd02-4774-ac7f-d7a2aa97c7ef] Total Number Of records added to
> store: 640000
> 17/08/24 16:02:23 INFO store.CarbonFactDataHandlerColumnar:
> pool-43-thread-1 Number Of records processed: 32000
> 17/08/24 16:02:23 INFO store.CarbonFactDataHandlerColumnar:
> pool-43-thread-5 Number Of records processed: 32000
> 17/08/24 16:02:23 INFO store.CarbonFactDataHandlerColumnar: [Executor
> task launch worker-4][partitionID:default_carbon_catalog_returns2_
> 03f3c47c-cd02-4774-ac7f-d7a2aa97c7ef] Total Number Of records added to
> store: 672000
> 17/08/24 16:02:23 INFO store.CarbonFactDataHandlerColumnar:
> pool-43-thread-6 Number Of records processed: 32000
> 17/08/24 16:02:23 INFO store.CarbonFactDataHandlerColumnar: [Executor
> task launch worker-4][partitionID:default_carbon_catalog_returns2_
> 03f3c47c-cd02-4774-ac7f-d7a2aa97c7ef] Total Number Of records added to
> store: 704000
> 17/08/24 16:02:23 INFO store.CarbonFactDataHandlerColumnar: [Executor
> task launch worker-4][partitionID:default_carbon_catalog_returns2_
> 03f3c47c-cd02-4774-ac7f-d7a2aa97c7ef] Total Number Of records added to
> store: 736000
> 17/08/24 16:02:23 INFO store.CarbonFactDataHandlerColumnar:
> pool-43-thread-2 Number Of records processed: 32000
> 17/08/24 16:02:23 INFO store.CarbonFactDataHandlerColumnar: [Executor
> task launch worker-4][partitionID:default_carbon_catalog_returns2_
> 03f3c47c-cd02-4774-ac7f-d7a2aa97c7ef] Total Number Of records added to
> store: 768000
> 17/08/24 16:02:23 INFO store.CarbonFactDataHandlerColumnar:
> pool-43-thread-4 Number Of records processed: 32000
> 17/08/24 16:02:23 INFO store.CarbonFactDataHandlerColumnar: [Executor
> task launch worker-4][partitionID:default_carbon_catalog_returns2_
> 03f3c47c-cd02-4774-ac7f-d7a2aa97c7ef] Total Number Of records added to
> store: 800000
> 17/08/24 16:02:23 INFO store.CarbonFactDataHandlerColumnar:
> pool-43-thread-3 Number Of records processed: 32000
> 17/08/24 16:02:23 INFO v3.CarbonFactDataWriterImplV3: pool-44-thread-1
> Number of Pages for blocklet is: 19 :Rows Added: 608000
> 17/08/24 16:02:23 INFO store.CarbonFactDataHandlerColumnar:
> pool-43-thread-1 Number Of records processed: 32000
> 17/08/24 16:02:23 INFO store.CarbonFactDataHandlerColumnar:
> pool-43-thread-5 Number Of records processed: 32000
> 17/08/24 16:02:23 INFO store.CarbonFactDataHandlerColumnar:
> pool-43-thread-6 Number Of records processed: 32000
> 17/08/24 16:02:23 INFO store.CarbonFactDataHandlerColumnar:
> pool-43-thread-2 Number Of records processed: 32000
> 17/08/24 16:02:23 ERROR store.CarbonFactDataHandlerColumnar:
> pool-44-thread-1 Problem while writing the carbon data file
> java.nio.BufferUnderflowException
>  at java.nio.Buffer.nextGetIndex(Buffer.java:506)
>  at java.nio.HeapByteBuffer.getLong(HeapByteBuffer.java:412)
>  at org.apache.carbondata.core.util.CarbonMetadataUtil.compareMeasureData(
> CarbonMetadataUtil.java:899)
>  at org.apache.carbondata.core.util.CarbonMetadataUtil.getBlockletIndex(
> CarbonMetadataUtil.java:286)
>  at org.apache.carbondata.processing.store.writer.v3.
> CarbonFactDataWriterImplV3.writeDataToFile(CarbonFactDataWriterImplV3.
> java:508)
>  at org.apache.carbondata.processing.store.writer.v3.
> CarbonFactDataWriterImplV3.writeDataToFile(CarbonFactDataWriterImplV3.
> java:396)
>  at org.apache.carbondata.processing.store.writer.v3.
> CarbonFactDataWriterImplV3.writeBlockletData(CarbonFactDataWriterImplV3.
> java:345)
>  at org.apache.carbondata.processing.store.CarbonFactDataHandlerColumnar$
> Consumer.call(CarbonFactDataHandlerColumnar.java:1400)
>  at org.apache.carbondata.processing.store.CarbonFactDataHandlerColumnar$
> Consumer.call(CarbonFactDataHandlerColumnar.java:1380)
>  at java.util.concurrent.FutureTask.run(FutureTask.java:266)
>  at java.util.concurrent.ThreadPoolExecutor.runWorker(
> ThreadPoolExecutor.java:1142)
>  at java.util.concurrent.ThreadPoolExecutor$Worker.run(
> ThreadPoolExecutor.java:617)
>  at java.lang.Thread.run(Thread.java:745)
> 17/08/24 16:02:23 ERROR store.CarbonFactDataHandlerColumnar:
> pool-43-thread-4 null
> java.lang.InterruptedException
>  at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.
> reportInterruptAfterWait(AbstractQueuedSynchronizer.java:2014)
>  at java.util.concurrent.locks.AbstractQueuedSynchronizer$
> ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2088)
>  at java.util.concurrent.ThreadPoolExecutor.awaitTermination(
> ThreadPoolExecutor.java:1465)
>  at org.apache.carbondata.processing.store.CarbonFactDataHandlerColumnar.
> createNodeHolderObjectWithOutKettle(CarbonFactDataHandlerColumnar.
> java:811)
>  at org.apache.carbondata.processing.store.CarbonFactDataHandlerColumnar.
> processDataRows(CarbonFactDataHandlerColumnar.java:648)
>  at org.apache.carbondata.processing.store.CarbonFactDataHandlerColumnar.
> access$1100(CarbonFactDataHandlerColumnar.java:86)
>  at org.apache.carbondata.processing.store.CarbonFactDataHandlerColumnar$
> Producer.call(CarbonFactDataHandlerColumnar.java:1362)
>  at org.apache.carbondata.processing.store.CarbonFactDataHandlerColumnar$
> Producer.call(CarbonFactDataHandlerColumnar.java:1339)
>  at java.util.concurrent.FutureTask.run(FutureTask.java:266)
>  at java.util.concurrent.ThreadPoolExecutor.runWorker(
> ThreadPoolExecutor.java:1142)
>  at java.util.concurrent.ThreadPoolExecutor$Worker.run(
> ThreadPoolExecutor.java:617)
>  at java.lang.Thread.run(Thread.java:745)
> 17/08/24 16:02:23 ERROR steps.DataWriterProcessorStepImpl: [Executor task
> launch worker-4][partitionID:default_carbon_catalog_returns2_
> 03f3c47c-cd02-4774-ac7f-d7a2aa97c7ef] Failed for table:
> carbon_catalog_returns2 in DataWriterProcessorStepImpl
> org.apache.carbondata.processing.newflow.exception.CarbonDataLoadingException:
> unable to generate the mdkey
>  at org.apache.carbondata.processing.newflow.steps.
> DataWriterProcessorStepImpl.processBatch(DataWriterProcessorStepImpl.
> java:181)
>  at org.apache.carbondata.processing.newflow.steps.
> DataWriterProcessorStepImpl.execute(DataWriterProcessorStepImpl.java:111)
>  at org.apache.carbondata.processing.newflow.DataLoadExecutor.execute(
> DataLoadExecutor.java:48)
>  at org.apache.carbondata.spark.rdd.NewCarbonDataLoadRDD$$anon$1.<init>(
> NewCarbonDataLoadRDD.scala:243)
>  at org.apache.carbondata.spark.rdd.NewCarbonDataLoadRDD.
> compute(NewCarbonDataLoadRDD.scala:220)
>  at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:323)
>  at org.apache.spark.rdd.RDD.iterator(RDD.scala:287)
>  at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:87)
>  at org.apache.spark.scheduler.Task.run(Task.scala:99)
>  at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:282)
>  at java.util.concurrent.ThreadPoolExecutor.runWorker(
> ThreadPoolExecutor.java:1142)
>  at java.util.concurrent.ThreadPoolExecutor$Worker.run(
> ThreadPoolExecutor.java:617)
>  at java.lang.Thread.run(Thread.java:745)
> Caused by: java.util.concurrent.RejectedExecutionException: Task
> java.util.concurrent.FutureTask@67098e0 rejected from
> java.util.concurrent.ThreadPoolExecutor@5b91b608[Shutting down, pool size
> = 1, active threads = 1, queued tasks = 0, completed tasks = 24]
>  at java.util.concurrent.ThreadPoolExecutor$AbortPolicy.rejectedExecution(
> ThreadPoolExecutor.java:2047)
>  at java.util.concurrent.ThreadPoolExecutor.reject(
> ThreadPoolExecutor.java:823)
>  at java.util.concurrent.ThreadPoolExecutor.execute(
> ThreadPoolExecutor.java:1369)
>  at java.util.concurrent.AbstractExecutorService.submit(
> AbstractExecutorService.java:134)
>  at org.apache.carbondata.processing.store.CarbonFactDataHandlerColumnar.
> addDataToStore(CarbonFactDataHandlerColumnar.java:466)
>  at org.apache.carbondata.processing.newflow.steps.
> DataWriterProcessorStepImpl.processBatch(DataWriterProcessorStepImpl.
> java:178)
>  ... 12 more
> 17/08/24 16:02:23 INFO newflow.AbstractDataLoadProcessorStep: [Executor
> task launch worker-4][partitionID:default_carbon_catalog_returns2_
> 03f3c47c-cd02-4774-ac7f-d7a2aa97c7ef] Total rows processed in step Data
> Writer: 831000
> 17/08/24 16:02:23 INFO newflow.AbstractDataLoadProcessorStep: [Executor
> task launch worker-4][partitionID:default_carbon_catalog_returns2_
> 03f3c47c-cd02-4774-ac7f-d7a2aa97c7ef] Total rows processed in step Sort
> Processor: 11946858
> 17/08/24 16:02:23 INFO newflow.AbstractDataLoadProcessorStep: [Executor
> task launch worker-4][partitionID:default_carbon_catalog_returns2_
> 03f3c47c-cd02-4774-ac7f-d7a2aa97c7ef] Total rows processed in step Data
> Converter: 11946858
> 17/08/24 16:02:23 INFO newflow.AbstractDataLoadProcessorStep: [Executor
> task launch worker-4][partitionID:default_carbon_catalog_returns2_
> 03f3c47c-cd02-4774-ac7f-d7a2aa97c7ef] Total rows processed in step Input
> Processor: 11946858
> 17/08/24 16:02:23 INFO rdd.NewCarbonDataLoadRDD: DataLoad failure
>
> org.apache.carbondata.processing.newflow.exception.CarbonDataLoadingException:
> There is an unexpected error: unable to generate the mdkey
>  at org.apache.carbondata.processing.newflow.steps.
> DataWriterProcessorStepImpl.execute(DataWriterProcessorStepImpl.java:125)
>  at org.apache.carbondata.processing.newflow.DataLoadExecutor.execute(
> DataLoadExecutor.java:48)
>  at org.apache.carbondata.spark.rdd.NewCarbonDataLoadRDD$$anon$1.<init>(
> NewCarbonDataLoadRDD.scala:243)
>  at org.apache.carbondata.spark.rdd.NewCarbonDataLoadRDD.
> compute(NewCarbonDataLoadRDD.scala:220)
>  at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:323)
>  at org.apache.spark.rdd.RDD.iterator(RDD.scala:287)
>  at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:87)
>  at org.apache.spark.scheduler.Task.run(Task.scala:99)
>  at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:282)
>  at java.util.concurrent.ThreadPoolExecutor.runWorker(
> ThreadPoolExecutor.java:1142)
>  at java.util.concurrent.ThreadPoolExecutor$Worker.run(
> ThreadPoolExecutor.java:617)
>  at java.lang.Thread.run(Thread.java:745)
> Caused by: org.apache.carbondata.processing.newflow.exception.CarbonDataLoadingException:
> unable to generate the mdkey
>  at org.apache.carbondata.processing.newflow.steps.
> DataWriterProcessorStepImpl.processBatch(DataWriterProcessorStepImpl.
> java:181)
>  at org.apache.carbondata.processing.newflow.steps.
> DataWriterProcessorStepImpl.execute(DataWriterProcessorStepImpl.java:111)
>  ... 11 more
> Caused by: java.util.concurrent.RejectedExecutionException: Task
> java.util.concurrent.FutureTask@67098e0 rejected from
> java.util.concurrent.ThreadPoolExecutor@5b91b608[Shutting down, pool size
> = 1, active threads = 1, queued tasks = 0, completed tasks = 24]
>  at java.util.concurrent.ThreadPoolExecutor$AbortPolicy.rejectedExecution(
> ThreadPoolExecutor.java:2047)
>  at java.util.concurrent.ThreadPoolExecutor.reject(
> ThreadPoolExecutor.java:823)
>  at java.util.concurrent.ThreadPoolExecutor.execute(
> ThreadPoolExecutor.java:1369)
>  at java.util.concurrent.AbstractExecutorService.submit(
> AbstractExecutorService.java:134)
>  at org.apache.carbondata.processing.store.CarbonFactDataHandlerColumnar.
> addDataToStore(CarbonFactDataHandlerColumnar.java:466)
>  at org.apache.carbondata.processing.newflow.steps.
> DataWriterProcessorStepImpl.processBatch(DataWriterProcessorStepImpl.
> java:178)
>  ... 12 more
> 17/08/24 16:02:23 ERROR rdd.NewCarbonDataLoadRDD: [Executor task launch
> worker-4][partitionID:default_carbon_catalog_returns2_
> 03f3c47c-cd02-4774-ac7f-d7a2aa97c7ef]
> org.apache.carbondata.processing.newflow.exception.CarbonDataLoadingException:
> There is an unexpected error: unable to generate the mdkey
>  at org.apache.carbondata.processing.newflow.steps.
> DataWriterProcessorStepImpl.execute(DataWriterProcessorStepImpl.java:125)
>  at org.apache.carbondata.processing.newflow.DataLoadExecutor.execute(
> DataLoadExecutor.java:48)
>  at org.apache.carbondata.spark.rdd.NewCarbonDataLoadRDD$$anon$1.<init>(
> NewCarbonDataLoadRDD.scala:243)
>  at org.apache.carbondata.spark.rdd.NewCarbonDataLoadRDD.
> compute(NewCarbonDataLoadRDD.scala:220)
>  at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:323)
>  at org.apache.spark.rdd.RDD.iterator(RDD.scala:287)
>  at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:87)
>  at org.apache.spark.scheduler.Task.run(Task.scala:99)
>  at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:282)
>  at java.util.concurrent.ThreadPoolExecutor.runWorker(
> ThreadPoolExecutor.java:1142)
>  at java.util.concurrent.ThreadPoolExecutor$Worker.run(
> ThreadPoolExecutor.java:617)
>  at java.lang.Thread.run(Thread.java:745)
> Caused by: org.apache.carbondata.processing.newflow.exception.CarbonDataLoadingException:
> unable to generate the mdkey
>  at org.apache.carbondata.processing.newflow.steps.
> DataWriterProcessorStepImpl.processBatch(DataWriterProcessorStepImpl.
> java:181)
>  at org.apache.carbondata.processing.newflow.steps.
> DataWriterProcessorStepImpl.execute(DataWriterProcessorStepImpl.java:111)
>  ... 11 more
> Caused by: java.util.concurrent.RejectedExecutionException: Task
> java.util.concurrent.FutureTask@67098e0 rejected from
> java.util.concurrent.ThreadPoolExecutor@5b91b608[Shutting down, pool size
> = 1, active threads = 1, queued tasks = 0, completed tasks = 24]
>  at java.util.concurrent.ThreadPoolExecutor$AbortPolicy.rejectedExecution(
> ThreadPoolExecutor.java:2047)
>  at java.util.concurrent.ThreadPoolExecutor.reject(
> ThreadPoolExecutor.java:823)
>  at java.util.concurrent.ThreadPoolExecutor.execute(
> ThreadPoolExecutor.java:1369)
>  at java.util.concurrent.AbstractExecutorService.submit(
> AbstractExecutorService.java:134)
>  at org.apache.carbondata.processing.store.CarbonFactDataHandlerColumnar.
> addDataToStore(CarbonFactDataHandlerColumnar.java:466)
>  at org.apache.carbondata.processing.newflow.steps.
> DataWriterProcessorStepImpl.processBatch(DataWriterProcessorStepImpl.
> java:178)
>  ... 12 more
> 17/08/24 16:02:23 INFO load.CarbonLoaderUtil: pool-71-thread-1 Deleted the
> local store location/fs/data10/yarn/usercache/hadoop/appcache/
> application_1499044620720_0086/container_1499044620720_0086_01_000007/tmp/4688236258085316/0
> : TIme taken: 3
> 17/08/24 16:02:23 INFO store.CarbonFactDataHandlerColumnar:
> pool-43-thread-4 Number Of records processed: 32000
> 17/08/24 16:02:23 ERROR executor.Executor: Exception in task 0.0 in stage
> 2.0 (TID 24)
>
> org.apache.carbondata.processing.newflow.exception.CarbonDataLoadingException:
> There is an unexpected error: unable to generate the mdkey
>  at org.apache.carbondata.processing.newflow.steps.
> DataWriterProcessorStepImpl.execute(DataWriterProcessorStepImpl.java:125)
>  at org.apache.carbondata.processing.newflow.DataLoadExecutor.execute(
> DataLoadExecutor.java:48)
>  at org.apache.carbondata.spark.rdd.NewCarbonDataLoadRDD$$anon$1.<init>(
> NewCarbonDataLoadRDD.scala:243)
>  at org.apache.carbondata.spark.rdd.NewCarbonDataLoadRDD.
> compute(NewCarbonDataLoadRDD.scala:220)
>  at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:323)
>  at org.apache.spark.rdd.RDD.iterator(RDD.scala:287)
>  at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:87)
>  at org.apache.spark.scheduler.Task.run(Task.scala:99)
>  at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:282)
>  at java.util.concurrent.ThreadPoolExecutor.runWorker(
> ThreadPoolExecutor.java:1142)
>  at java.util.concurrent.ThreadPoolExecutor$Worker.run(
> ThreadPoolExecutor.java:617)
>  at java.lang.Thread.run(Thread.java:745)
> Caused by: org.apache.carbondata.processing.newflow.exception.CarbonDataLoadingException:
> unable to generate the mdkey
>  at org.apache.carbondata.processing.newflow.steps.
> DataWriterProcessorStepImpl.processBatch(DataWriterProcessorStepImpl.
> java:181)
>  at org.apache.carbondata.processing.newflow.steps.
> DataWriterProcessorStepImpl.execute(DataWriterProcessorStepImpl.java:111)
>  ... 11 more
> Caused by: java.util.concurrent.RejectedExecutionException: Task
> java.util.concurrent.FutureTask@67098e0 rejected from
> java.util.concurrent.ThreadPoolExecutor@5b91b608[Shutting down, pool size
> = 1, active threads = 1, queued tasks = 0, completed tasks = 24]
>  at java.util.concurrent.ThreadPoolExecutor$AbortPolicy.rejectedExecution(
> ThreadPoolExecutor.java:2047)
>  at java.util.concurrent.ThreadPoolExecutor.reject(
> ThreadPoolExecutor.java:823)
>  at java.util.concurrent.ThreadPoolExecutor.execute(
> ThreadPoolExecutor.java:1369)
>  at java.util.concurrent.AbstractExecutorService.submit(
> AbstractExecutorService.java:134)
>  at org.apache.carbondata.processing.store.CarbonFactDataHandlerColumnar.
> addDataToStore(CarbonFactDataHandlerColumnar.java:466)
>  at org.apache.carbondata.processing.newflow.steps.
> DataWriterProcessorStepImpl.processBatch(DataWriterProcessorStepImpl.
> java:178)
>  ... 12 more
> 17/08/24 16:03:57 INFO executor.CoarseGrainedExecutorBackend: Driver
> commanded a shutdown
> 17/08/24 16:03:57 INFO executor.CoarseGrainedExecutorBackend: Driver from
> 192.168.0.142:51695 disconnected during shutdown
> 17/08/24 16:03:57 INFO executor.CoarseGrainedExecutorBackend: Driver from
> 192.168.0.142:51695 disconnected during shutdown
> 17/08/24 16:03:57 INFO memory.MemoryStore: MemoryStore cleared
> 17/08/24 16:03:57 INFO storage.BlockManager: BlockManager stopped
> 17/08/24 16:03:57 INFO util.ShutdownHookManager: Shutdown hook called
> End of LogType:stderr
>
> LogType:stdout
> Log Upload Time:Thu Aug 24 16:03:58 +0800 2017
> LogLength:0
> Log Contents:
> End of LogType:stdout
> 2017-08-24
> ------------------------------
> lk_hadoop
> ------------------------------
>
> *发件人:*Ravindra Pesala <ravi.pesala@gmail.com>
> *发送时间:*2017-08-24 22:34
> *主题:*Re: get error when load tpcds data catalog_returns
> *收件人:*"user"<user@carbondata.apache.org>
> *抄送:*
>
> Hi,
>
> Which version of carbon and spark are you using? How much data are you
> loading and what is the machine configuration?
>
> I have tried loading catlog_returns with 20 MB data in my local machine
> and it is successful. I used the latest master branch and spark-2.1
> version.
>
> Also please send the complete log as the log information you provided does
> not say the actual cause.
>
> Regards,
> Ravindra.
>
> On 24 August 2017 at 14:02, lk_hadoop <lk_hadoop@163.com> wrote:
>
>> hi,all
>>        I want to test carbondata by using tpc-ds data , I try to load
>> table : catalog_returns
>>        I got error :
>>        org.apache.carbondata.processing.newflow.exception.CarbonDataLoadingException:
>> There is an unexpected error: unable to generate the mdkey
>>  at org.apache.carbondata.processing.newflow.steps.DataWriterPro
>> cessorStepImpl.execute(DataWriterProcessorStepImpl.java:125)
>>  at org.apache.carbondata.processing.newflow.DataLoadExecutor.
>> execute(DataLoadExecutor.java:48)
>>  at org.apache.carbondata.spark.rdd.NewCarbonDataLoadRDD$$anon$
>> 1.<init>(NewCarbonDataLoadRDD.scala:243)
>>  at org.apache.carbondata.spark.rdd.NewCarbonDataLoadRDD.compute
>> (NewCarbonDataLoadRDD.scala:220)
>>  at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:323)
>>  at org.apache.spark.rdd.RDD.iterator(RDD.scala:287)
>>  at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:87)
>>  at org.apache.spark.scheduler.Task.run(Task.scala:99)
>>  at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:282)
>>  at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPool
>> Executor.java:1142)
>>  at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoo
>> lExecutor.java:617)
>>  at java.lang.Thread.run(Thread.java:745)
>> Caused by: org.apache.carbondata.processing.newflow.exception.CarbonDataLoadingException:
>> unable to generate the mdkey
>>  at org.apache.carbondata.processing.newflow.steps.DataWriterPro
>> cessorStepImpl.processBatch(DataWriterProcessorStepImpl.java:181)
>>  at org.apache.carbondata.processing.newflow.steps.DataWriterPro
>> cessorStepImpl.execute(DataWriterProcessorStepImpl.java:111)
>>  ... 11 more
>> Caused by: java.util.concurrent.RejectedExecutionException: Task
>> java.util.concurrent.FutureTask@67098e0 rejected from
>> java.util.concurrent.ThreadPoolExecutor@5b91b608[Shutting down, pool
>> size = 1, active threads = 1, queued tasks = 0, completed tasks = 24]
>>  at java.util.concurrent.ThreadPoolExecutor$AbortPolicy.
>> rejectedExecution(ThreadPoolExecutor.java:2047)
>>  at java.util.concurrent.ThreadPoolExecutor.reject(ThreadPoolExe
>> cutor.java:823)
>>  at java.util.concurrent.ThreadPoolExecutor.execute(ThreadPoolEx
>> ecutor.java:1369)
>>  at java.util.concurrent.AbstractExecutorService.submit(Abstract
>> ExecutorService.java:134)
>>  at org.apache.carbondata.processing.store.CarbonFactDataHandler
>> Columnar.addDataToStore(CarbonFactDataHandlerColumnar.java:466)
>>  at org.apache.carbondata.processing.newflow.steps.DataWriterPro
>> cessorStepImpl.processBatch(DataWriterProcessorStepImpl.java:178)
>>  ... 12 more
>>
>> my table creat with sql :
>> create table if not exists tpcds_carbon.catalog_returns
>> (
>>     cr_returned_date_sk       int                       ,
>>     cr_returned_time_sk       int                       ,
>>     cr_item_sk                int               ,
>>     cr_refunded_customer_sk   int                       ,
>>     cr_refunded_cdemo_sk      int                       ,
>>     cr_refunded_hdemo_sk      int                       ,
>>     cr_refunded_addr_sk       int                       ,
>>     cr_returning_customer_sk  int                       ,
>>     cr_returning_cdemo_sk     int                       ,
>>     cr_returning_hdemo_sk     int                       ,
>>     cr_returning_addr_sk      int                       ,
>>     cr_call_center_sk         int                       ,
>>     cr_catalog_page_sk        int                       ,
>>     cr_ship_mode_sk           int                       ,
>>     cr_warehouse_sk           int                       ,
>>     cr_reason_sk              int                       ,
>>     cr_order_number           int               ,
>>     cr_return_quantity        int                       ,
>>     cr_return_amount          decimal(7,2)                  ,
>>     cr_return_tax             decimal(7,2)                  ,
>>     cr_return_amt_inc_tax     decimal(7,2)                  ,
>>     cr_fee                    decimal(7,2)                  ,
>>     cr_return_ship_cost       decimal(7,2)                  ,
>>     cr_refunded_cash          decimal(7,2)                  ,
>>     cr_reversed_charge        decimal(7,2)                  ,
>>     cr_store_credit           decimal(7,2)                  ,
>>     cr_net_loss               decimal(7,2)
>> )STORED BY 'carbondata'
>> TBLPROPERTIES ('DICTIONARY_INCLUDE'='cr_item_sk,cr_order_number')
>> //because these two cols are the real PK
>>
>> and I load data with :
>>
>> carbon.sql("load data inpath 'hdfs://AAA:9000/tpcds/source/
>> catalog_returns/catalog_returns_1_4.dat' into table
>> carbon_catalog_returns2 OPTIONS('DELIMITER'='|','fileh
>> eader'='cr_returned_date_sk,cr_returned_time_sk,cr_item_
>> sk,cr_refunded_customer_sk,cr_refunded_cdemo_sk,cr_refunded_
>> hdemo_sk,cr_refunded_addr_sk,cr_returning_customer_sk,cr_
>> returning_cdemo_sk,cr_returning_hdemo_sk,cr_returning_addr_
>> sk,cr_call_center_sk,cr_catalog_page_sk,cr_ship_mode_
>> sk,cr_warehouse_sk,cr_reason_sk,cr_order_number,cr_return_
>> quantity,cr_return_amount,cr_return_tax,cr_return_amt_inc_
>> tax,cr_fee,cr_return_ship_cost,cr_refunded_cash,cr_
>> reversed_charge,cr_store_credit,cr_net_loss')")
>>
>> any one know what was wrong?
>>
>>
>>
>>
>> 2017-08-24
>> ------------------------------
>> lk_hadoop
>>
>
>
>
> --
> Thanks & Regards,
> Ravi
>
>


-- 
Thanks & Regards,
Ravi

Mime
View raw message