ignite-user mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From percent620 <percent...@163.com>
Subject Yarn deployment for memory capacity make a bigger than before: Urgent!!!
Date Wed, 24 Aug 2016 14:28:49 GMT
Here is my detailed steps
1)root@sparkup1 config]# cat default-config.xml
<?xml version="1.0" encoding="UTF-8"?>
<beans xmlns="http://www.springframework.org/schema/beans"
       xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance"
xmlns:util="http://www.springframework.org/schema/util"
       xsi:schemaLocation="http://www.springframework.org/schema/beans
       http://www.springframework.org/schema/beans/spring-beans-4.1.xsd">
<bean class="org.apache.ignite.configuration.IgniteConfiguration">
                    <property name="discoverySpi">
                         <bean
class="org.apache.ignite.spi.discovery.tcp.TcpDiscoverySpi">
                                   <property name="ipFinder">
                                           <bean
class="org.apache.ignite.spi.discovery.tcp.ipfinder.vm.TcpDiscoveryVmIpFinder">       
                          
                                                <property name="addresses">
                                                   <list>
                                                       
<value>172.16.186.200:47500..47509</value>
                                                       
<value>172.16.186.201:47500..47509</value>
                                                       
<value>172.16.186.202:47500..47509</value>
                                                   </list>
                                                </property>
                                           </bean>
                                    </property>
                         </bean>
                   </property>
          </bean>
</beans>
[root@sparkup1 config]# 

2)yarn contains log as below

i started 3 ignite node and for every is 1024G 


$cat cluster.properties
# The number of nodes in the cluster.
IGNITE_NODE_COUNT=3
# The number of CPU Cores for each Apache Ignite node.
IGNITE_RUN_CPU_PER_NODE=1
# The number of Megabytes of RAM for each Apache Ignite node.
IGNITE_MEMORY_PER_NODE=1024
# The version ignite which will be run on nodes.
IGNITE_VERSION=1.0.6
IGNITE_WORK_DIR=/u01/yueyi/apache-ignite-hadoop-1.6.0-bin
IGNITE_XML_CONFIG=/ignite/releases/apache-ignite-hadoop-1.6.0-bin/config/default-config.xml
IGNITE_RELEASES_DIR=/ignite/releases/
IGNITE_USERS_LIBS=/ignite/releases/apache-ignite-hadoop-1.6.0-bin/libs/
#IGNITE_HOSTNAME_CONSTRAINT=vmsecdomain010194070026.cm10
IGNITE_PATH=/ignite/releases/


[root@sparkup3 config]# tail -f
/usr/hadoop-2.4.1/logs/userlogs/application_1472047995043_0001/container_1472047995043_0001_01_000003/stdout
[07:13:45] Configured plugins:
[07:13:45]   ^-- None
[07:13:45]
[07:13:46] Security status [authentication=off, tls/ssl=off]
[07:13:47] To start Console Management & Monitoring run
ignitevisorcmd.{sh|bat}
[07:13:47]
[07:13:47] Ignite node started OK (id=20fb73be)
[07:13:47] Topology snapshot [ver=1, servers=1, clients=0, CPUs=1,
heap=1.0GB]
[07:13:48] Topology snapshot [ver=2, servers=2, clients=0, CPUs=2,
heap=2.0GB]
[07:13:50] Topology snapshot [ver=3, servers=3, clients=0, CPUs=3,
heap=3.0GB]
==========================================
Thanks is ok for above steps



3)
spark-submit *--driver-memory 4G* --class com.ignite.testIgniteSharedRDD
--master yarn --executor-cores 2 --executor-memory 1000m --num-executors 2
--conf spark.rdd.compress=false --conf spark.shuffle.compress=false --conf
spark.broadcast.compress=false
/root/limu/ignite/spark-project-jar-with-dependencies.jar


4)Yarn logs become is 
[07:13:46] Security status [authentication=off, tls/ssl=off]
[07:13:47] To start Console Management & Monitoring run
ignitevisorcmd.{sh|bat}
[07:13:47]
[07:13:47] Ignite node started OK (id=20fb73be)
[07:13:47] Topology snapshot [ver=1, servers=1, clients=0, CPUs=1,
heap=1.0GB]
[07:13:48] Topology snapshot [ver=2, servers=2, clients=0, CPUs=2,
heap=2.0GB]
[07:13:50] Topology snapshot [ver=3, servers=3, clients=0, CPUs=3,
heap=3.0GB]
*[07:16:54] Topology snapshot [ver=4, servers=3, clients=1, CPUs=3,
heap=7.0GB] correct

*
/[07:17:06] Topology snapshot [ver=5, servers=4, clients=1, CPUs=3,
heap=8.0GB]
[07:17:07] Topology snapshot [ver=6, servers=5, clients=1, CPUs=3,
heap=9.0GB]/

is not correct, why become 5 servers and 9GB memory



details log 
nohup: ignoring input
16/08/24 07:16:17 INFO spark.SparkContext: Running Spark version 1.6.1
16/08/24 07:16:18 WARN util.NativeCodeLoader: Unable to load native-hadoop
library for your platform... using builtin-java classes where applicable
16/08/24 07:16:18 INFO spark.SecurityManager: Changing view acls to: root
16/08/24 07:16:18 INFO spark.SecurityManager: Changing modify acls to: root
16/08/24 07:16:18 INFO spark.SecurityManager: SecurityManager:
authentication disabled; ui acls disabled; users with view permissions:
Set(root); users with modify permissions: Set(root)
16/08/24 07:16:19 INFO util.Utils: Successfully started service
'sparkDriver' on port 56368.
16/08/24 07:16:20 INFO slf4j.Slf4jLogger: Slf4jLogger started
16/08/24 07:16:20 INFO Remoting: Starting remoting
16/08/24 07:16:20 INFO Remoting: Remoting started; listening on addresses
:[akka.tcp://sparkDriverActorSystem@172.16.186.200:45375]
16/08/24 07:16:20 INFO util.Utils: Successfully started service
'sparkDriverActorSystem' on port 45375.
16/08/24 07:16:20 INFO spark.SparkEnv: Registering MapOutputTracker
16/08/24 07:16:20 INFO spark.SparkEnv: Registering BlockManagerMaster
16/08/24 07:16:20 INFO storage.DiskBlockManager: Created local directory at
/tmp/blockmgr-8739bc15-9e06-4496-91fc-beea794a57d7
16/08/24 07:16:20 INFO storage.MemoryStore: MemoryStore started with
capacity 2.7 GB
16/08/24 07:16:21 INFO spark.SparkEnv: Registering OutputCommitCoordinator
16/08/24 07:16:21 INFO server.Server: jetty-8.y.z-SNAPSHOT
16/08/24 07:16:21 INFO server.AbstractConnector: Started
SelectChannelConnector@0.0.0.0:4040
16/08/24 07:16:21 INFO util.Utils: Successfully started service 'SparkUI' on
port 4040.
16/08/24 07:16:21 INFO ui.SparkUI: Started SparkUI at
http://172.16.186.200:4040
16/08/24 07:16:21 INFO spark.HttpFileServer: HTTP File server directory is
/tmp/spark-07a3f373-5d80-42b0-8544-33583e9fff7d/httpd-4d4119d2-ac26-4f49-90a9-03b88e24f9d9
16/08/24 07:16:21 INFO spark.HttpServer: Starting HTTP Server
16/08/24 07:16:21 INFO server.Server: jetty-8.y.z-SNAPSHOT
16/08/24 07:16:21 INFO server.AbstractConnector: Started
SocketConnector@0.0.0.0:38541
16/08/24 07:16:21 INFO util.Utils: Successfully started service 'HTTP file
server' on port 38541.
16/08/24 07:16:22 INFO spark.SparkContext: Added JAR
file:/root/limu/ignite/spark-project-jar-with-dependencies.jar at
http://172.16.186.200:38541/jars/spark-project-jar-with-dependencies.jar
with timestamp 1472048182938
16/08/24 07:16:23 INFO client.RMProxy: Connecting to ResourceManager at
sparkup1/172.16.186.200:8032
16/08/24 07:16:23 INFO yarn.Client: Requesting a new application from
cluster with 3 NodeManagers
16/08/24 07:16:23 INFO yarn.Client: Verifying our application has not
requested more than the maximum memory capability of the cluster (8192 MB
per container)
16/08/24 07:16:23 INFO yarn.Client: Will allocate AM container, with 896 MB
memory including 384 MB overhead
16/08/24 07:16:23 INFO yarn.Client: Setting up container launch context for
our AM
16/08/24 07:16:23 INFO yarn.Client: Setting up the launch environment for
our AM container
16/08/24 07:16:23 INFO yarn.Client: Preparing resources for our AM container
16/08/24 07:16:24 INFO yarn.Client: Uploading resource
file:/usr/spark-1.6.1-bin-hadoop2.6/lib/spark-assembly-1.6.1-hadoop2.6.0.jar
->
hdfs://sparkup1:9000/user/root/.sparkStaging/application_1472047995043_0002/spark-assembly-1.6.1-hadoop2.6.0.jar
16/08/24 07:16:31 INFO yarn.Client: Uploading resource
file:/tmp/spark-07a3f373-5d80-42b0-8544-33583e9fff7d/__spark_conf__9140885938827716798.zip
->
hdfs://sparkup1:9000/user/root/.sparkStaging/application_1472047995043_0002/__spark_conf__9140885938827716798.zip
16/08/24 07:16:31 INFO spark.SecurityManager: Changing view acls to: root
16/08/24 07:16:31 INFO spark.SecurityManager: Changing modify acls to: root
16/08/24 07:16:31 INFO spark.SecurityManager: SecurityManager:
authentication disabled; ui acls disabled; users with view permissions:
Set(root); users with modify permissions: Set(root)
16/08/24 07:16:31 INFO yarn.Client: Submitting application 2 to
ResourceManager
16/08/24 07:16:31 INFO impl.YarnClientImpl: Submitted application
application_1472047995043_0002
16/08/24 07:16:32 INFO yarn.Client: Application report for
application_1472047995043_0002 (state: ACCEPTED)
16/08/24 07:16:32 INFO yarn.Client:
     client token: N/A
     diagnostics: N/A
     ApplicationMaster host: N/A
     ApplicationMaster RPC port: -1
     queue: default
     start time: 1472048191273
     final status: UNDEFINED
     tracking URL:
http://sparkup1:8088/proxy/application_1472047995043_0002/
     user: root
16/08/24 07:16:33 INFO yarn.Client: Application report for
application_1472047995043_0002 (state: ACCEPTED)
16/08/24 07:16:34 INFO yarn.Client: Application report for
application_1472047995043_0002 (state: ACCEPTED)
16/08/24 07:16:35 INFO yarn.Client: Application report for
application_1472047995043_0002 (state: ACCEPTED)
16/08/24 07:16:36 INFO yarn.Client: Application report for
application_1472047995043_0002 (state: ACCEPTED)
16/08/24 07:16:37 INFO yarn.Client: Application report for
application_1472047995043_0002 (state: ACCEPTED)
16/08/24 07:16:38 INFO cluster.YarnSchedulerBackend$YarnSchedulerEndpoint:
ApplicationMaster registered as NettyRpcEndpointRef(null)
16/08/24 07:16:38 INFO cluster.YarnClientSchedulerBackend: Add WebUI Filter.
org.apache.hadoop.yarn.server.webproxy.amfilter.AmIpFilter, Map(PROXY_HOSTS
-> sparkup1, PROXY_URI_BASES ->
http://sparkup1:8088/proxy/application_1472047995043_0002),
/proxy/application_1472047995043_0002
16/08/24 07:16:38 INFO ui.JettyUtils: Adding filter:
org.apache.hadoop.yarn.server.webproxy.amfilter.AmIpFilter
16/08/24 07:16:38 INFO yarn.Client: Application report for
application_1472047995043_0002 (state: RUNNING)
16/08/24 07:16:38 INFO yarn.Client:
     client token: N/A
     diagnostics: N/A
     ApplicationMaster host: 172.16.186.201
     ApplicationMaster RPC port: 0
     queue: default
     start time: 1472048191273
     final status: UNDEFINED
     tracking URL:
http://sparkup1:8088/proxy/application_1472047995043_0002/
     user: root
16/08/24 07:16:38 INFO cluster.YarnClientSchedulerBackend: Application
application_1472047995043_0002 has started running.
16/08/24 07:16:38 INFO util.Utils: Successfully started service
'org.apache.spark.network.netty.NettyBlockTransferService' on port 41610.
16/08/24 07:16:38 INFO netty.NettyBlockTransferService: Server created on
41610
16/08/24 07:16:38 INFO storage.BlockManagerMaster: Trying to register
BlockManager
16/08/24 07:16:38 INFO storage.BlockManagerMasterEndpoint: Registering block
manager 172.16.186.200:41610 with 2.7 GB RAM, BlockManagerId(driver,
172.16.186.200, 41610)
16/08/24 07:16:38 INFO storage.BlockManagerMaster: Registered BlockManager
16/08/24 07:16:46 INFO cluster.YarnClientSchedulerBackend: Registered
executor NettyRpcEndpointRef(null) (sparkup2:56516) with ID 1
16/08/24 07:16:46 INFO storage.BlockManagerMasterEndpoint: Registering block
manager sparkup2:43778 with 500.0 MB RAM, BlockManagerId(1, sparkup2, 43778)
16/08/24 07:16:47 INFO cluster.YarnClientSchedulerBackend: Registered
executor NettyRpcEndpointRef(null) (sparkup1:46131) with ID 2
16/08/24 07:16:47 INFO cluster.YarnClientSchedulerBackend: SchedulerBackend
is ready for scheduling beginning after reached minRegisteredResourcesRatio:
0.8
16/08/24 07:16:47 INFO storage.BlockManagerMasterEndpoint: Registering block
manager sparkup1:35739 with 500.0 MB RAM, BlockManagerId(2, sparkup1, 35739)
16/08/24 07:16:48 INFO xml.XmlBeanDefinitionReader: Loading XML bean
definitions from URL
[file:/usr/apache-ignite-fabric-1.6.0-bin/config/default-config.xml]
16/08/24 07:16:52 INFO support.GenericApplicationContext: Refreshing
org.springframework.context.support.GenericApplicationContext@467048e:
startup date [Wed Aug 24 07:16:52 PDT 2016]; root of context hierarchy
16/08/24 07:16:53 INFO internal.IgniteKernal:

>>>    __________  ________________  
>>>   /  _/ ___/ |/ /  _/_  __/ __/  
>>>  _/ // (7 7    // /  / / / _/    
>>> /___/\___/_/|_/___/ /_/ /___/   
>>>
>>> ver. 1.6.0#20160518-sha1:0b22c45b
>>> 2016 Copyright(C) Apache Software Foundation
>>>
>>> Ignite documentation: http://ignite.apache.org

16/08/24 07:16:53 INFO internal.IgniteKernal: Config URL: n/a
16/08/24 07:16:53 INFO internal.IgniteKernal: Daemon mode: off
16/08/24 07:16:53 INFO internal.IgniteKernal: OS: Linux
2.6.32-431.el6.x86_64 amd64
16/08/24 07:16:53 INFO internal.IgniteKernal: OS user: root
16/08/24 07:16:53 INFO internal.IgniteKernal: Language runtime: Java
Platform API Specification ver. 1.7
16/08/24 07:16:53 INFO internal.IgniteKernal: VM information: Java(TM) SE
Runtime Environment 1.7.0_71-b14 Oracle Corporation Java HotSpot(TM) 64-Bit
Server VM 24.71-b01
16/08/24 07:16:53 INFO internal.IgniteKernal: VM total memory: 3.9GB
16/08/24 07:16:53 INFO internal.IgniteKernal: Remote Management [restart:
off, REST: on, JMX (remote: off)]
16/08/24 07:16:53 INFO internal.IgniteKernal:
IGNITE_HOME=/usr/apache-ignite-fabric-1.6.0-bin
16/08/24 07:16:53 INFO internal.IgniteKernal: VM arguments: [-Xms4G, -Xmx4G,
-XX:MaxPermSize=256m]
16/08/24 07:16:53 INFO internal.IgniteKernal: Configured caches
['ignite-marshaller-sys-cache', 'ignite-sys-cache',
'ignite-atomics-sys-cache']
16/08/24 07:16:53 INFO internal.IgniteKernal: 3-rd party licenses can be
found at: /usr/apache-ignite-fabric-1.6.0-bin/libs/licenses
16/08/24 07:16:53 INFO internal.IgniteKernal: Non-loopback local IPs:
172.16.186.200, fe80:0:0:0:20c:29ff:fecd:2f25%2
16/08/24 07:16:53 INFO internal.IgniteKernal: Enabled local MACs:
000C29CD2F25
16/08/24 07:16:53 INFO plugin.IgnitePluginProcessor: Configured plugins:
16/08/24 07:16:53 INFO plugin.IgnitePluginProcessor:   ^-- None
16/08/24 07:16:53 INFO plugin.IgnitePluginProcessor:
16/08/24 07:16:53 INFO tcp.TcpCommunicationSpi: IPC shared memory server
endpoint started [port=48101,
tokDir=/usr/apache-ignite-fabric-1.6.0-bin/work/ipc/shmem/529a5fe1-5f95-4aeb-a0dc-a8f814f63369-25100]
16/08/24 07:16:53 INFO tcp.TcpCommunicationSpi: Successfully bound shared
memory communication to TCP port [port=48101, locHost=0.0.0.0/0.0.0.0]
16/08/24 07:16:53 INFO tcp.TcpCommunicationSpi: Successfully bound to TCP
port [port=47101, locHost=0.0.0.0/0.0.0.0]
16/08/24 07:16:53 WARN noop.NoopCheckpointSpi: Checkpoints are disabled (to
enable configure any GridCheckpointSpi implementation)
16/08/24 07:16:53 WARN collision.GridCollisionManager: Collision resolution
is disabled (all jobs will be activated upon arrival).
16/08/24 07:16:53 WARN noop.NoopSwapSpaceSpi: Swap space is disabled. To
enable use FileSwapSpaceSpi.
16/08/24 07:16:53 INFO internal.IgniteKernal: Security status
[authentication=off, tls/ssl=off]
16/08/24 07:16:54 INFO tcp.GridTcpRestProtocol: Command protocol
successfully started [name=TCP binary, host=0.0.0.0/0.0.0.0, port=11212]
16/08/24 07:16:54 WARN internal.IgniteKernal: Attempting to start more nodes
than physical RAM available on current host (this can cause significant
slowdown)
16/08/24 07:16:54 INFO cache.GridCacheProcessor: Started cache
[name=ignite-marshaller-sys-cache, mode=REPLICATED]
16/08/24 07:16:54 INFO cache.GridCacheProcessor: Started cache
[name=ignite-atomics-sys-cache, mode=PARTITIONED]
16/08/24 07:16:54 INFO cache.GridCacheProcessor: Started cache
[name=ignite-sys-cache, mode=REPLICATED]
16/08/24 07:16:55 INFO internal.IgniteKernal: To start Console Management &
Monitoring run ignitevisorcmd.{sh|bat}
16/08/24 07:16:55 INFO internal.IgniteKernal:
16/08/24 07:16:55 INFO internal.IgniteKernal:

>>> +----------------------------------------------------------------------+
>>> Ignite ver. 1.6.0#20160518-sha1:0b22c45bb9b97692208fd0705ddf8045ff34a031
>>> +----------------------------------------------------------------------+
>>> OS name: Linux 2.6.32-431.el6.x86_64 amd64
>>> CPU(s): 1
>>> Heap: 4.0GB
>>> VM name: 25100@sparkup1
>>> Local node [ID=529A5FE1-5F95-4AEB-A0DC-A8F814F63369, order=4,
>>> clientMode=true]
>>> Local node addresses: [sparkup1/0:0:0:0:0:0:0:1%1, /127.0.0.1,
>>> /172.16.186.200]
>>> Local ports: TCP:11212 TCP:47101 TCP:48101

16/08/24 07:16:55 INFO discovery.GridDiscoveryManager: Topology snapshot
[ver=4, servers=3, clients=1, CPUs=3, heap=7.0GB]
16/08/24 07:16:56 INFO cache.GridCacheProcessor: Started cache
[name=sharedIgniteRDD-ling-sha111o, mode=PARTITIONED]
16/08/24 07:16:57 INFO spark.SparkContext: Starting job: count at
testIgniteSharedRDD.scala:19
16/08/24 07:16:57 INFO scheduler.DAGScheduler: Got job 0 (count at
testIgniteSharedRDD.scala:19) with 1024 output partitions
16/08/24 07:16:57 INFO scheduler.DAGScheduler: Final stage: ResultStage 0
(count at testIgniteSharedRDD.scala:19)
16/08/24 07:16:57 INFO scheduler.DAGScheduler: Parents of final stage:
List()
16/08/24 07:16:57 INFO scheduler.DAGScheduler: Missing parents: List()
16/08/24 07:16:57 INFO scheduler.DAGScheduler: Submitting ResultStage 0
(IgniteRDD[0] at RDD at IgniteAbstractRDD.scala:31), which has no missing
parents
16/08/24 07:16:57 INFO storage.MemoryStore: Block broadcast_0 stored as
values in memory (estimated size 1872.0 B, free 1872.0 B)
16/08/24 07:16:57 INFO storage.MemoryStore: Block broadcast_0_piece0 stored
as bytes in memory (estimated size 1853.0 B, free 3.6 KB)
16/08/24 07:16:57 INFO storage.BlockManagerInfo: Added broadcast_0_piece0 in
memory on 172.16.186.200:41610 (size: 1853.0 B, free: 2.7 GB)
16/08/24 07:16:57 INFO spark.SparkContext: Created broadcast 0 from
broadcast at DAGScheduler.scala:1006
16/08/24 07:16:57 INFO scheduler.DAGScheduler: Submitting 1024 missing tasks
from ResultStage 0 (IgniteRDD[0] at RDD at IgniteAbstractRDD.scala:31)
16/08/24 07:16:57 INFO cluster.YarnScheduler: Adding task set 0.0 with 1024
tasks
16/08/24 07:16:57 INFO scheduler.TaskSetManager: Starting task 7.0 in stage
0.0 (TID 0, sparkup2, partition 7,NODE_LOCAL, 1967 bytes)
16/08/24 07:16:57 INFO scheduler.TaskSetManager: Starting task 1.0 in stage
0.0 (TID 1, sparkup1, partition 1,NODE_LOCAL, 1967 bytes)
16/08/24 07:16:57 INFO scheduler.TaskSetManager: Starting task 8.0 in stage
0.0 (TID 2, sparkup2, partition 8,NODE_LOCAL, 1967 bytes)
16/08/24 07:16:57 INFO scheduler.TaskSetManager: Starting task 3.0 in stage
0.0 (TID 3, sparkup1, partition 3,NODE_LOCAL, 1967 bytes)
16/08/24 07:17:00 INFO storage.BlockManagerInfo: Added broadcast_0_piece0 in
memory on sparkup1:35739 (size: 1853.0 B, free: 500.0 MB)
16/08/24 07:17:01 INFO storage.BlockManagerInfo: Added broadcast_0_piece0 in
memory on sparkup2:43778 (size: 1853.0 B, free: 500.0 MB)
[07:17:04] New version is available at ignite.apache.org: 1.7.0
16/08/24 07:17:04 WARN cluster.GridUpdateNotifier: New version is available
at ignite.apache.org: 1.7.0
*16/08/24 07:17:06 INFO discovery.GridDiscoveryManager: Added new node to
topology: TcpDiscoveryNode [id=18575d1b-d764-43ca-9fcb-2f130b159557,
addrs=[0:0:0:0:0:0:0:1%1, 127.0.0.1, 172.16.186.201],
sockAddrs=[sparkup2/172.16.186.201:47501, /0:0:0:0:0:0:0:1%1:47501,
/127.0.0.1:47501, /172.16.186.201:47501], discPort=47501, order=5,
intOrder=5, lastExchangeTime=1472048225977, loc=false,
ver=1.6.0#20160518-sha1:0b22c45b, isClient=false]
16/08/24 07:17:06 INFO discovery.GridDiscoveryManager: Topology snapshot
[ver=5, servers=4, clients=1, CPUs=3, heap=8.0GB]
16/08/24 07:17:07 INFO discovery.GridDiscoveryManager: Added new node to
topology: TcpDiscoveryNode [id=a8fbf4cd-7fe3-40d6-ad64-b1c53479785f,
addrs=[0:0:0:0:0:0:0:1%1, 127.0.0.1, 172.16.186.200],
sockAddrs=[sparkup1/172.16.186.200:47501, /0:0:0:0:0:0:0:1%1:47501,
/127.0.0.1:47501, /172.16.186.200:47501], discPort=47501, order=6,
intOrder=6, lastExchangeTime=1472048226549, loc=false,
ver=1.6.0#20160518-sha1:0b22c45b, isClient=false]
16/08/24 07:17:07 INFO discovery.GridDiscoveryManager: Topology snapshot
[ver=6, servers=5, clients=1, CPUs=3, heap=9.0GB]*

*why local port 127.0.0.1 join yarn ignite topology, is this correct? I
think that this is not correct.
*
16/08/24 07:17:11 INFO scheduler.TaskSetManager: Starting task 17.0 in stage
0.0 (TID 4, sparkup2, partition 17,NODE_LOCAL, 1967 bytes)
16/08/24 07:17:11 INFO scheduler.TaskSetManager: Starting task 18.0 in stage
0.0 (TID 5, sparkup2, partition 18,NODE_LOCAL, 1967 bytes)
16/08/24 07:17:11 INFO scheduler.TaskSetManager: Finished task 8.0 in stage
0.0 (TID 2) in 13294 ms on sparkup2 (1/1024)
16/08/24 07:17:11 INFO scheduler.TaskSetManager: Finished task 7.0 in stage
0.0 (TID 0) in 13364 ms on sparkup2 (2/1024)
16/08/24 07:17:11 INFO scheduler.TaskSetManager: Starting task 4.0 in stage
0.0 (TID 6, sparkup1, partition 4,NODE_LOCAL, 1967 bytes)
16/08/24 07:17:11 INFO scheduler.TaskSetManager: Starting task 6.0 in stage
0.0 (TID 7, sparkup1, partition 6,NODE_LOCAL, 1967 bytes)
16/08/24 07:17:11 INFO scheduler.TaskSetManager: Finished task 1.0 in stage
0.0 (TID 1) in 13763 ms on sparkup1 (3/1024)
16/08/24 07:17:11 INFO scheduler.TaskSetManager: Finished task 3.0 in stage
0.0 (TID 3) in 13759 ms on sparkup1 (4/1024)
16/08/24 07:17:12 INFO scheduler.TaskSetManager: Starting task 20.0 in stage
0.0 (TID 8, sparkup2, partition 20,NODE_LOCAL, 



--
View this message in context: http://apache-ignite-users.70518.x6.nabble.com/Yarn-deployment-for-memory-capacity-make-a-bigger-than-before-Urgent-tp7275.html
Sent from the Apache Ignite Users mailing list archive at Nabble.com.

Mime
View raw message