Home > other >  The Spark run gets stuck, help you a great god
The Spark run gets stuck, help you a great god

Time:09-18

My cluster configuration for 7, with 5 machine is 8 gb of memory, the other two for the virtual machine,
After don't write programs through the spark - submit submit it, you can run successfully, but today when to run there was a problem, the problem is as follows:
17/03/26 10:10:32 INFO BlockManagerInfo: Added broadcast_3_piece0 in memory on 196.168.168.104:59612 (size: 119.0 B, free: 3.0 GB)
17/03/26 10:10:32 INFO BlockManagerInfo: Added broadcast_1730_piece0 in memory on 196.168.168.104:59612 (size: 119.0 B, free: 3.0 GB)
17/03/26 10:10:32 INFO BlockManagerInfo: Added broadcast_1732_piece0 in memory on 196.168.168.104:59612 (size: 119.0 B, free: 3.0 GB)
17/03/26 10:10:32 INFO BlockManagerInfo: Added broadcast_1733_piece0 in memory on 196.168.168.104:59612 (size: 119.0 B, free: 3.0 GB)
Has been stuck in this place, I tried a lot of methods can't solve, may not know the reason, so you need to which a great god, give some advice, thank you

CodePudding user response:

By looking at the log found
17/03/25 22:52:32 INFO ExternalSorter: Thread 82 spilling in - the memory map of 473.6 MB to disk (25 times so far)
17/03/25 22:52:37 INFO ExternalSorter: Thread 71 spilling in - the memory map of 392.0 MB to disk (26 times so far)
17/03/25 22:52:52 INFO ExternalSorter: Thread 80 spilling in - the memory map of 392.0 MB to disk (22 times so far)
17/03/25 22:53:07 INFO ExternalSorter: Thread 70 spilling in - the memory map of 392.0 MB to disk (24 times so far)
17/03/25 22:53:38 INFO ExternalSorter: Thread 79 spilling in - the memory map of 401.9 MB to disk (27 times so far)
17/03/25 22:53:49 INFO ExternalSorter: Thread 83 spilling in - the memory map of 416.0 MB to disk (24 times so far)
17/03/25 22:53:53 INFO ExternalSorter: Thread 82 spilling in - the memory map of 396.8 MB to disk (26 times so far)
Don't know how to solve?

CodePudding user response:

I also encountered this problem, the card for a long time, in this step, the building Lord solved

CodePudding user response:

Not yet, haven't been solved, good confused, need to save a great spirit

CodePudding user response:

The landlord to provide the information is not complete,
First spark the version number, application code, in which the task, the memory configuration?
It seems like a memory disk writes frequently,

CodePudding user response:

The spark 2.1
Below is sumbit the content of the submitted
/root/spark - 2.1.0 - bin - hadoop2.6/bin/spark - submit \
- class com. Sirc. ZWZ. CSRJava. ChangeDataStruction. SCSR \
- num - executors 100 \
- driver - the memory 6 g \
- executor - 6 g memory \
- executor - 8 cores \
/root/jars/SparkCSR_JAVA - 0.0.1 - the SNAPSHOT. Jar \
7 cluster, 1 master and slave, 6 sets of 4 sets of each 8 gb of memory, may provide the spark to run the maximum amount of memory is 6 g (per), the other two are virtual machine each 2 gb of memory, each providing 1 g to calculate

It's part of the log information:
PrimaryNodeIndex=1, replicas=[ReplicaUnderConstruction [[DISK] dbc57 DS - 017-6553-43 ea - 8 a2d - 3555 fccd663d: NORMAL: 196.168.168.103:50010 | RBW], ReplicaUnderConstruction [[DISK] DS - 6 eb004b2 - b3dc - 42 df - b212 - ffa2fd6b5572: NORMAL: 196.168.168.27:50010 | RBW], ReplicaUnderConstruction [[DISK] DS - 5785 ace1 a611-479 - b - b360-79562081 feb1: NORMAL: 196.168.168.104:50010 | RBW]]}
2017-03-28 11:21:23,382 INFO org.apache.hadoop.hdfs.StateChange: BLOCK* allocateBlock:/SRResult/N25E118/_temporary/0/_temporary/attempt_20170327232029_0002_m_000017_21/part-00017. BP-2089499914-196.168.168.100-1490492430641 blk_1073742807_1983{blockUCState=UNDER_CONSTRUCTION, primaryNodeIndex=-1, replicas=[ReplicaUnderConstruction[[DISK]DS-5785ace1-a611-479b-b360-79562081feb1:NORMAL:196.168.168.104:50010|RBW], ReplicaUnderConstruction[[DISK]DS-411c0e4c-86c5-4203-94d8-d6d7a95df7da:NORMAL:196.168.168.102:50010|RBW], ReplicaUnderConstruction[[DISK]DS-6eb004b2-b3dc-42df-b212-ffa2fd6b5572:NORMAL:196.168.168.27:50010|RBW]]}
2017-03-28 11:21:23,459 INFO org.apache.hadoop.hdfs.StateChange: BLOCK* allocateBlock:/SRResult/N25E118/_temporary/0/_temporary/attempt_20170327232029_0002_m_000029_33/part-00029. BP-2089499914-196.168.168.100-1490492430641 blk_1073742808_1984{blockUCState=UNDER_CONSTRUCTION, primaryNodeIndex=-1, replicas=[ReplicaUnderConstruction[[DISK]DS-5785ace1-a611-479b-b360-79562081feb1:NORMAL:196.168.168.104:50010|RBW], ReplicaUnderConstruction[[DISK]DS-411c0e4c-86c5-4203-94d8-d6d7a95df7da:NORMAL:196.168.168.102:50010|RBW], ReplicaUnderConstruction[[DISK]DS-017dbc57-6553-43ea-8a2d-3555fccd663d:NORMAL:196.168.168.103:50010|RBW]]}
2017-03-28 11:21:23,509 INFO org.apache.hadoop.hdfs.StateChange: BLOCK* allocateBlock:/SRResult/N25E118/_temporary/0/_temporary/attempt_20170327232029_0002_m_000026_30/part-00026. BP-2089499914-196.168.168.100-1490492430641 blk_1073742809_1985{blockUCState=UNDER_CONSTRUCTION, primaryNodeIndex=-1, replicas=[ReplicaUnderConstruction[[DISK]DS-97e7de7f-fbcd-44bb-821d-4d245f1ce82c:NORMAL:196.168.168.101:50010|RBW], ReplicaUnderConstruction[[DISK]DS-6eb004b2-b3dc-42df-b212-ffa2fd6b5572:NORMAL:196.168.168.27:50010|RBW], ReplicaUnderConstruction[[DISK]DS-411c0e4c-86c5-4203-94d8-d6d7a95df7da:NORMAL:196.168.168.102:50010|RBW]]}
nullnullnullnullnullnullnullnullnullnullnullnullnullnullnullnullnullnullnullnullnullnullnullnullnullnullnullnullnullnullnullnullnullnullnullnullnullnullnullnullnullnullnullnullnullnullnullnullnullnullnullnullnullnullnullnullnullnullnullnullnullnullnullnullnullnullnullnullnullnullnullnullnullnullnullnullnullnullnullnullnullnullnullnullnullnullnullnullnullnullnullnullnullnullnullnullnullnullnullnullnullnullnullnullnullnullnullnullnullnullnullnullnullnullnullnullnullnullnullnullnull
  • Related