Skip to content

Latest commit

 

History

History
37 lines (28 loc) · 1.54 KB

spark-application-configuration-through-spark-submit.md

File metadata and controls

37 lines (28 loc) · 1.54 KB
title description keywords services documentationcenter author manager editor ms.assetid ms.service ms.workload ms.tgt_pltfrm ms.devlang ms.topic ms.date ms.author
How do I configure Spark application through spark-submit on HDInsight clusters? | Microsoft Docs
Use the Spark FAQ for answers to common questions on Spark on Azure HDInsight platform.
Azure HDInsight, Spark, FAQ, troubleshooting guide, common problems, application configuration, spark-submit
Azure HDInsight
na
arijitt
0DBF3DBF-FD6C-4F87-A8AE-7F1D74B094A2
multiple
na
na
na
article
03/30/2017
arijitt

How do I configure Spark application through spark-submit on HDInsight clusters?

Issue:

Need to configure at submit time through spark-submit, the amount of memory and number of cores that a Spark application can use on HDInsight clusters.

  1. Refer to the topic Why did my Spark application fail with OutOfMemoryError? to determine which Spark configurations need to be set and to what values.

  2. Launch spark-shell with a command similar to the following (change the actual value of the configurations as applicable):

spark-submit --master yarn-cluster --class com.microsoft.spark.application --num-executors 4 --executor-memory 4g --executor-cores 2 --driver-memory 8g --driver-cores 4 /home/user/spark/sparkapplication.jar

Further Reading:

Spark job submission on HDInsight clusters