title | description | keywords | services | documentationcenter | author | manager | editor | ms.assetid | ms.service | ms.workload | ms.tgt_pltfrm | ms.devlang | ms.topic | ms.date | ms.author |
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
How do I configure Spark application through LIVY on HDInsight clusters? | Microsoft Docs |
Use the Spark FAQ for answers to common questions on Spark on Azure HDInsight platform. |
Azure HDInsight, Spark, FAQ, troubleshooting guide, common problems, application configuration, livy submission, remote submission |
Azure HDInsight |
na |
arijitt |
D7D9AE44-2C92-484F-8032-3F4CA8A223A3 |
multiple |
na |
na |
na |
article |
03/30/2017 |
arijitt |
Need to configure at submit time through LIVY, the amount of memory and number of cores that a Spark application can use on HDInsight clusters.
-
Refer to the topic Why did my Spark application fail with OutOfMemoryError? to determine which Spark configurations need to be set and to what values.
-
Submit the Spark application to LIVY using a REST client like CURL with a command similar to the following (change the actual values as applicable):
curl -k --user 'username:password' -v -H 'Content-Type: application/json' -X POST -d '{ "file":"wasb://container@storageaccountname.blob.core.windows.net/example/jars/sparkapplication.jar", "className":"com.microsoft.spark.application", "numExecutors":4, "executorMemory":"4g", "executorCores":2, "driverMemory":"8g", "driverCores":4}'