Set Library to launch Yarn master

0 votes
I have a custom library that I want to use to launch Spark application master. I am running Spark application in Client mode. I can see that no library is set to be used. How can I set the custom library for the application?
Mar 28 in Apache Spark by Krishna
22 views

1 answer to this question.

0 votes

You can make use of Special Library path to make your application use your custom library for Yarn application master. To do this, run the below command in Spark shell after running your Spark application.

val sc = new SparkContext(new SparkConf())

./bin/spark-submit <all your existing options> --spark.yarn.am.extraLibraryPath= <path to custom library>
answered Mar 28 by Raj

Related Questions In Apache Spark

0 votes
1 answer

How to set executors for static allocation in Spark Yarn?

Open Spark shell and run the following ...READ MORE

answered Mar 28 in Apache Spark by Raj
206 views
0 votes
1 answer

Thread to use Yarn application master is limited

This is because the maximum number of ...READ MORE

answered Mar 28 in Apache Spark by Raj
60 views
0 votes
1 answer
0 votes
1 answer

When running Spark on Yarn, do I need to install Spark on all nodes of Yarn Cluster?

No, it is not necessary to install ...READ MORE

answered Jun 14, 2018 in Apache Spark by nitinrawat895
• 10,730 points
1,241 views
0 votes
1 answer
0 votes
1 answer

Hadoop Mapreduce word count Program

Firstly you need to understand the concept ...READ MORE

answered Mar 16, 2018 in Data Analytics by nitinrawat895
• 10,730 points
3,371 views
0 votes
1 answer

hadoop.mapred vs hadoop.mapreduce?

org.apache.hadoop.mapred is the Old API  org.apache.hadoop.mapreduce is the ...READ MORE

answered Mar 16, 2018 in Data Analytics by nitinrawat895
• 10,730 points
406 views
0 votes
10 answers

hadoop fs -put command?

put syntax: put <localSrc> <dest> copy syntax: copyFr ...READ MORE

answered Dec 7, 2018 in Big Data Hadoop by Aditya
16,760 views
0 votes
1 answer

Increasing memory to use for Yarn application master?

You can increase the memory dynamically by ...READ MORE

answered Mar 26 in Apache Spark by Tina
34 views
0 votes
1 answer

How to set max executors for dynamic allocation?

You can set it by assigning the ...READ MORE

answered Mar 13 in Apache Spark by Venu
137 views