Setting directory for Spark recovery

0 votes
Hello guys! I need help with my Spark application, especially with the recovery part. I have set the recovery mode to Zookeeper and have set the URL. Everything is fine and the connection is also happening as expected. I am stuck with the last part. I want to set the directory where the recovery state is stored. Please mention the commands to set this directory. Thanks
Mar 25, 2019 in Apache Spark by Rahul
766 views

1 answer to this question.

0 votes

You can set the directory to store the recovery state dynamically. Here is the command you can use to do it:

val sc = new SparkContext(new SparkConf())

./bin/spark-submit <all your existing options> --spark.deploy.zookeeper.dir=<path to directory>
answered Mar 25, 2019 by Hari

Related Questions In Apache Spark

0 votes
1 answer

How to connect to Zookeeper after setting Spark recovery mode?

You have set Zookeeper as the recovery ...READ MORE

answered Mar 25, 2019 in Apache Spark by Hari
1,100 views
0 votes
1 answer
0 votes
1 answer

Hadoop mandatory for Spark?

No not mandatory, but there is no ...READ MORE

answered Jun 1, 2018 in Apache Spark by kurt_cobain
• 9,390 points
415 views
0 votes
2 answers

Which cluster type should I choose for Spark?

Spark is agnostic to the underlying cluster ...READ MORE

answered Aug 21, 2018 in Apache Spark by zombie
• 3,790 points
1,649 views
+1 vote
2 answers
+1 vote
1 answer

Hadoop Mapreduce word count Program

Firstly you need to understand the concept ...READ MORE

answered Mar 16, 2018 in Data Analytics by nitinrawat895
• 11,380 points
10,559 views
0 votes
1 answer

hadoop.mapred vs hadoop.mapreduce?

org.apache.hadoop.mapred is the Old API  org.apache.hadoop.mapreduce is the ...READ MORE

answered Mar 16, 2018 in Data Analytics by nitinrawat895
• 11,380 points
2,185 views
+2 votes
11 answers

hadoop fs -put command?

Hi, You can create one directory in HDFS ...READ MORE

answered Mar 16, 2018 in Big Data Hadoop by nitinrawat895
• 11,380 points
104,216 views
0 votes
1 answer

Setting port for SSL for Spark application

You can set the port in the ...READ MORE

answered Mar 15, 2019 in Apache Spark by Karan
519 views
+1 vote
3 answers

Which cluster type should I choose for Spark?

According to me, start with a standalone ...READ MORE

answered Jun 27, 2018 in Apache Spark by nitinrawat895
• 11,380 points
1,212 views
webinar REGISTER FOR FREE WEBINAR X
REGISTER NOW
webinar_success Thank you for registering Join Edureka Meetup community for 100+ Free Webinars each month JOIN MEETUP GROUP