Spark multiple version without major version

–1 vote

Is it possible to use multiple spark version without setting the SPARK_MAJOR_VERSION?

Dec 27, 2018 in Big Data Hadoop by digger
• 26,740 points
827 views

1 answer to this question.

0 votes

Yes. It is not necessary to set SPARK_MAJOR_VERSION. You can type different command to run different spark.

spark-shell

loads Spark 1.6

and

spark2-shell

loads Spark 2.0

answered Dec 27, 2018 by Omkar
• 69,220 points

Related Questions In Big Data Hadoop

0 votes
3 answers

Can we run Spark without using Hadoop?

No, you can run spark without hadoop. ...READ MORE

answered May 7, 2019 in Big Data Hadoop by pradeep
2,363 views
0 votes
1 answer

Joining Multiple Spark Dataframes

You can run the below code to ...READ MORE

answered Mar 26, 2018 in Big Data Hadoop by Bharani
• 4,660 points
2,865 views
0 votes
1 answer

How to read Spark elements having multiple lines each?

Try this: val new_records = sc.newAPIHadoopRDD(hadoopConf,classOf[ ...READ MORE

answered Dec 12, 2018 in Big Data Hadoop by Omkar
• 69,220 points
1,421 views
0 votes
1 answer

Can you build “Spark” with any particular Hadoop version?

Yes, one can build “Spark” for a specific ...READ MORE

answered Dec 14, 2018 in Big Data Hadoop by Frankie
• 9,830 points
1,411 views
+1 vote
2 answers
+1 vote
1 answer

Hadoop Mapreduce word count Program

Firstly you need to understand the concept ...READ MORE

answered Mar 16, 2018 in Data Analytics by nitinrawat895
• 11,380 points
11,077 views
0 votes
1 answer

hadoop.mapred vs hadoop.mapreduce?

org.apache.hadoop.mapred is the Old API  org.apache.hadoop.mapreduce is the ...READ MORE

answered Mar 16, 2018 in Data Analytics by nitinrawat895
• 11,380 points
2,575 views
+2 votes
11 answers

hadoop fs -put command?

Hi, You can create one directory in HDFS ...READ MORE

answered Mar 16, 2018 in Big Data Hadoop by nitinrawat895
• 11,380 points
109,073 views
–1 vote
1 answer

How to use multiple spark version?

You can use the  SPARK_MAJOR_VERSION for this. Suppose ...READ MORE

answered Dec 27, 2018 in Big Data Hadoop by Omkar
• 69,220 points
2,266 views
0 votes
1 answer

Hadoop Spark: What is version to find SparkSession in library Spark?

you need both core and SQL artifacts <repositories> ...READ MORE

answered Nov 13, 2018 in Big Data Hadoop by Omkar
• 69,220 points
2,138 views
webinar REGISTER FOR FREE WEBINAR X
REGISTER NOW
webinar_success Thank you for registering Join Edureka Meetup community for 100+ Free Webinars each month JOIN MEETUP GROUP