PySpark not starting No active sparkcontext

0 votes

Hi,
I am not able to start a pyspark session in edureka VM session.

I am using the command "pyspark"
Error Message is :

"Connection Refused: localhost/127.0.0.1:7077"
No active sparkcontext

I didn't get the spark logo at the end.
Can you help me in resolving it?

Jul 30, 2019 in Apache Spark by Karan
4,591 views

1 answer to this question.

0 votes

Seems like Spark hadoop daemons are not running. Start it first and then start pyspark. Refer to the below commands:

$ cd /usr/lib/spark-2.1.1-bin-hadoop2.7

$ cd sbin

$ ./start-all.sh


Hope this helps!

Join Pyspark training online today to know more about Pyspark.

Thanks.

answered Jul 30, 2019 by Jishan

Related Questions In Apache Spark

0 votes
1 answer

SparkContext.addFile() not able to update file.

Spark by default won't let you overwrite ...READ MORE

answered Mar 10, 2019 in Apache Spark by Siri
3,540 views
+1 vote
1 answer

Error: value textfile is not a member of org.apache.spark.SparkContext

Hi, Regarding this error, you just need to change ...READ MORE

answered Jul 4, 2019 in Apache Spark by Gitika
• 65,890 points
4,273 views
0 votes
1 answer

Spark Streaming Pyspark code not working

The address you are using in the ...READ MORE

answered Jul 11, 2019 in Apache Spark by Shir
2,292 views
0 votes
1 answer

env: ‘python’: No such file or directory in pyspark.

Hi@akhtar, This error occurs because your python version ...READ MORE

answered Apr 7, 2020 in Apache Spark by MD
• 95,460 points
6,356 views
+1 vote
2 answers
+1 vote
1 answer

Hadoop Mapreduce word count Program

Firstly you need to understand the concept ...READ MORE

answered Mar 16, 2018 in Data Analytics by nitinrawat895
• 11,380 points
10,931 views
0 votes
1 answer

hadoop.mapred vs hadoop.mapreduce?

org.apache.hadoop.mapred is the Old API  org.apache.hadoop.mapreduce is the ...READ MORE

answered Mar 16, 2018 in Data Analytics by nitinrawat895
• 11,380 points
2,457 views
+2 votes
11 answers

hadoop fs -put command?

Hi, You can create one directory in HDFS ...READ MORE

answered Mar 16, 2018 in Big Data Hadoop by nitinrawat895
• 11,380 points
108,354 views
0 votes
5 answers

How to change the spark Session configuration in Pyspark?

You aren't actually overwriting anything with this ...READ MORE

answered Dec 14, 2020 in Apache Spark by Gitika
• 65,890 points
125,208 views
–1 vote
1 answer

Not able to use sc in spark shell

Seems like master and worker are not ...READ MORE

answered Jan 3, 2019 in Apache Spark by Omkar
• 69,230 points
1,679 views
webinar REGISTER FOR FREE WEBINAR X
REGISTER NOW
webinar_success Thank you for registering Join Edureka Meetup community for 100+ Free Webinars each month JOIN MEETUP GROUP