Amazon AWS Web Services

0 votes

For a Java class final project, we need to setup Hadoop and implement an n-gram processor. I have found a number of 'Hadoop on AWS' tutorials, but am uncertain how to deploy Hadoop while staying in the free tier. I tried a while ago, and received a bill for over $250 USD. Now I am too nervous to test, and would like help to ensure I do not rack up another bill.

From what I understand, these are the limits of the free tier: Allowable Instances: t1.micro or t2.micro Maximum instance-hours per month: 750 OS: Standard-priced versions of Linux or Windows on EC2 Storage: 30 GB of EBS Many of the Hadoop tutorials use instances other than those two shown above. The AWS tutorial here mentions it will cost about $1, stating it will run for one hour. I need to keep this/these instances active for over 2 weeks, but only really use them for a few minutes at a time. I do not believe that I will exceed even one-tenth of that 750 hours.

We can get bonus points if we use "more than 1 machine". Can I do that within the free tier? Does the free tier have a limit on the number of instances it can spin up?

Does anyone have a tutorial that stays in the free tier? Or should I skip AWS and try a local Hadoop solution?

Jul 8 in Big Data Hadoop by nitinrawat895
• 10,510 points
25 views

1 answer to this question.

0 votes

If you want to limit your hadoop cluster nodes only to t2.micro instances and total EBS volumes size to 30 GB, then you can run [in theory] a hadoop cluster within free tier. Do note that the hardware on t2.micro are of meagre.

The thing about free tier on AWS is that you are allowed only t2.micro for 750 hours per month. That means you can run for example 10 nodes for 75 hrs in a month for free after which you would be billed.

Here is a post that does exactly what you want with 4 nodes "Spinning Up a Free Hadoop Cluster: Step by Step". So you should be able to run this cluster within free limit for around 1 week

answered Jul 8 by ravikiran
• 4,200 points

Related Questions In Big Data Hadoop

0 votes
1 answer

How can I use my host machine’s web browser to check my HDFS services running in the VM?

The sole purpose of the virtual machine ...READ MORE

answered Apr 18, 2018 in Big Data Hadoop by Shubham
• 13,290 points
120 views
0 votes
1 answer

Killing an Application from ResourceManager Web UI

In hadoop v2.7.1, if you open the ...READ MORE

answered Apr 23, 2018 in Big Data Hadoop by Shubham
• 13,290 points
102 views
0 votes
1 answer

Is there any way to access Hadoop web UI in linux?

In this case what you can do ...READ MORE

answered May 9, 2018 in Big Data Hadoop by nitinrawat895
• 10,510 points
620 views
0 votes
1 answer

Cant find slave on ResourceManager Web UI

Seems like it is running on default ...READ MORE

answered Dec 4, 2018 in Big Data Hadoop by Omkar
• 67,290 points
10 views
0 votes
1 answer

Deploy Docker Containers from Docker Cloud

To solve this problem, I followed advice ...READ MORE

answered Sep 3, 2018 in AWS by Priyaj
• 56,520 points
136 views
+13 votes
2 answers

Git management technique when there are multiple customers and need multiple customization?

Consider this - In 'extended' Git-Flow, (Git-Multi-Flow, ...READ MORE

answered Mar 26, 2018 in DevOps & Agile by DragonLord999
• 8,380 points
124 views
0 votes
1 answer

Can I use Amazon Web Services for free?

If you would like to use AWS ...READ MORE

answered Apr 25 in Big Data Hadoop by ravikiran
• 4,200 points
50 views