Brief Introduction to Oozie

Recommended by 46 users

Sep 11, 2014
Brief Introduction to Oozie
Add to Bookmark Email this Post 10K    5

Oozie is a workflow scheduler system to manage Apache Hadoop jobs. Oozie is integrated with the rest of the Hadoop stack supporting several types of Hadoop jobs such as Java MapReduce, Streaming MapReduce, Pig, Hive and Sqoop. Oozie is a scalable, reliable and extensible system. Oozie is used in production at Yahoo!, running more than 200,000 jobs every day.

Features of Oozie:

  • Execute and monitor workflows in Hadoop
  • Periodic scheduling of workflows
  • Trigger execution of data availability
  • HTTP and command line interface and web console

Oozie Workflow – Directed Acyclic Graph of Jobs:

Oozie Workflow Example:

Oozie Workflow Example

<workflow-app nome='wordcount –wf’>
 <start to= ‘wordcount’/>
<action name=’Wordcount'>
 <value> ${outputDir}</value>
<ok to='end’/>
 <error to='kill'/>
<kill name='kill'/>
<end name='end'/>

Workflow Definition:

A workflow definition is a DAG with control flow nodes or action nodes, where the nodes are connected by transitions arrows.

Control Flow Nodes:     

The control flow provides a way to control the Workflow execution path. Flow control operations within the workflow applications can be done through the following nodes:

  • Start/end/kill
  • Decision
  • Fork/join

Action Nodes:

  • Map-reduce
  • Pig
  • HDFS
  • Sub-workflow
  • Java – Run custom Java code

Oozie Workflow Application:

Workflow application is a ZIP file that includes the workflow definition and the necessary files to run all the actions. It contains the following files:

  • Configuration file – config-default.xml
  • App files – lib/ directory with JAR and SO files
  • Pig scripts

Application Deployment:

$ hadoop fs-put wordcount-wf hdfs://

Workflow Job Parameters:

$ cat job.properites

Job Execution:

$ oozie job –run –config

Got a question for us? Mention them in the comments section and we will get back to you. 

Related Posts:

Big Data and Hadoop Training

Why Learn Hadoop?

Hadoop 2.0 FAQs

Introduction to Hadoop 2.0

Share on
  • Rajiv

    sir how to schedule job using crontab

    • EdurekaSupport

      Hey Rajiv, thanks for checking out our blog. Please refer to the steps given below to step up cron job:
      1. Prepare SQL to be run on using CRON
      2. See below for example of code which needs to be added to SQL code for a cron job
      .logon server/user_id, Teradata password
      For example :
      .logon Mozart/akatarni,Welcome1
      3. WinSCP – this is the file transfer application that is used to transfer the .SQL code file to the server.
      a. Open “WinSCP”, Server name =
      b. Give login id and SAS password
      c. Copy the code from your system to server window, in the attached snap shot we have copied “ask_lstg.sql” from genpact(personal system) to server window.
      i. Left window shows your personal computer and right one is server
      4. Open “Putty”. Use server
      5. At the prompt, enter SAS credentials. After entering the password , you will see the attached window.
      6. To open the editor :
      a. Type export EDITOR=vi <hit enter>
      b. Type crontab -e <hit enter>
      i. This command edits your crontab file, or create one if it doesn’t already exist.
      c. Press “i” to start typing
      d. Press <ESC> to get out of insert mode
      7. Then make the cron job entry:
      A crontab entry has five fields for specifying day, date and time followed by the command to be run at that interval.
      00 06 * * * /usr/bin/bteq <fake_lstg.sql> fake_lstg.LOG 2>&1
      The above will run the code at 06:00 hours every day
      In the above example, “fake_lstg.sql” is SQL file, “fake_lstg.LOG” is the log file where results will appear
      15 20 * * 0 /usr/bin/bteq <fake_lstg.sql> fake_lstg.LOG 2>&1
      The above will run the code at 20:15 hours every Sunday
      8. Keep adding lines to the crontab file to schedule more job.
      a. The easiest way to add a line is to be at the first character in the file, then in ESC mode,
      click <shift> + O (case sensitive). This adds a new line above the current one.
      9. To move around the file, in ESC mode
      “l” – move right
      “h” – move left
      “j” – move down
      “k” – move up
      10. To save the crontab file and exit, press <ESC>, then :wq
      a. To exit the file WITHOUT saving, press <ESC>, the :q!
      11. Type Exit at the Unix prompt to exit Putty.
      12. The cron job should run at the specified time
      13. Check the *.LOG file to make sure code ran successfully.
      Hope this helps. Cheers!

      • Rajiv

        sir thanks for giving answer to my question..its helpful form me…good and fine description..thanks to u sir

  • Sankalp Tomar


    Suppose we want to use the output of Hive Job as an input to Mapreduce Job. How can we achieve this??

    • EdurekaSupport

      Hey Sankalp, thanks for checking out our blog. With regard to your query, first we can store the output of hive in hdfs and then we can execute it as an input file for mapreduce code.
      Storing the output of hive.
      INSERT OVERWRITE DIRECTORY ‘/path/to/output/dir’
      select books from table;
      Hope this helps. Cheers!