~~ Licensed under the Apache License, Version 2.0 (the "License"); ~~ you may not use this file except in compliance with the License. ~~ You may obtain a copy of the License at ~~ ~~ http://www.apache.org/licenses/LICENSE-2.0 ~~ ~~ Unless required by applicable law or agreed to in writing, software ~~ distributed under the License is distributed on an "AS IS" BASIS, ~~ WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. ~~ See the License for the specific language governing permissions and ~~ limitations under the License. See accompanying LICENSE file. --- Oozie Quick Start --- Oozie Quick Start These instructions install and run Oozie using an embedded Tomcat server and an embedded Derby database. For detailed install and configuration instructions refer to {{{AG_Install}Oozie Install}}. * System Requirements * Unix (tested in Linux and Mac OS X) * Java 1.6+ * Hadoop * {{{http://hadoop.apache.org}Apache Hadoop}} (tested with 0.20.2) * {{{http://developer.yahoo.com/hadoop}Yahoo! Hadoop}} (tested with 0.20.104.2) * ExtJS library (optional, to enable Oozie webconsole) * {{{http://extjs.com/deploy/ext-2.2.zip}ExtJS 2.2}} The Java 1.6+ directory should be in the command path. * Server Installation <> Oozie ignores any set value for , Oozie computes its home automatically. * Download or build an Oozie binary distribution * Download a Hadoop binary distribution * Download ExtJS library (it must be version 2.2) [] <> The ExtJS library is not bundled with Oozie because it uses a different license. <> It is recommended to use a Oozie Unix user for the Oozie server. Expand the Oozie distribution . Expand the Hadoop distribution (as the Oozie Unix user). Oozie is bundled without Hadoop JAR files and without the ExtJS library. * The Hadoop JARs are required to run Oozie * The ExtJS library is optional (only required for the Oozie web-console to work) [] <> all Oozie server scripts (, , and ) run only under the Unix user that owns the Oozie installation directory, if necessary use when invoking the scripts. Use the script to add the Hadoop JARs and the ExtJS library to Oozie. ----------------------------------------------------------------------------- $ bin/oozie-setup.sh -hadoop 0.20.200 ${HADOOP_HOME} -extjs /tmp/ext-2.2.zip ----------------------------------------------------------------------------- To start Oozie as a daemon process run: ----------------------------------------------------------------------------- $ bin/oozie-start.sh ----------------------------------------------------------------------------- To start Oozie as a foreground process run: ----------------------------------------------------------------------------- $ bin/oozie-run.sh ----------------------------------------------------------------------------- Check the Oozie log file to ensure Oozie started properly. Using the Oozie command line tool check the status of Oozie: ----------------------------------------------------------------------------- $ bin/oozie admin -oozie http://localhost:11000/oozie -status ----------------------------------------------------------------------------- Using a browser go to the {{{http://localhost:11000/oozie}Oozie web console}}, Oozie status should be <>. Refer to the {{{DG_Examples}Running the Examples}} document for details on running the examples. * Client Installation Copy and expand the TAR.GZ file bundled with the distribution. Add the directory to the . Refer to the {{{DG_CommandLineTool}Command Line Interface Utilities}} document for a a full reference of the command line tool. NOTE: The Oozie server installation includes the Oozie client. The Oozie client should be installed in remote machines only. * Oozie Share Lib Installation Expand the TAR.GZ file bundled with the distribution. The directory must be copied to the Oozie HOME directory in HDFS: ----------------------------------------------------------------------------- $ hadoop fs -put share share ----------------------------------------------------------------------------- <> This must be done using the Oozie Hadoop (HDFS) user. If a directory already exists in HDFS, it must be deleted before copying it again.