Hi Dan, Here is an example I wrote few months back. Let me know if it works for you. http://enstartup.com/post/12298119544/execute-oozie-tests-like-what-you-test-hadoop-using
Thanks, Angelo On Tue, May 1, 2012 at 9:27 AM, Alejandro Abdelnur <[email protected]>wrote: > Dan, > > Alternatively you could do as the OozieExamples testcases that fires > up a mini cluster. > > Thxs. > > Alejandro > > On Tue, May 1, 2012 at 6:06 AM, Harsh J <[email protected]> wrote: > > Hey Dan, > > > > Until https://issues.apache.org/jira/browse/OOZIE-477 goes in, I don't > > believe you can run off of local FS. > > > > I usually solve this by bootstrapping some FS-API calls/commands > > before re-running the Oozie workflow. > > > > Something like the following one liner in bash history: > > > > $ hadoop fs -rmr examples/apps/java-app; hadoop fs -put java-app > > examples/apps; oozie job -run java-app/job.properties > > > > On Tue, May 1, 2012 at 10:06 AM, Dan Feldman <[email protected]> > wrote: > >> I was wondering if there is something analogous to pig's local mode for > >> oozie? For instance, I couldn't get oozie working while my workflow.xml > and > >> job.properties files were sitting on the local system - I was getting > >> "wrong FS" error just like the OP here > >> > https://groups.google.com/a/cloudera.org/group/scm-users/browse_thread/thread/c96a6a6149e1e86e/c399307616b920f0?#c399307616b920f0 > . > >> But now that the files are on the hdfs, it seems like I have to delete > and > >> re-upload them to hdfs every time I need to make modifications to any of > >> the files. > >> > >> I'm hoping to use oozie to automate analysis of cassandra data using a > >> bunch of pig and ruby scripts and as of now, local mode is more than > enough > >> to handle that... > >> > >> Thanks! > >> Dan > > > > > > > > -- > > Harsh J > > > > -- > Alejandro >
