Hi Wang, maybe you can consider to use an integration framework like Apache Camel in order to run differents jobs...
Alonso Isidoro Roman [image: https://]about.me/alonso.isidoro.roman <https://about.me/alonso.isidoro.roman?promo=email_sig&utm_source=email_sig&utm_medium=email_sig&utm_campaign=external_links> 2016-06-16 13:08 GMT+02:00 Jacek Laskowski <ja...@japila.pl>: > Hi, > > When you say "several ETL types of things", what is this exactly? What > would an example of "dependency between these jobs" be? > > Pozdrawiam, > Jacek Laskowski > ---- > https://medium.com/@jaceklaskowski/ > Mastering Apache Spark http://bit.ly/mastering-apache-spark > Follow me at https://twitter.com/jaceklaskowski > > > On Thu, Jun 16, 2016 at 11:36 AM, Haopu Wang <hw...@qilinsoft.com> wrote: > > Hi, > > > > > > > > Suppose I have a spark application which is doing several ETL types of > > things. > > > > I understand Spark can analyze and generate several jobs to execute. > > > > The question is: is it possible to control the dependency between these > > jobs? > > > > > > > > Thanks! > > > > > > --------------------------------------------------------------------- > To unsubscribe, e-mail: user-unsubscr...@spark.apache.org > For additional commands, e-mail: user-h...@spark.apache.org > >