hi all, I am a very newbie of apache spark, recently I have tried spark on yarn, it works for batch process. Now we want to try streaming process using spark-streaming, and still, use yarn for resource scheduler as we want to manager all the resource of the cluster used for computing tasks in a unified way. Can this works? any suggestions are welcome!
-------------------------------------------------------------------------------- Best Regards!
