If you look at the threads, the other 30 are almost surely not Spark worker threads. They're the JVM finalizer, GC threads, Jetty listeners, etc. Nothing wrong with this. Your OS has hundreds of threads running now, most of which are idle, and up to 4 of which can be executing. In a one-machine cluster, I don't think you would expect any difference in number of running threads. More data does not mean more threads, no. Your executor probably takes as many threads as cores in both cases, 4.
On Sat, Feb 7, 2015 at 10:14 AM, Deep Pradhan <pradhandeep1...@gmail.com> wrote: > Hi, > I am using YourKit tool to profile Spark jobs that is run in my Single Node > Spark Cluster. > When I see the YourKit UI Performance Charts, the thread count always > remains at > All threads: 34 > Daemon threads: 32 > > Here are my questions: > > 1. My system can run only 4 threads simultaneously, and obviously my system > does not have 34 threads. What could 34 threads mean? > > 2. I tried running the same job with four different datasets, two small and > two relatively big. But in the UI the thread count increases by two, > irrespective of data size. Does this mean that the number of threads > allocated to each job depending on data size is not taken care by the > framework? > > Thank You --------------------------------------------------------------------- To unsubscribe, e-mail: user-unsubscr...@spark.apache.org For additional commands, e-mail: user-h...@spark.apache.org