Github user mgummelt commented on the issue:

    https://github.com/apache/spark/pull/11157
  
    > How about the rest of the ports?
    Unused resources in an offer are implicitly declined.
    > We will have only one executor per job right?
    if they specify a port, then one executor per node per job, yes
    > Right now I get conflicts without the isolator in my local tests.
    what conflicts?
    > So why consume all port resources?
    See above.  Unused resources in an offer are implicitly declined.  So if we 
launch a task using a single port, we're implicitly declining all other ports.
    
    > Moreover if we have the isolator enabled I think we should allow task 
creation independently of a port offer
    
    That's a good point.  I spoke with our networking team, and unfortunately 
there's no way for us to discover if IP-per-container is enabled, so we'd have 
to expose a spark.mesos.ip-per-container or something, to prevent using port 
offers.  I think this is an orthogonal problem though, since we don't even 
support CNI yet.  Let's solve it in a separate PR.
    



---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastruct...@apache.org or file a JIRA ticket
with INFRA.
---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org

Reply via email to