[jira] [Commented] (SPARK-44845) spark job copies jars repeatedly if fs.defaultFS and application jar are same url

2023-08-17 Thread zheju_he (Jira)


[ 
https://issues.apache.org/jira/browse/SPARK-44845?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17755404#comment-17755404
 ] 

zheju_he commented on SPARK-44845:
--

my pr https://github.com/apache/spark/pull/42529

> spark job copies jars repeatedly if fs.defaultFS and application jar are same 
> url
> -
>
> Key: SPARK-44845
> URL: https://issues.apache.org/jira/browse/SPARK-44845
> Project: Spark
>  Issue Type: Bug
>  Components: YARN
>Affects Versions: 3.4.1
>Reporter: zheju_he
>Priority: Major
>
> In the org.apache.spark.deploy.yarn.Client#compareUri method, 
> hdfs://hadoop81:8020 and hdfs://192.168.0.81:8020 are regarded as different 
> file systems (hadoop81 corresponds to 192.168.0.81). The specific reason is 
> that in the last pr, different URIs of user information are also regarded as 
> different file systems. Uri.getauthority is used to determine the user 
> information, but authority contains the host so the URI above must be 
> different from authority. To determine whether the user authentication 
> information is different, you only need to determine URI.getUserInfo.
>  
> the last pr and issue link:
> https://issues.apache.org/jira/browse/SPARK-22587
> https://github.com/apache/spark/pull/19885



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Created] (SPARK-44845) spark job copies jars repeatedly if fs.defaultFS and application jar are same url

2023-08-17 Thread zheju_he (Jira)
zheju_he created SPARK-44845:


 Summary: spark job copies jars repeatedly if fs.defaultFS and 
application jar are same url
 Key: SPARK-44845
 URL: https://issues.apache.org/jira/browse/SPARK-44845
 Project: Spark
  Issue Type: Bug
  Components: YARN
Affects Versions: 3.4.1
Reporter: zheju_he


In the org.apache.spark.deploy.yarn.Client#compareUri method, 
hdfs://hadoop81:8020 and hdfs://192.168.0.81:8020 are regarded as different 
file systems (hadoop81 corresponds to 192.168.0.81). The specific reason is 
that in the last pr, different URIs of user information are also regarded as 
different file systems. Uri.getauthority is used to determine the user 
information, but authority contains the host so the URI above must be different 
from authority. To determine whether the user authentication information is 
different, you only need to determine URI.getUserInfo.

 

the last pr and issue link:
https://issues.apache.org/jira/browse/SPARK-22587

https://github.com/apache/spark/pull/19885



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Commented] (SPARK-42088) Running python3 setup.py sdist on windows reports a permission error

2023-01-16 Thread zheju_he (Jira)


[ 
https://issues.apache.org/jira/browse/SPARK-42088?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17677214#comment-17677214
 ] 

zheju_he commented on SPARK-42088:
--

This is my pr address https://github.com/apache/spark/pull/39603

> Running python3 setup.py sdist on windows reports a permission error
> 
>
> Key: SPARK-42088
> URL: https://issues.apache.org/jira/browse/SPARK-42088
> Project: Spark
>  Issue Type: Bug
>  Components: Build
>Affects Versions: 3.4.0
>Reporter: zheju_he
>Priority: Minor
>
> My system version is windows 10, and I can run setup.py with administrator 
> permissions, so there will be no error. However, it may be troublesome for us 
> to upgrade permissions with Windows Server, so we need to modify the code of 
> setup.py to ensure no error. To avoid the hassle of compiling for the user, I 
> suggest modifying the following code to enable the out-of-the-box effect
> {code:python}
> def _supports_symlinks():
> """Check if the system supports symlinks (e.g. *nix) or not."""
> return getattr(os, "symlink", None) is not None and 
> ctypes.windll.shell32.IsUserAnAdmin() != 0 if sys.platform == "win32" else 
> True
> {code}



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Created] (SPARK-42088) Running python3 setup.py sdist on windows reports a permission error

2023-01-16 Thread zheju_he (Jira)
zheju_he created SPARK-42088:


 Summary: Running python3 setup.py sdist on windows reports a 
permission error
 Key: SPARK-42088
 URL: https://issues.apache.org/jira/browse/SPARK-42088
 Project: Spark
  Issue Type: Bug
  Components: Build
Affects Versions: 3.4.0
Reporter: zheju_he


My system version is windows 10, and I can run setup.py with administrator 
permissions, so there will be no error. However, it may be troublesome for us 
to upgrade permissions with Windows Server, so we need to modify the code of 
setup.py to ensure no error. To avoid the hassle of compiling for the user, I 
suggest modifying the following code to enable the out-of-the-box effect
{code:python}
def _supports_symlinks():
"""Check if the system supports symlinks (e.g. *nix) or not."""
return getattr(os, "symlink", None) is not None and 
ctypes.windll.shell32.IsUserAnAdmin() != 0 if sys.platform == "win32" else True
{code}



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org