[jira] [Assigned] (SPARK-36610) Add `thousands` argument to `ps.read_csv`.

2021-09-03 Thread Apache Spark (Jira)


 [ 
https://issues.apache.org/jira/browse/SPARK-36610?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Apache Spark reassigned SPARK-36610:


Assignee: Apache Spark

> Add `thousands` argument to `ps.read_csv`.
> --
>
> Key: SPARK-36610
> URL: https://issues.apache.org/jira/browse/SPARK-36610
> Project: Spark
>  Issue Type: Sub-task
>  Components: PySpark
>Affects Versions: 3.2.0
>Reporter: Haejoon Lee
>Assignee: Apache Spark
>Priority: Major
>
> When reading csv file in pandas, pandas automatically detect the thousand 
> separator if `thousands` argument is specified.
> {code:java}
> >>> pd.read_csv(path, sep=";")
> name  agejob  money
> 0  Jorge   30  Developer  1,000,000
> 1Bob   32  Developer100
> >>> pd.read_csv(path, sep=";", thousands=",")
> name  agejobmoney
> 0  Jorge   30  Developer  100
> 1Bob   32  Developer  100{code}
> However, pandas-on-Spark doesn't support it.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Assigned] (SPARK-36610) Add `thousands` argument to `ps.read_csv`.

2021-09-03 Thread Apache Spark (Jira)


 [ 
https://issues.apache.org/jira/browse/SPARK-36610?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Apache Spark reassigned SPARK-36610:


Assignee: (was: Apache Spark)

> Add `thousands` argument to `ps.read_csv`.
> --
>
> Key: SPARK-36610
> URL: https://issues.apache.org/jira/browse/SPARK-36610
> Project: Spark
>  Issue Type: Sub-task
>  Components: PySpark
>Affects Versions: 3.2.0
>Reporter: Haejoon Lee
>Priority: Major
>
> When reading csv file in pandas, pandas automatically detect the thousand 
> separator if `thousands` argument is specified.
> {code:java}
> >>> pd.read_csv(path, sep=";")
> name  agejob  money
> 0  Jorge   30  Developer  1,000,000
> 1Bob   32  Developer100
> >>> pd.read_csv(path, sep=";", thousands=",")
> name  agejobmoney
> 0  Jorge   30  Developer  100
> 1Bob   32  Developer  100{code}
> However, pandas-on-Spark doesn't support it.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org