[
https://issues.apache.org/jira/browse/SPARK-26645?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17238798#comment-17238798
]
Punit Shah commented on SPARK-26645:
Hello [~dongjoon] If we can get this PR then this would be
[
https://issues.apache.org/jira/browse/SPARK-33445?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17234919#comment-17234919
]
Punit Shah commented on SPARK-33445:
Thank you very much [~dongjoon]
> Can't parse decimal type
[
https://issues.apache.org/jira/browse/SPARK-33445?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17234851#comment-17234851
]
Punit Shah commented on SPARK-33445:
My apologies [~dongjoon] for the incorrect tags. Please let me
[
https://issues.apache.org/jira/browse/SPARK-33445?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17233589#comment-17233589
]
Punit Shah edited comment on SPARK-33445 at 11/17/20, 1:38 PM:
---
[
https://issues.apache.org/jira/browse/SPARK-33445?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
Punit Shah reopened SPARK-33445:
As per the issue description, the call to spark_session.schema results in
error. Not
[
https://issues.apache.org/jira/browse/SPARK-33445?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
Punit Shah updated SPARK-33445:
---
Attachment: tsd.csv
> Can't parse decimal type from csv file
>
Punit Shah created SPARK-33445:
--
Summary: Can't parse decimal type from csv file
Key: SPARK-33445
URL: https://issues.apache.org/jira/browse/SPARK-33445
Project: Spark
Issue Type: Bug
[
https://issues.apache.org/jira/browse/SPARK-33327?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
Punit Shah updated SPARK-33327:
---
Description:
The attached csv file has two columns, namely "User" and "FromDate". The
import
[
https://issues.apache.org/jira/browse/SPARK-33327?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
Punit Shah updated SPARK-33327:
---
Description:
The attached csv file has two columns, namely "User" and "FromDate". The
import
[
https://issues.apache.org/jira/browse/SPARK-33327?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
Punit Shah updated SPARK-33327:
---
Description:
The attached csv file has two columns, namely "User" and "FromDate". The
import
[
https://issues.apache.org/jira/browse/SPARK-33327?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
Punit Shah updated SPARK-33327:
---
Description:
The attached csv file has two columns, namely "User" and "FromDate". The
import
[
https://issues.apache.org/jira/browse/SPARK-33327?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
Punit Shah updated SPARK-33327:
---
Description:
The attached csv file has two columns, namely "User" and "FromDate". The
import
[
https://issues.apache.org/jira/browse/SPARK-33327?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17226712#comment-17226712
]
Punit Shah commented on SPARK-33327:
The correct behaviour of running the query should be:
cnt,
[
https://issues.apache.org/jira/browse/SPARK-33327?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
Punit Shah updated SPARK-33327:
---
Attachment: users.csv
> grouped by first and last against date column returns incorrect results
>
Punit Shah created SPARK-33327:
--
Summary: grouped by first and last against date column returns
incorrect results
Key: SPARK-33327
URL: https://issues.apache.org/jira/browse/SPARK-33327
Project: Spark
[
https://issues.apache.org/jira/browse/SPARK-32965?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
Punit Shah reopened SPARK-32965:
The linked duplicate issue won't be fixed because the issue was mixed with a
multiline feature
[
https://issues.apache.org/jira/browse/SPARK-32965?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17200743#comment-17200743
]
Punit Shah commented on SPARK-32965:
It looks similar. I've attached a utf-16le file to this
[
https://issues.apache.org/jira/browse/SPARK-32965?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
Punit Shah updated SPARK-32965:
---
Attachment: 32965.png
> pyspark reading csv files with utf_16le encoding
>
[
https://issues.apache.org/jira/browse/SPARK-32965?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
Punit Shah updated SPARK-32965:
---
Attachment: 16le.csv
> pyspark reading csv files with utf_16le encoding
>
Punit Shah created SPARK-32965:
--
Summary: pyspark reading csv files with utf_16le encoding
Key: SPARK-32965
URL: https://issues.apache.org/jira/browse/SPARK-32965
Project: Spark
Issue Type: Bug
[
https://issues.apache.org/jira/browse/SPARK-32956?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17200163#comment-17200163
]
Punit Shah commented on SPARK-32956:
That may work
> Duplicate Columns in a csv file
>
Punit Shah created SPARK-32956:
--
Summary: Duplicate Columns in a csv file
Key: SPARK-32956
URL: https://issues.apache.org/jira/browse/SPARK-32956
Project: Spark
Issue Type: Bug
[
https://issues.apache.org/jira/browse/SPARK-32888?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
Punit Shah closed SPARK-32888.
--
Resolved by adding documentation
> reading a parallized rdd with two identical records results in a zero
[
https://issues.apache.org/jira/browse/SPARK-32888?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17197069#comment-17197069
]
Punit Shah commented on SPARK-32888:
Thank you for your reply [~viirya] However what I've noticed
[
https://issues.apache.org/jira/browse/SPARK-32888?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17196985#comment-17196985
]
Punit Shah edited comment on SPARK-32888 at 9/16/20, 2:55 PM:
--
Why do we
[
https://issues.apache.org/jira/browse/SPARK-32888?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17196985#comment-17196985
]
Punit Shah commented on SPARK-32888:
Why do we remove lines that are the same as the header? The
Punit Shah created SPARK-32888:
--
Summary: reading a parallized rdd with two identical records
results in a zero count df when read via spark.read.csv
Key: SPARK-32888
URL:
27 matches
Mail list logo