[ 
https://issues.apache.org/jira/browse/CARBONDATA-1168?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16053380#comment-16053380
 ] 

Kingsley Chen commented on CARBONDATA-1168:
-------------------------------------------

I tested it in the version of spark1.6+CarbonData-1.0.0-incubating, it can 
process 100,000+ in one batch. So is that a bug of version 1.1.0?? Is there 
anyone can help me???

> Driver Delete data operation is failed due to failure in creating delete 
> delta file for segment
> -----------------------------------------------------------------------------------------------
>
>                 Key: CARBONDATA-1168
>                 URL: https://issues.apache.org/jira/browse/CARBONDATA-1168
>             Project: CarbonData
>          Issue Type: Bug
>          Components: sql
>    Affects Versions: 1.1.0
>         Environment: spark1.6+carbon1.1.0
> we have a 20 nodes cluster and 32GB ram each
>            Reporter: Kingsley Chen
>              Labels: features
>             Fix For: NONE
>
>
> We use spark code to delete data from table as below
> ------------------spark code----------------------
> val deleteSql = s"DELETE FROM $tableName WHERE $rowkeyName IN 
> (${rowKeyVals.mkString(",")})"
>     cc.sql(deleteSql).show()
> ------------------spark code----------------------
> when the array size of rowKeyVals is larger than 200, the delete operation 
> will be failed, and print the log:
> Delete data request has been received for default.item
> Delete data operation is failed for default.item
> Driver Delete data operation is failed due to failure in creating delete 
> delta file for segment : null block : null
> ++
> ||
> ++
> ++
> That is to say, it only delete success maximun at 200 a batch,and took about 
> 1min which is
> too slow. So my question is how to tuning the performance to make the batch 
> larger 
> and delete faster



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

Reply via email to