Hi Josh,
using a regular DELETE is not a bulk delete. Bulk operation use jdbc
statement batch, collecting up to 1000 statements into a single execution
reducing the amount of communication overhead and improving performance.
What i really want to do is add a deleteFromPhoenix() method into ProductRDD
to simplify the scenario where an RDD doesn't contains new data to save into
but old data to delete from Phoenix.
 
FF.
 
  _____  

Da: Josh Mahonin [mailto:jmaho...@gmail.com] 
Inviato: mercoledì 28 settembre 2016 20:29
A: user@phoenix.apache.org
Oggetto: Re: bulk-delete spark phoenix


Hi Fabio, 

You could probably just execute a regular DELETE query from a JDBC call,
which is generally safe to do either from the Spark driver or within an
executor. As long as auto-commit is enabled, it's an entirely server side
operation: https://phoenix.apache.org/language/#delete

Josh

On Wed, Sep 28, 2016 at 2:13 PM, fabio ferrante <fabio.ferra...@eng.it>
wrote:



Hi,
 
I would like to perform a bulk delete to HBase using Apache Phoenix from
Spark. Using Phoenix-Spark plugin i can successfully perform a bulk load
using saveToPhoenix method from PhoenixRDD but how i can perform a bulk
delete? There isn't a deleteFromPhoenix method in PhoenixRDD. Is that
correct? Implement such method is a trivial task?
 
Thanks in advance,
 Fabio.


Reply via email to