Hi Josh, using a regular DELETE is not a bulk delete. Bulk operation use jdbc statement batch, collecting up to 1000 statements into a single execution reducing the amount of communication overhead and improving performance. What i really want to do is add a deleteFromPhoenix() method into ProductRDD to simplify the scenario where an RDD doesn't contains new data to save into but old data to delete from Phoenix. FF. _____
Da: Josh Mahonin [mailto:jmaho...@gmail.com] Inviato: mercoledì 28 settembre 2016 20:29 A: user@phoenix.apache.org Oggetto: Re: bulk-delete spark phoenix Hi Fabio, You could probably just execute a regular DELETE query from a JDBC call, which is generally safe to do either from the Spark driver or within an executor. As long as auto-commit is enabled, it's an entirely server side operation: https://phoenix.apache.org/language/#delete Josh On Wed, Sep 28, 2016 at 2:13 PM, fabio ferrante <fabio.ferra...@eng.it> wrote: Hi, I would like to perform a bulk delete to HBase using Apache Phoenix from Spark. Using Phoenix-Spark plugin i can successfully perform a bulk load using saveToPhoenix method from PhoenixRDD but how i can perform a bulk delete? There isn't a deleteFromPhoenix method in PhoenixRDD. Is that correct? Implement such method is a trivial task? Thanks in advance, Fabio.