This is an automated email from the ASF dual-hosted git repository. vinoth pushed a commit to branch asf-site in repository https://gitbox.apache.org/repos/asf/hudi.git
The following commit(s) were added to refs/heads/asf-site by this push: new 2bc927f Travis CI build asf-site 2bc927f is described below commit 2bc927f04b1f5f288a3cfe8d7e63a2168e82dcaa Author: CI <ci...@hudi.apache.org> AuthorDate: Sat Aug 22 22:59:24 2020 +0000 Travis CI build asf-site --- content/docs/0.5.3-quick-start-guide.html | 2 +- content/docs/quick-start-guide.html | 2 +- 2 files changed, 2 insertions(+), 2 deletions(-) diff --git a/content/docs/0.5.3-quick-start-guide.html b/content/docs/0.5.3-quick-start-guide.html index 216b7b2..a7334ac 100644 --- a/content/docs/0.5.3-quick-start-guide.html +++ b/content/docs/0.5.3-quick-start-guide.html @@ -767,7 +767,7 @@ specific commit time and beginTime to “000” (denoting earliest possible comm <span class="kn">from</span> <span class="nn">pyspark.sql.functions</span> <span class="kn">import</span> <span class="n">lit</span> <span class="n">deletes</span> <span class="o">=</span> <span class="nb">list</span><span class="p">(</span><span class="nb">map</span><span class="p">(</span><span class="k">lambda</span> <span class="n">row</span><span class="p">:</span> <span class="p">(</span><span class="n">row</span><span class="p">[</span><span class="mi">0</span><span class="p">],</span> <span class="n">row</span><span class="p">[</span><span class="mi">1</span><span class="p">]),</span> <span class="n">ds</span> [...] -<span class="n">df</span> <span class="o">=</span> <span class="n">spark</span><span class="o">.</span><span class="n">sparkContext</span><span class="o">.</span><span class="n">parallelize</span><span class="p">(</span><span class="n">deletes</span><span class="p">)</span><span class="o">.</span><span class="n">toDF</span><span class="p">([</span><span class="s">'partitionpath'</span><span class="p">,</span> <span class="s">'uuid'</span><span class="p">])</span><span class="o">.</span>< [...] +<span class="n">df</span> <span class="o">=</span> <span class="n">spark</span><span class="o">.</span><span class="n">sparkContext</span><span class="o">.</span><span class="n">parallelize</span><span class="p">(</span><span class="n">deletes</span><span class="p">)</span><span class="o">.</span><span class="n">toDF</span><span class="p">([</span><span class="s">'uuid'</span><span class="p">,</span> <span class="s">'partitionpath'</span><span class="p">])</span><span class="o">.</span>< [...] <span class="n">df</span><span class="o">.</span><span class="n">write</span><span class="o">.</span><span class="nb">format</span><span class="p">(</span><span class="s">"hudi"</span><span class="p">)</span><span class="o">.</span> \ <span class="n">options</span><span class="p">(</span><span class="o">**</span><span class="n">hudi_delete_options</span><span class="p">)</span><span class="o">.</span> \ <span class="n">mode</span><span class="p">(</span><span class="s">"append"</span><span class="p">)</span><span class="o">.</span> \ diff --git a/content/docs/quick-start-guide.html b/content/docs/quick-start-guide.html index ba3f188..2a5c393 100644 --- a/content/docs/quick-start-guide.html +++ b/content/docs/quick-start-guide.html @@ -786,7 +786,7 @@ specific commit time and beginTime to “000” (denoting earliest possible comm <span class="kn">from</span> <span class="nn">pyspark.sql.functions</span> <span class="kn">import</span> <span class="n">lit</span> <span class="n">deletes</span> <span class="o">=</span> <span class="nb">list</span><span class="p">(</span><span class="nb">map</span><span class="p">(</span><span class="k">lambda</span> <span class="n">row</span><span class="p">:</span> <span class="p">(</span><span class="n">row</span><span class="p">[</span><span class="mi">0</span><span class="p">],</span> <span class="n">row</span><span class="p">[</span><span class="mi">1</span><span class="p">]),</span> <span class="n">ds</span> [...] -<span class="n">df</span> <span class="o">=</span> <span class="n">spark</span><span class="o">.</span><span class="n">sparkContext</span><span class="o">.</span><span class="n">parallelize</span><span class="p">(</span><span class="n">deletes</span><span class="p">)</span><span class="o">.</span><span class="n">toDF</span><span class="p">([</span><span class="s">'partitionpath'</span><span class="p">,</span> <span class="s">'uuid'</span><span class="p">])</span><span class="o">.</span>< [...] +<span class="n">df</span> <span class="o">=</span> <span class="n">spark</span><span class="o">.</span><span class="n">sparkContext</span><span class="o">.</span><span class="n">parallelize</span><span class="p">(</span><span class="n">deletes</span><span class="p">)</span><span class="o">.</span><span class="n">toDF</span><span class="p">([</span><span class="s">'uuid'</span><span class="p">,</span> <span class="s">'partitionpath'</span><span class="p">])</span><span class="o">.</span>< [...] <span class="n">df</span><span class="o">.</span><span class="n">write</span><span class="o">.</span><span class="nb">format</span><span class="p">(</span><span class="s">"hudi"</span><span class="p">)</span><span class="o">.</span> \ <span class="n">options</span><span class="p">(</span><span class="o">**</span><span class="n">hudi_delete_options</span><span class="p">)</span><span class="o">.</span> \ <span class="n">mode</span><span class="p">(</span><span class="s">"append"</span><span class="p">)</span><span class="o">.</span> \