This is an automated email from the ASF dual-hosted git repository.

srowen pushed a commit to branch branch-3.3
in repository https://gitbox.apache.org/repos/asf/spark.git


The following commit(s) were added to refs/heads/branch-3.3 by this push:
     new b40e408ea13 [SPARK-43005][PYSPARK] Fix typo in pyspark/pandas/config.py
b40e408ea13 is described below

commit b40e408ea13d9d2d55d407f0758940cbba5ade3e
Author: thyecust <t...@mail.ecust.edu.cn>
AuthorDate: Mon Apr 3 08:24:17 2023 -0500

    [SPARK-43005][PYSPARK] Fix typo in pyspark/pandas/config.py
    
    By comparing compute.isin_limit and plotting.max_rows, `v is v` is likely 
to be a typo.
    
    ### What changes were proposed in this pull request?
    
    fix `v is v >= 0` with `v >= 0`.
    
    ### Why are the changes needed?
    
    By comparing compute.isin_limit and plotting.max_rows, `v is v` is likely 
to be a typo.
    
    ### Does this PR introduce _any_ user-facing change?
    
    No.
    
    ### How was this patch tested?
    
    By GitHub Actions.
    
    Closes #40620 from thyecust/patch-2.
    
    Authored-by: thyecust <t...@mail.ecust.edu.cn>
    Signed-off-by: Sean Owen <sro...@gmail.com>
    (cherry picked from commit 5ac2b0fc024ae499119dfd5ab2ee4d038418c5fd)
    Signed-off-by: Sean Owen <sro...@gmail.com>
---
 python/pyspark/pandas/config.py | 2 +-
 1 file changed, 1 insertion(+), 1 deletion(-)

diff --git a/python/pyspark/pandas/config.py b/python/pyspark/pandas/config.py
index a0b8db67758..b96828333ef 100644
--- a/python/pyspark/pandas/config.py
+++ b/python/pyspark/pandas/config.py
@@ -233,7 +233,7 @@ _options: List[Option] = [
         default=1000,
         types=int,
         check_func=(
-            lambda v: v is v >= 0,
+            lambda v: v >= 0,
             "'plotting.max_rows' should be greater than or equal to 0.",
         ),
     ),


---------------------------------------------------------------------
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org

Reply via email to