I am seeing a bunch of python related (43) failures in the sql module (for
example [1]) ... I am currently on Python 3.11.6, java 8.
Not sure if ubuntu modified anything from under me, thoughts ?

I am currently testing this against an older branch to make sure it is not
an issue with my desktop.

Regards,
Mridul


[1]


org.apache.spark.sql.IntegratedUDFTestUtils.shouldTestGroupedAggPandasUDFs
was false (QueryCompilationErrorsSuite.scala:112)
Traceback (most recent call last):
  File "/home/mridul/work/apache/vote/spark/python/pyspark/serializers.py",
line 458, in dumps
    return cloudpickle.dumps(obj, pickle_protocol)
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File
"/home/mridul/work/apache/vote/spark/python/pyspark/cloudpickle/cloudpickle_fast.py",
line 73, in dumps
    cp.dump(obj)
  File
"/home/mridul/work/apache/vote/spark/python/pyspark/cloudpickle/cloudpickle_fast.py",
line 602, in dump
    return Pickler.dump(self, obj)
           ^^^^^^^^^^^^^^^^^^^^^^^
  File
"/home/mridul/work/apache/vote/spark/python/pyspark/cloudpickle/cloudpickle_fast.py",
line 692, in reducer_override
    return self._function_reduce(obj)
           ^^^^^^^^^^^^^^^^^^^^^^^^^^
  File
"/home/mridul/work/apache/vote/spark/python/pyspark/cloudpickle/cloudpickle_fast.py",
line 565, in _function_reduce
    return self._dynamic_function_reduce(obj)
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File
"/home/mridul/work/apache/vote/spark/python/pyspark/cloudpickle/cloudpickle_fast.py",
line 546, in _dynamic_function_reduce
    state = _function_getstate(func)
            ^^^^^^^^^^^^^^^^^^^^^^^^
  File
"/home/mridul/work/apache/vote/spark/python/pyspark/cloudpickle/cloudpickle_fast.py",
line 157, in _function_getstate
    f_globals_ref = _extract_code_globals(func.__code__)
                    ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File
"/home/mridul/work/apache/vote/spark/python/pyspark/cloudpickle/cloudpickle.py",
line 334, in _extract_code_globals
    out_names = {names[oparg]: None for _, oparg in _walk_global_ops(co)}
                ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File
"/home/mridul/work/apache/vote/spark/python/pyspark/cloudpickle/cloudpickle.py",
line 334, in <dictcomp>
    out_names = {names[oparg]: None for _, oparg in _walk_global_ops(co)}
                 ~~~~~^^^^^^^
IndexError: tuple index out of range
Traceback (most recent call last):
  File "/home/mridul/work/apache/vote/spark/python/pyspark/serializers.py",
line 458, in dumps
    return cloudpickle.dumps(obj, pickle_protocol)
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File
"/home/mridul/work/apache/vote/spark/python/pyspark/cloudpickle/cloudpickle_fast.py",
line 73, in dumps
    cp.dump(obj)
  File
"/home/mridul/work/apache/vote/spark/python/pyspark/cloudpickle/cloudpickle_fast.py",
line 602, in dump
    return Pickler.dump(self, obj)
           ^^^^^^^^^^^^^^^^^^^^^^^
  File
"/home/mridul/work/apache/vote/spark/python/pyspark/cloudpickle/cloudpickle_fast.py",
line 692, in reducer_override
    return self._function_reduce(obj)
           ^^^^^^^^^^^^^^^^^^^^^^^^^^
  File
"/home/mridul/work/apache/vote/spark/python/pyspark/cloudpickle/cloudpickle_fast.py",
line 565, in _function_reduce
    return self._dynamic_function_reduce(obj)
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File
"/home/mridul/work/apache/vote/spark/python/pyspark/cloudpickle/cloudpickle_fast.py",
line 546, in _dynamic_function_reduce
    state = _function_getstate(func)
            ^^^^^^^^^^^^^^^^^^^^^^^^
  File
"/home/mridul/work/apache/vote/spark/python/pyspark/cloudpickle/cloudpickle_fast.py",
line 157, in _function_getstate
    f_globals_ref = _extract_code_globals(func.__code__)
                    ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File
"/home/mridul/work/apache/vote/spark/python/pyspark/cloudpickle/cloudpickle.py",
line 334, in _extract_code_globals
    out_names = {names[oparg]: None for _, oparg in _walk_global_ops(co)}
                ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File
"/home/mridul/work/apache/vote/spark/python/pyspark/cloudpickle/cloudpickle.py",
line 334, in <dictcomp>
    out_names = {names[oparg]: None for _, oparg in _walk_global_ops(co)}
                 ~~~~~^^^^^^^
IndexError: tuple index out of range

During handling of the above exception, another exception occurred:

Traceback (most recent call last):
  File "<string>", line 1, in <module>
  File "/home/mridul/work/apache/vote/spark/python/pyspark/serializers.py",
line 468, in dumps
    raise pickle.PicklingError(msg)
_pickle.PicklingError: Could not serialize object: IndexError: tuple index
out of range
- UNSUPPORTED_FEATURE: Using Python UDF with unsupported join condition ***
FAILED ***



On Sun, Dec 10, 2023 at 9:05 PM L. C. Hsieh <vii...@gmail.com> wrote:

> +1
>
> On Sun, Dec 10, 2023 at 6:15 PM Kent Yao <y...@apache.org> wrote:
> >
> > +1(non-binding
> >
> > Kent Yao
> >
> > Yuming Wang <yumw...@apache.org> 于2023年12月11日周一 09:33写道:
> > >
> > > +1
> > >
> > > On Mon, Dec 11, 2023 at 5:55 AM Dongjoon Hyun <dongj...@apache.org>
> wrote:
> > >>
> > >> +1
> > >>
> > >> Dongjoon
> > >>
> > >> On 2023/12/08 21:41:00 Dongjoon Hyun wrote:
> > >> > Please vote on releasing the following candidate as Apache Spark
> version
> > >> > 3.3.4.
> > >> >
> > >> > The vote is open until December 15th 1AM (PST) and passes if a
> majority +1
> > >> > PMC votes are cast, with a minimum of 3 +1 votes.
> > >> >
> > >> > [ ] +1 Release this package as Apache Spark 3.3.4
> > >> > [ ] -1 Do not release this package because ...
> > >> >
> > >> > To learn more about Apache Spark, please see
> https://spark.apache.org/
> > >> >
> > >> > The tag to be voted on is v3.3.4-rc1 (commit
> > >> > 18db204995b32e87a650f2f09f9bcf047ddafa90)
> > >> > https://github.com/apache/spark/tree/v3.3.4-rc1
> > >> >
> > >> > The release files, including signatures, digests, etc. can be found
> at:
> > >> >
> > >> > https://dist.apache.org/repos/dist/dev/spark/v3.3.4-rc1-bin/
> > >> >
> > >> >
> > >> > Signatures used for Spark RCs can be found in this file:
> > >> >
> > >> > https://dist.apache.org/repos/dist/dev/spark/KEYS
> > >> >
> > >> >
> > >> > The staging repository for this release can be found at:
> > >> >
> > >> >
> https://repository.apache.org/content/repositories/orgapachespark-1451/
> > >> >
> > >> >
> > >> > The documentation corresponding to this release can be found at:
> > >> >
> > >> > https://dist.apache.org/repos/dist/dev/spark/v3.3.4-rc1-docs/
> > >> >
> > >> >
> > >> > The list of bug fixes going into 3.3.4 can be found at the
> following URL:
> > >> >
> > >> > https://issues.apache.org/jira/projects/SPARK/versions/12353505
> > >> >
> > >> >
> > >> > This release is using the release script of the tag v3.3.4-rc1.
> > >> >
> > >> >
> > >> > FAQ
> > >> >
> > >> >
> > >> > =========================
> > >> >
> > >> > How can I help test this release?
> > >> >
> > >> > =========================
> > >> >
> > >> >
> > >> >
> > >> > If you are a Spark user, you can help us test this release by taking
> > >> >
> > >> > an existing Spark workload and running on this release candidate,
> then
> > >> >
> > >> > reporting any regressions.
> > >> >
> > >> >
> > >> >
> > >> > If you're working in PySpark you can set up a virtual env and
> install
> > >> >
> > >> > the current RC and see if anything important breaks, in the
> Java/Scala
> > >> >
> > >> > you can add the staging repository to your projects resolvers and
> test
> > >> >
> > >> > with the RC (make sure to clean up the artifact cache before/after
> so
> > >> >
> > >> > you don't end up building with a out of date RC going forward).
> > >> >
> > >> >
> > >> >
> > >> > ===========================================
> > >> >
> > >> > What should happen to JIRA tickets still targeting 3.3.4?
> > >> >
> > >> > ===========================================
> > >> >
> > >> >
> > >> >
> > >> > The current list of open tickets targeted at 3.3.4 can be found at:
> > >> >
> > >> > https://issues.apache.org/jira/projects/SPARK and search for
> "Target
> > >> > Version/s" = 3.3.4
> > >> >
> > >> >
> > >> > Committers should look at those and triage. Extremely important bug
> > >> >
> > >> > fixes, documentation, and API tweaks that impact compatibility
> should
> > >> >
> > >> > be worked on immediately. Everything else please retarget to an
> > >> >
> > >> > appropriate release.
> > >> >
> > >> >
> > >> >
> > >> > ==================
> > >> >
> > >> > But my bug isn't fixed?
> > >> >
> > >> > ==================
> > >> >
> > >> >
> > >> >
> > >> > In order to make timely releases, we will typically not hold the
> > >> >
> > >> > release unless the bug in question is a regression from the previous
> > >> >
> > >> > release. That being said, if there is something which is a
> regression
> > >> >
> > >> > that has not been correctly targeted please ping me or a committer
> to
> > >> >
> > >> > help target the issue.
> > >> >
> > >>
> > >> ---------------------------------------------------------------------
> > >> To unsubscribe e-mail: dev-unsubscr...@spark.apache.org
> > >>
> >
> > ---------------------------------------------------------------------
> > To unsubscribe e-mail: dev-unsubscr...@spark.apache.org
> >
>
> ---------------------------------------------------------------------
> To unsubscribe e-mail: dev-unsubscr...@spark.apache.org
>
>

Reply via email to