It's merged now and in developer tools page -
http://spark.apache.org/developer-tools.html#individual-tests
Have some func with PySpark testing!

2018년 12월 5일 (수) 오후 4:30, Hyukjin Kwon <gurwls...@gmail.com>님이 작성:

> Hey all, I kind of met the goal with a minimised fix with keeping
> available framework and options. See
>
> https://github.com/apache/spark/pull/23203
> https://github.com/apache/spark-website/pull/161
>
> I know it's not perfect and other Python testing framework provide many
> good other features but should be good enough for now.
> Thanks!
>
>
> 2017년 8월 17일 (목) 오전 2:38, Nicholas Chammas <nicholas.cham...@gmail.com>님이
> 작성:
>
>> Looks like it doesn’t take too much work to get pytest working on our
>> code base, since it knows how to run unittest tests.
>>
>> https://github.com/apache/spark/compare/master…nchammas:pytest
>> <https://github.com/apache/spark/compare/master...nchammas:pytest>
>>
>> For example I was able to do this from that branch and it did the right
>> thing, running only the tests with string in their name:
>>
>> python [pytest *]$ ../bin/spark-submit ./pytest-run-tests.py 
>> ./pyspark/sql/tests.py -v -k string
>>
>> However, looking more closely at the whole test setup, I’m hesitant to
>> work any further on this.
>>
>> My intention was to see if we could leverage pytest, tox, and other test
>> tools that are standard in the Python ecosystem to replace some of the
>> homegrown stuff we have. We have our own test dependency tracking code, our
>> own breakdown of tests into module-scoped chunks, and our own machinery to
>> parallelize test execution. It seems like it would be a lot of work to reap
>> the benefits of using the standard tools while ensuring that we don’t lose
>> any of the benefits our current test setup provides.
>>
>> Nick
>>
>> On Tue, Aug 15, 2017 at 3:26 PM Bryan Cutler cutl...@gmail.com
>> <http://mailto:cutl...@gmail.com> wrote:
>>
>> This generally works for me to just run tests within a class or even a
>>> single test.  Not as flexible as pytest -k, which would be nice..
>>>
>>> $ SPARK_TESTING=1 bin/pyspark pyspark.sql.tests ArrowTests
>>> On Tue, Aug 15, 2017 at 5:49 AM, Nicholas Chammas <
>>> nicholas.cham...@gmail.com> wrote:
>>>
>>>> Pytest does support unittest-based tests
>>>> <https://docs.pytest.org/en/latest/unittest.html>, allowing for
>>>> incremental adoption. I'll see how convenient it is to use with our current
>>>> test layout.
>>>>
>>>> On Tue, Aug 15, 2017 at 1:03 AM Hyukjin Kwon <gurwls...@gmail.com>
>>>> wrote:
>>>>
>>>>> For me, I would like this if this can be done with relatively small
>>>>> changes.
>>>>> How about adding more granular options, for example, specifying or
>>>>> filtering smaller set of test goals in the run-tests.py script?
>>>>> I think it'd be quite small change and we could roughly reach this
>>>>> goal if I understood correctly.
>>>>>
>>>>>
>>>>> 2017-08-15 3:06 GMT+09:00 Nicholas Chammas <nicholas.cham...@gmail.com
>>>>> >:
>>>>>
>>>>>> Say you’re working on something and you want to rerun the PySpark
>>>>>> tests, focusing on a specific test or group of tests. Is there a way to 
>>>>>> do
>>>>>> that?
>>>>>>
>>>>>> I know that you can test entire modules with this:
>>>>>>
>>>>>> ./python/run-tests --modules pyspark-sql
>>>>>>
>>>>>> But I’m looking for something more granular, like pytest’s -k option.
>>>>>>
>>>>>> On that note, does anyone else think it would be valuable to use a
>>>>>> test runner like pytest to run our Python tests? The biggest benefits 
>>>>>> would
>>>>>> be the use of fixtures
>>>>>> <https://docs.pytest.org/en/latest/fixture.html>, and more
>>>>>> flexibility on test running and reporting. Just wondering if we’ve 
>>>>>> already
>>>>>> considered this.
>>>>>>
>>>>>> Nick
>>>>>> ​
>>>>>>
>>>>>
>>>>> ​
>>
>

Reply via email to