It's merged now and in developer tools page - http://spark.apache.org/developer-tools.html#individual-tests Have some func with PySpark testing!
2018년 12월 5일 (수) 오후 4:30, Hyukjin Kwon <gurwls...@gmail.com>님이 작성: > Hey all, I kind of met the goal with a minimised fix with keeping > available framework and options. See > > https://github.com/apache/spark/pull/23203 > https://github.com/apache/spark-website/pull/161 > > I know it's not perfect and other Python testing framework provide many > good other features but should be good enough for now. > Thanks! > > > 2017년 8월 17일 (목) 오전 2:38, Nicholas Chammas <nicholas.cham...@gmail.com>님이 > 작성: > >> Looks like it doesn’t take too much work to get pytest working on our >> code base, since it knows how to run unittest tests. >> >> https://github.com/apache/spark/compare/master…nchammas:pytest >> <https://github.com/apache/spark/compare/master...nchammas:pytest> >> >> For example I was able to do this from that branch and it did the right >> thing, running only the tests with string in their name: >> >> python [pytest *]$ ../bin/spark-submit ./pytest-run-tests.py >> ./pyspark/sql/tests.py -v -k string >> >> However, looking more closely at the whole test setup, I’m hesitant to >> work any further on this. >> >> My intention was to see if we could leverage pytest, tox, and other test >> tools that are standard in the Python ecosystem to replace some of the >> homegrown stuff we have. We have our own test dependency tracking code, our >> own breakdown of tests into module-scoped chunks, and our own machinery to >> parallelize test execution. It seems like it would be a lot of work to reap >> the benefits of using the standard tools while ensuring that we don’t lose >> any of the benefits our current test setup provides. >> >> Nick >> >> On Tue, Aug 15, 2017 at 3:26 PM Bryan Cutler cutl...@gmail.com >> <http://mailto:cutl...@gmail.com> wrote: >> >> This generally works for me to just run tests within a class or even a >>> single test. Not as flexible as pytest -k, which would be nice.. >>> >>> $ SPARK_TESTING=1 bin/pyspark pyspark.sql.tests ArrowTests >>> On Tue, Aug 15, 2017 at 5:49 AM, Nicholas Chammas < >>> nicholas.cham...@gmail.com> wrote: >>> >>>> Pytest does support unittest-based tests >>>> <https://docs.pytest.org/en/latest/unittest.html>, allowing for >>>> incremental adoption. I'll see how convenient it is to use with our current >>>> test layout. >>>> >>>> On Tue, Aug 15, 2017 at 1:03 AM Hyukjin Kwon <gurwls...@gmail.com> >>>> wrote: >>>> >>>>> For me, I would like this if this can be done with relatively small >>>>> changes. >>>>> How about adding more granular options, for example, specifying or >>>>> filtering smaller set of test goals in the run-tests.py script? >>>>> I think it'd be quite small change and we could roughly reach this >>>>> goal if I understood correctly. >>>>> >>>>> >>>>> 2017-08-15 3:06 GMT+09:00 Nicholas Chammas <nicholas.cham...@gmail.com >>>>> >: >>>>> >>>>>> Say you’re working on something and you want to rerun the PySpark >>>>>> tests, focusing on a specific test or group of tests. Is there a way to >>>>>> do >>>>>> that? >>>>>> >>>>>> I know that you can test entire modules with this: >>>>>> >>>>>> ./python/run-tests --modules pyspark-sql >>>>>> >>>>>> But I’m looking for something more granular, like pytest’s -k option. >>>>>> >>>>>> On that note, does anyone else think it would be valuable to use a >>>>>> test runner like pytest to run our Python tests? The biggest benefits >>>>>> would >>>>>> be the use of fixtures >>>>>> <https://docs.pytest.org/en/latest/fixture.html>, and more >>>>>> flexibility on test running and reporting. Just wondering if we’ve >>>>>> already >>>>>> considered this. >>>>>> >>>>>> Nick >>>>>> >>>>>> >>>>> >>>>> >> >